Abstract
The differentiation of autoimmune pancreatitis (AIP) and pancreatic ductal adenocarcinoma (PDAC) poses a relevant diagnostic challenge and can lead to misdiagnosis and consequently poor patient outcome. Recent studies have shown that radiomics-based models can achieve high sensitivity and specificity in predicting both entities. However, radiomic features can only capture low level representations of the input image. In contrast, convolutional neural networks (CNNs) can learn and extract more complex representations which have been used for image classification to great success. In our retrospective observational study, we performed a deep learning-based feature extraction using CT-scans of both entities and compared the predictive value against traditional radiomic features. In total, 86 patients, 44 with AIP and 42 with PDACs, were analyzed. Whole pancreas segmentation was automatically performed on CT-scans during the portal venous phase. The segmentation masks were manually checked and corrected if necessary. In total, 1411 radiomic features were extracted using PyRadiomics and 256 features (deep features) were extracted using an intermediate layer of a convolutional neural network (CNN). After feature selection and normalization, an extremely randomized trees algorithm was trained and tested using a two-fold shuffle-split cross-validation with a test sample of 20% (n = 18) to discriminate between AIP or PDAC. Feature maps were plotted and visual difference was noted. The machine learning (ML) model achieved a sensitivity, specificity, and ROC-AUC of 0.89 ± 0.11, 0.83 ± 0.06, and 0.90 ± 0.02 for the deep features and 0.72 ± 0.11, 0.78 ± 0.06, and 0.80 ± 0.01 for the radiomic features. Visualization of feature maps indicated different activation patterns for AIP and PDAC. We successfully trained a machine learning model using deep feature extraction from CT-images to differentiate between AIP and PDAC. In comparison to traditional radiomic features, deep features achieved a higher sensitivity, specificity, and ROC-AUC. Visualization of deep features could further improve the diagnostic accuracy of non-invasive differentiation of AIP and PDAC.
Keywords: deep learning, radiomics, pancreatic cancer, autoimmune pancreatitis
1. Introduction
Autoimmune pancreatitis (AIP) is a rare inflammatory disease that can be classified into two subtypes. AIP type 1 is associated with increased levels of IgG-4 and typically shows other organ involvement; in contrast, AIP type 2 commonly affects only the pancreas. While different diagnostic guidelines were established [1,2,3], the accuracy of these diagnostic scoring systems remains unsatisfactory, with some patients not matching any of the common scoring systems [4,5]. Moreover, different studies show that cases of AIP are at risk for being misdiagnosed as pancreatic ductal adenocarcinoma (PDAC), falsely leading to pancreatic resection [6,7]. This is mainly due to the overlapping clinical and radiological features of both entities. AIP patients commonly present with painless jaundice, making malignancy of the pancreaticobiliary system a likely differential. While AIP typically responds well to corticosteroid therapy, PDACs require resection and/or chemotherapeutic treatment. Computed tomography (CT) is frequently used in the work-up of patients presenting with painless jaundice and different imaging parameters are already implemented in the above-mentioned scoring systems. While some imaging characteristics are suggestive of AIP, for example, “sausage-like” enlargement of the pancreas, AIP can also cause focal mass or abnormal enhancement, making it difficult to differentiate it from pancreatic cancer.
In recent years, radiomics, a computer-based analysis of quantitative imaging features, has been widely used to increase diagnostic accuracy and successfully predict patient outcome and therapy response. Recent studies show that a differentiation between AIP and PDAC based on radiomic features is possible [8,9,10]. Furthermore, radiomic features were successfully used to predict therapeutically relevant subtypes of PDAC, and overall and progression-free survival of PDAC patients [11,12].
Convolutional neural networks (CNNs), which have been used for image classification tasks to great success, are being increasingly applied for medical image analysis, outperforming traditional machine learning (ML) algorithms on large datasets [13,14]. However, small datasets and label uncertainty, a typical limitation seen in the medical field, can hinder the successful end-to-end training of CNNs. Alternatively, CNNs pretrained on a large dataset, for example ImageNet [15], can be used as a feature extractor. These features (deep features) provided promising results for different tasks in image analysis [16,17]. However, few studies have investigated the utility of deep features in medical imaging.
In our study, we trained a machine learning model using deep features and traditional radiomic features and compared their predictive value in the differentiation of AIP and PDAC in portal venous CT-scans.
2. Experimental Section
The study was designed as a retrospective cohort study. Ethics approval for patient recruitment and data processing and analysis was obtained (180/17), and the requirement for individual written consent was waived. All procedures and analyses were carried out in accordance with pertinent laws and regulations. A patient recruitment flowchart and the STROBE (The STrengthening the Reporting of OBservational studies in Epidemiology) checklist [18] are included in the Supplementary Materials. A total of 86 patients, 44 with AIP and 42 with PDAC, were found eligible. Patients without a baseline CT-scan, inadequate or incomplete imaging data, and insufficient ground truth data were excluded. AIP was either diagnosed histopathologically (n = 22) or based on a combination of serology, imaging data, and therapy response to corticosteroids (n = 22) in accordance with the international consensus diagnostic criteria (ICDC) for autoimmune pancreatitis [1]. PDACs were histopathologically diagnosed after tumor resection and the pathological parameters were noted. Based on the 7th edition of the TNM-staging for pancreatic ductal adenocarcinoma, only PDACs limited to the pancreas (tumor stage (T): 1/2) and without regional lymph node metastasis (nodal status (n): 0) were selected. Age at diagnosis, sex, localization, diffuse or focal involvement (if applicable), and suspected malignancy were noted for each patient. For the clinical confounders, the chi-squared test and the Students t-test were used for cross-tabulations and continuous variables, respectively. A significance level of p < 0.05 was chosen.
Participants were screened for eligibility based on a search of the hospital picture archiving system (PACS). A total of 86 patients with portal venous CT-scans (70 s post injection of iodinated contrast media) were obtained between the 1 November 2004 and 1 March 2020. The imaging data were exported in pseudonymized form. Pancreas segmentation was done automatically using an in-house algorithm. Each segmentation mask was controlled under radiological reporting conditions by two experienced observers (S.Z., G.K.); if necessary, segmentation masks were adapted using ITK-SNAP v. 3.8.0 (Penn Image Computing and Science Laboratory (PICSL), University of Pennsylvania, PA, USA and Scientific Computing and Imaging Institute (SCI), University of Utah, UT, USA) [19].
Radiomic feature extraction was done using the Python package PyRadiomics v 3.0 [20]. The detailed settings for the feature extraction can be found in the Supplementary Materials. In total, 1411 features were extracted from the CT-images. The following feature preprocessing steps were applied to eliminate unstable and non-informative features. Low variance features (below 0.1) were excluded. Correlation of the remaining 943 features was calculated using Spearman’s correlation coefficient; features with a correlation coefficient above or below ± 0.9 were excluded. The resulting 299 features were normalized to the (0.1) interval.
For the deep-feature extraction, a VGG19 (CNN architecture which consist of 19 layers (16 convolution layers, 3 Fully connected layer, 5 MaxPool layers, 1 SoftMax layer) [21] pretrained on ImageNet was used. For every patient, the CT-image-stack was resized and extended by two channels to match the input shape of 224 (height) × 224 (width) × 3 (channels). The Hounsfield units were rescaled to (0,255). For each CT-image, 256 feature maps of shape 56 × 56 were calculated and extracted on the layer “block3conv4”, yielding 590,080 parameters. The feature maps were summed and normalized to the (0.1) interval. The mean value was calculated for every feature map, yielding 256 features per patient. Features were preprocessed as described for the radiomic features, yielding 79 features. Feature maps were plotted, and visual differences were noted by two experienced radiologists (S.Z., G.K.).
For the prediction of AIP or PDAC, an extremely randomized trees classifier [22] was fit on the radiomic and deep features. Hyperparameter tuning was performed using grid search with a 5-fold cross-validation. The best parameters were retained (hyperparameters can be found in the Supplementary Materials. Two-fold shuffle-split cross-validation with a test sample fraction of 0.2 was performed and the average sensitivity, specificity and ROC-AUC were obtained. ROC curves of the average performance of the algorithm were plotted for both feature groups. The permutation feature importance was assessed using the feature importance function from scikit-learn [23]. All statistical and machine learning analysis was performed using Python v.3.7.9 (Python Software Foundation, DE, USA) and Keras with a TensorFlow [24] backend.
3. Results
In total, 86 patients (AIP: n = 44, PDAC: n = 42) were analyzed. The distribution of the clinical and histological parameters can be found in Table 1. Of the 42 resected PDACs, 14 were classified as T1 and 28 as T2 tumors. All PDACs were classified as focal lesions.
Table 1.
Variable | AIP (n = 44) | PDAC (n = 42) |
---|---|---|
Age (Years) | Mean: 57 ± 17.3 | Mean:67 ± 10.6 |
Range: 26–82 | Range: 34–88 | |
Sex | Male: 29 (66%) | Male: 19 (45%) |
Female: 15 (34%) | Female: 23 (55%) | |
Focal/Multifocal/Diffuse | Focal: 30 (68%) | |
Multifocal: 2 (5%) | ||
Diffuse: 12 (27%) | ||
Localisation (focal) | Head: 13 (43%) | Head: 30 (71%) |
Body: 4 (14%) | Body: 9 (21%) | |
Tail: 13 (43%) | Tail: 3 (8%) |
After feature extraction and selection, the extremely randomized trees algorithm was trained on deep features (n = 79) and radiomic features (n = 299), achieving higher sensitivity (mean ± std), specificity (mean ± std) and ROC-AUC (mean ± std) over the cross-validation folds for the deep features (0.89 ± 0.11, 0.83 ± 0.06, and 0.90 ± 0.02) compared to radiomic features (0.72 ± 0.11, 0.78 ± 0.06, and 0.80 ± 0.01). ROC curves of the average performance of the machine learning algorithm for the deep features and radiomic features are shown in Figure 1.
Age as a clinical feature was significantly different between groups (p = 0.001). The analysis was repeated for the deep features and the radiomic features with age as an additional feature. No increase in sensitivity, specificity, and ROC-AUC was observed. Feature permutation importance was computed. No feature reached an importance above 0.056 for the deep features and 0.022 for the radiomic features. The feature importance of age was 0.00 for both the deep features and radiomic features, indicating that age did not contribute to the accuracy of the prediction. The 256 feature maps were plotted, and visual differences were noted. A visualization of a subset (n = 64) of the feature maps for exemplary patients can be found in the Supplementary Materials. Figure 2 shows CT-images of exemplary patients and the visualization of a representative feature map, indicating different activation patterns. In the displayed feature map, the PDAC lesion shows the highest activation in the transition zone between normal and tumor tissue, whereas the primary lesion shows no activation. In contrast, the AIP lesion shows activation in the inner structure of the lesion and low activation in the transition zone. This indicates that the deep neural network learns features discriminative between AIP and PDAC, owing to lesion architecture manifesting as imaging features.
4. Discussion
Due to overlapping clinical and imaging characteristics of AIP and PDAC, safely ruling out malignancy in AIP patients is often difficult. Misdiagnosis can dramatically worsen the outcome of both entities. Therefore, further diagnostic tools are needed that increase diagnostic accuracy.
In this study, we compared the predictive value of deep features and radiomic features in the differentiation of AIP and PDAC in portal venous CT-scans using a machine learning algorithm. For our cohort, the prediction based on the deep features achieved a higher sensitivity, specificity, and ROC-AUC compared to handcrafted radiomic features.
Radiomic feature analyses have been used for different tasks in a variety of radiological datasets, achieving promising results, including the differentiation between AIP and PDAC [8,9,10]. In the cited publications, the authors were able to predict AIP and PDAC on CT/PET-CT images with high sensitivity and specificity by combining traditional radiomics feature extraction and selection with a ML model. However, there are certain drawbacks inherent to the methodology of radiomics studies. Radiomic features are based on predefined functions and, therefore, are highly dependent on the defined region of interest (ROI), which requires profound domain knowledge and can be time consuming depending on the sample size. Dependency on bin number and ROI size further complicates a standardization [25,26]. Furthermore, different phantom studies have shown that even under experimental conditions, radiomic features can be unstable [27,28]. In combination with the loss of information, due to the spatial and intensity discretization and the limited complexity of the features, radiomic features may only capture a rudimentary representation of the input image. In contrast, CNNs combine low- and high-level features, which results in a more abstract and complex representation of the input image. CNN-based features are learned specifically to improve the performance on the task given and are therefore more versatile in use, which also allows a transfer of features that have been trained on a similar task. In our study, we used a VGG19 pretrained for image classification on the ImageNet dataset. Several studies have shown that using deep features of a pretrained CNN combined with a machine learning classifier can achieve high sensitivity and specificity on different imaging tasks [29,30] and permit generalization to various other tasks [31].
A major challenge in the successful translation of radiomics and deep learning into clinical practice is the lack of visualization and interpretability. In contrast to radiomic features, the feature maps of CNNs can be plotted for visual interpretation by an observer. In our study, we visualized the 256 feature maps and noted that individual feature maps enable a visual differentiation of both entities. The feature map presented in Figure 2 indicates activation in the internal structure of the AIP lesion and a more peripheral located activation for the PDAC, which is in line with a recent study on pancreatic findings in endoscopic ultrasound images in which high-scoring CNN activation was centrally located in the pancreas parenchyma for AIP [32]. The influence of these feature maps on radiological decision-making needs to be investigated in further studies. A recent study has shown that deep features can be replaced with semantic features defined by a radiologist, indicating the discriminatory ability of deep features and their potential to be visually interpretable [33].
The following limitations of our study have to be addressed. The retrospective nature and small patient cohort limit the generalizability of the results and would require multi-center prospective studies to validate the importance of deep features in the differentiation of AIP and PDAC. Consequently, a transfer learning approach had to be used for the deep feature extraction. The lack of an external validation set means our results cannot be used for predictive inference; the necessity for cross-validation may have yielded overly optimistic results which, however, do not influence the comparison of the radiomic and deep features, which was the main objective of our investigation.
5. Conclusions
In our study, we compared the performance of deep features and radiomic features in classifying AIP or PDAC in portal venous CT-scans. We found that deep features achieved the best performance in discriminating both entities. Feature map visualization shows visually detectable difference, which can further improve radiological decision-making. The benefit of CNNs and deep features trained on CT-images of large AIP and PDAC cohorts should be investigated in following studies.
Supplementary Materials
The following are available online at https://www.mdpi.com/2077-0383/9/12/4013/s1, Feature extraction and hyperparameter tuning; Figure S1: Flowchart showing patient recruitment of AIP and PDAC patients; Figure S2: Feature map visualization; Table S1: Exclusion criteria, Table S2: STROBE checklist.
Author Contributions
Conceptualization, S.Z., G.K. and R.B.; methodology, S.Z.; validation, F.H.; formal analysis, S.Z.; investigation, S.Z. and G.K.; resources, F.H., F.J. and T.M.; data curation, T.M., F.J., F.H.; writing—original draft preparation, S.Z.; writing—review and editing, R.B. and G.K.; visualization, S.Z.; supervision, R.B., M.M. and G.K.; funding acquisition, R.B. and G.K. All authors have read and agreed to the published version of the manuscript.
Funding
This research was funded by the German Research Foundation, Collaborative Research Centers (SFB824, C6 and SPP2177, P4) and the German Cancer Consortium (DKTK).
Conflicts of Interest
The authors declare no conflict of interest. The funders had no role in either the design of the study or the collection, analyses, interpretation of data or in the writing of the manuscript.
Footnotes
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.
References
- 1.Shimosegawa T., Chari S.T., Frulloni L., Kamisawa T., Kawa S., Mino-Kenudson M., Kim M.-H., Klöppel G., Lerch M.M., Löhr M., et al. International Consensus Diagnostic Criteria for Autoimmune Pancreatitis: Guidelines of the international association of pancreatology. Pancreas. 2011;40:352–358. doi: 10.1097/MPA.0b013e3182142fd2. [DOI] [PubMed] [Google Scholar]
- 2.Chari S.T., Smyrk T.C., Levy M.J., Topazian M.D., Takahashi N., Zhang L., Clain J.E., Pearson R.K., Petersen B.T., Vege S.S. Diagnosis of Autoimmune Pancreatitis: The Mayo Clinic Experience. Clin. Gastroenterol. Hepatol. 2006;4:1010–1016. doi: 10.1016/j.cgh.2006.05.017. [DOI] [PubMed] [Google Scholar]
- 3.Pearson R.K., Longnecker D.S., Chari S.T., Smyrk T.C., Okazaki K., Frulloni L., Cavallini G. Controversies in Clinical Pancreatology: Autoimmune pancreatitis: Does it exist? Pancreas. 2003;27:1–13. doi: 10.1097/00006676-200307000-00001. [DOI] [PubMed] [Google Scholar]
- 4.Van Heerde M., Buijs J., Rauws E., Wenniger L., Hansen B., Biermann K., Verheij J., Vleggaar F., Brink M., Beuers U.H.W., et al. A Comparative Study of Diagnostic Scoring Systems for Autoimmune Pancreatitis. Pancreas. 2014;43:559–564. doi: 10.1097/MPA.0000000000000045. [DOI] [PubMed] [Google Scholar]
- 5.Madhani K., Desai H., Wong J., Lee-Felker S., Felker E., Farrell J.J. Tu1468 Evaluation of International Consensus Diagnostic Criteria in the Diagnosis of Autoimmune Pancreatitis: A Single Center North American Cohort Study. Gastroenterology. 2016;150:S910. doi: 10.1016/S0016-5085(16)33081-5. [DOI] [Google Scholar]
- 6.Hardacre J.M., Iacobuzio-Donahue C.A., Sohn T.A., Abraham S.C., Yeo C.J., Lillemoe K.D., Choti M.A., Campbell K.A., Schulick R.D., Hruban R.H., et al. Results of Pancreaticoduodenectomy for Lymphoplasmacytic Sclerosing Pancreatitis. Ann. Surg. 2003;237:853–859. doi: 10.1097/01.SLA.0000071516.54864.C1. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 7.Van Heerde M.J., Biermann K., Zondervan P.E., Kazemier G., Van Eijck C.H.J., Pek C., Kuipers E.J., Van Buuren H.R. Prevalence of Autoimmune Pancreatitis and Other Benign Disorders in Pancreatoduodenectomy for Presumed Malignancy of the Pancreatic Head. Dig. Dis. Sci. 2012;57:2458–2465. doi: 10.1007/s10620-012-2191-7. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8.Park S., Chu L., Hruban R., Vogelstein B., Kinzler K., Yuille A., Fouladi D., Shayesteh S., Ghandili S., Wolfgang C., et al. Differentiating autoimmune pancreatitis from pancreatic ductal adenocarcinoma with CT radiomics features. Diagn. Interv. Imaging. 2020;101:555–564. doi: 10.1016/j.diii.2020.03.002. [DOI] [PubMed] [Google Scholar]
- 9.Zhang Y., Cheng C., Liu Z., Wang L., Pan G., Sun G., Chang Y., Zuo C., Yang X. Radiomics analysis for the differentiation of autoimmune pancreatitis and pancreatic ductal adenocarcinoma in 18 F-FDG PET/CT. Med. Phys. 2019;46:4520–4530. doi: 10.1002/mp.13733. [DOI] [PubMed] [Google Scholar]
- 10.Linning E., Xu Y., Wu Z., Li L., Zhang N., Yang H., Schwartz L.H., Lu L., Zhao B. Differentiation of Focal-Type Autoimmune Pancreatitis From Pancreatic Ductal Adenocarcinoma Using Radiomics Based on Multiphasic Computed Tomography. J. Comput. Assist. Tomogr. 2020;44:511–518. doi: 10.1097/rct.0000000000001049. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 11.Kaissis G., Ziegelmayer S., Lohöfer F., Steiger K., Algül H., Muckenhuber A., Yen H.-Y., Rummeny E., Friess H., Schmid R., et al. A machine learning algorithm predicts molecular subtypes in pancreatic ductal adenocarcinoma with differential response to gemcitabine-based versus FOLFIRINOX chemotherapy. PLoS ONE. 2019;14:e0218642. doi: 10.1371/journal.pone.0218642. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 12.Kaissis G., Ziegelmayer S., Lohöfer F.K., Harder F., Jungmann F., Sasse D., Muckenhuber A., Yen H.-Y., Steiger K., Siveke J.T., et al. Image-Based Molecular Phenotyping of Pancreatic Ductal Adenocarcinoma. J. Clin. Med. 2020;9:724. doi: 10.3390/jcm9030724. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 13.Sun Q., Lin X., Zhao Y., Li L., Yan K., Liang D., Sun D., Li Z. Deep Learning vs. Radiomics for Predicting Axillary Lymph Node Metastasis of Breast Cancer Using Ultrasound Images: Don’t Forget the Peritumoral Region. Front. Oncol. 2020;10:53. doi: 10.3389/fonc.2020.00053. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14.Truhn D., Schrading S., Haarburger C., Schneider H., Merhof D., Kuhl C.K. Radiomic versus Convolutional Neural Networks Analysis for Classification of Contrast-enhancing Lesions at Multiparametric Breast MRI. Radiology. 2019;290:290–297. doi: 10.1148/radiol.2018181352. [DOI] [PubMed] [Google Scholar]
- 15.Russakovsky O., Deng J., Su H., Krause J., Satheesh S., Ma S., Huang Z., Karpathy A., Khosla A., Bernstein M., et al. ImageNet Large Scale Visual Recognition Challenge. Int. J. Comput. Vis. 2015;115:211–252. doi: 10.1007/s11263-015-0816-y. [DOI] [Google Scholar]
- 16.Lei J., Song X., Sun L., Song M., Li N., Chen C. Learning deep classifiers with deep features; Proceedings of the 2016 IEEE International Conference on Multimedia and Expo (ICME); Seattle, WA, USA. 11–15 July 2016; pp. 2–7. [DOI] [Google Scholar]
- 17.Wiggers K.L., Britto A.S., Heutte L., Koerich A.L., Oliveira L.E.S. Document image retrieval using deep features; Proceedings of the 2018 International Joint Conference on Neural Networks (IJCNN); Rio de Janeiro, Brazil. 8–13 July 2018; pp. 1–8. [DOI] [Google Scholar]
- 18.Von Elm E., Altman D., Egger M., Pocock S., Gøtzsche P., Vandenbroucke J. The strengthening the reporting of observational studies in epidemiology (strobe) statement: Guidelines for reporting observational studies. Ann. Intern. Med. 2007;147:573–577. doi: 10.7326/0003-4819-147-8-200710160-00010. [DOI] [PubMed] [Google Scholar]
- 19.Yushkevich P.A., Piven J., Hazlett H.C., Smith R.G., Ho S., Gee J.C., Gerig G. User-guided 3D active contour segmentation of anatomical structures: Significantly improved efficiency and reliability. NeuroImage. 2006;31:1116–1128. doi: 10.1016/j.neuroimage.2006.01.015. [DOI] [PubMed] [Google Scholar]
- 20.Van Griethuysen J.J., Fedorov A., Parmar C., Hosny A., Aucoin N., Narayan V., Beets-Tan R.G., Fillion-Robin J.-C., Pieper S., Aerts H.J.W.L. Computational Radiomics System to Decode the Radiographic Phenotype. Cancer Res. 2017;77:e104–e107. doi: 10.1158/0008-5472.CAN-17-0339. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 21.Simonyan K., Zisserman A. Very deep convolutional networks for large-scale image recognition. arXiv. 20141409.1556 [Google Scholar]
- 22.Geurts P., Ernst D., Wehenkel L. Extremely randomized trees. Mach. Learn. 2006;63:3–42. doi: 10.1007/s10994-006-6226-1. [DOI] [Google Scholar]
- 23.Pedregosa F., Varoquaux G., Gramfort A., Michel V., Thirion B., Grisel O., Blondel M., Prettenhofer P., Weiss R., Dubourg V., et al. Scikit-learn: Machine learning in Python. J. Mach. Learn. Res. 2011;12:2825–2830. [Google Scholar]
- 24.Abadi M., Agarwal A., Barham P., Brevdo E., Chen Z., Citro C., Corrado G.S., Davis A., Dean J., Devin M., et al. TensorFlow: Large-Scale Machine Learning on Heterogeneous Distributed Systems. arXiv. 20161603.04467 [Google Scholar]
- 25.Dercle L., Ammari S., Bateson M., Durand P.B., Haspinger E., Massard C., Jaudet C., Varga A., Deutsch E., Soria J.-C., et al. Limits of radiomic-based entropy as a surrogate of tumor heterogeneity: ROI-area, acquisition protocol and tissue site exert substantial influence. Sci. Rep. 2017;7:7952. doi: 10.1038/s41598-017-08310-5. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 26.Duron L., Balvay D., Perre S.V., Bouchouicha A., Savatovsky J., Sadik J.-C., Thomassin-Naggara I., Fournier L., Lecler A. Gray-level discretization impacts reproducible MRI radiomics texture features. PLoS ONE. 2019;14:e0213459. doi: 10.1371/journal.pone.0213459. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 27.Caramella C., Allorant A., Orlhac F., Bidault F., Asselain B., Ammari S., Jaranowski P., Moussier A., Balleyguier C., Lassau N., et al. Can we trust the calculation of texture indices of CT images? A phantom study. Med. Phys. 2018;45:1529–1536. doi: 10.1002/mp.12809. [DOI] [PubMed] [Google Scholar]
- 28.Berenguer R., Pastor-Juan M.D.R., Canales-Vázquez J., Castro-García M., Villas M.V., Legorburo F.M., Sabater S. Radiomics of CT Features May Be Nonreproducible and Redundant: Influence of CT Acquisition Parameters. Radiology. 2018;288:407–415. doi: 10.1148/radiol.2018172361. [DOI] [PubMed] [Google Scholar]
- 29.Xie M., Jean N., Burke M., Lobell D., Ermon S. Transfer learning from deep features for remote sensing and poverty mapping. arXiv. 20151510.00098 [Google Scholar]
- 30.Huang C., Loy C.C., Tang X. Local similarity-aware deep feature embedding. Adv. Neural Inf. Process. Syst. 2016;1:1270–1278. [Google Scholar]
- 31.Sharif Razavian A., Azizpour H., Sullivan J., Carlsson S. CNN features off-the-shelf: An astounding baseline for recognition; Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops; Columbus, OH, USA. 23–28 June 2014; pp. 512–519. [DOI] [Google Scholar]
- 32.Marya N.B., Powers P.D., Chari S.T., Gleeson F.C., Leggett C.L., Abu Dayyeh B.K., Chandrasekhara V., Iyer P.G., Majumder S., Pearson R.K., et al. Utilisation of artificial intelligence for the development of an EUS-convolutional neural network model trained to enhance the diagnosis of autoimmune pancreatitis. Gut. 2020 doi: 10.1136/gutjnl-2020-322821. [DOI] [PubMed] [Google Scholar]
- 33.Paul R., Liu Y., Li Q., Hall L.O., Goldgof D.B., Balagurunathan Y., Schabath M.B., Gillies R.J. Representation of deep features using radiologist defined semantic features; Proceedings of the 2018 International Joint Conference on Neural Networks (IJCNN); Rio de Janeiro, Brazil. 8–13 July 2018; pp. 1–7. [DOI] [PMC free article] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.