Abstract
High computational cost associated with digital pathology image analysis approaches is a challenge towards their translation in routine pathology clinic. Here, we propose a computationally efficient framework (SuperHistopath), designed to map global context features reflecting the rich tumor morphological heterogeneity. SuperHistopath efficiently combines i) a segmentation approach using the linear iterative clustering (SLIC) superpixels algorithm applied directly on the whole-slide images at low resolution (5x magnification) to adhere to region boundaries and form homogeneous spatial units at tissue-level, followed by ii) classification of superpixels using a convolution neural network (CNN). To demonstrate how versatile SuperHistopath was in accomplishing histopathology tasks, we classified tumor tissue, stroma, necrosis, lymphocytes clusters, differentiating regions, fat, hemorrhage and normal tissue, in 127 melanomas, 23 triple-negative breast cancers, and 73 samples from transgenic mouse models of high-risk childhood neuroblastoma with high accuracy (98.8%, 93.1% and 98.3% respectively). Furthermore, SuperHistopath enabled discovery of significant differences in tumor phenotype of neuroblastoma mouse models emulating genomic variants of high-risk disease, and stratification of melanoma patients (high ratio of lymphocyte-to-tumor superpixels (p = 0.015) and low stroma-to-tumor ratio (p = 0.028) were associated with a favorable prognosis). Finally, SuperHistopath is efficient for annotation of ground-truth datasets (as there is no need of boundary delineation), training and application (~5 min for classifying a whole-slide image and as low as ~30 min for network training). These attributes make SuperHistopath particularly attractive for research in rich datasets and could also facilitate its adoption in the clinic to accelerate pathologist workflow with the quantification of phenotypes, predictive/prognosis markers.
Keywords: deep learning, machine learning, digital pathology, computational pathology, tumor region classification, melanoma, neuroblastoma, breast cancer
Introduction
The analysis of histopathological images of surgical tissue specimens stained with hematoxylin and eosin (H&E) remains a critical decision-making tool used for the routine management of patients with cancer and the evaluation of new therapeutic strategies in clinical trials (1–3). In several precision medicine settings, there is an increasing demand for accurate quantification of histological features. However, in their diagnostic practice, pathologists exercise a predominantly qualitative or semi-quantitative assessment with an inherent degree of inter- and intra-observer variability, which occasionally hampers their consistency (4–7). In the new era of digital pathology, advanced computational image analysis techniques are revolutionizing the field of histopathology by providing objective, robust and reproducible quantification of tumor components, thereby assisting pathologists in tasks such as tumor identification and tumor grading (8, 9). Histopathological image analysis can now be performed in high-resolution H&E-stained whole-slide images (WSI) using state-of-the-art deep learning and classical machine learning approaches for single cell segmentation and/or classification. The new ability to map the spatial context of each single cell also opened new avenues for the study of the tumor micro-environment (10–16), which is key to guide the delivery of precision medicine including immunotherapy.
However, computational pathology is still not widely adopted in the oncological setting. One of the challenges lies in the gigabyte sizes of high-resolution WSIs, which result in computationally expensive approaches. WSIs need to be divided into images patches (typical size: 256x256) before being processed by deep networks such as convolutional neural networks (CNNs) (17). Secondly, single-cell approaches provide markers that are often hard-to-be-evaluated or even interpreted by the pathologists and can be prone to the generalization errors when applied in new unseen dataset. As a result, many promising markers eventually fail to reach the clinic due to a lack of cross-validation in new independent datasets. On the other hand, tissue classification approaches, which target multicellular assemblies and paucicellular areas where individual cells are incorporated into the region segmentation, would be accessible for visual validation by pathologists. Such algorithms would enable the characterization of the distribution and interrelationship of global features that are currently detectable by human perception but not quantifiable without artificial intelligence- (AI-)assisted numerical expression.
Current computed pathology tools primarily focus on individual cell analysis at high-resolution (40x/20x magnification) with limited local context features, whereas pathologists frequently employ collateral information, taking into account the overall tissue microarchitecture. Many established clinical markers are actually identified at low or intermediate magnifications, including tumor architecture-based grading systems (18, 19), stroma-tumor ratio (20, 21), infiltrating lymphocytes (TILs) (22, 23) and necrosis (24–26). This has not been yet fully emulated by computational pathology methodologies. However, some methods for the classification of tissue components have been suggested either using image patch classification typically with a CNN or pixel-level classification/segmentation typically with a U-Net-like architecture (27), mainly for tasks such as the dichotomized classification of tissue (e.g. cancerous vs non-cancerous) (28, 29), the segmentation of a feature of interest (e.g. glands) (16, 30) or multi-type tissue classification (9, 31–35). For segmentation purposes, U-Net-like architectures are usually preferred over CNNs, which have established limitations in conforming to object contours. Yet, CNNs have also resulted in promising segmentation approaches (36–38) with the enhanced capability of classifying a large number of categories (39). Multi-scale approaches incorporating information from various image resolutions have also been proposed (40–43). Different approaches have been explored for the classification of epithelium or stroma using superpixels-based segmentation of image patches with either hand-crafted or deep learning features (44, 45). Bejnordi and colleagues used a similar method for their multi-scale approach for the classification of tissue or non-tissue components on low resolution images and stroma and background regions from intermediate and high resolution images (46). However, these methods are typically performed on high-magnifications image patches (20-40x and more rarely 10x) and are associated with a high computational cost.
Here, we propose a framework (SuperHistopath), which can map most of the global context features that contribute to the rich tumor morphological heterogeneity visible to pathologists at low resolution and used for clinical decision making in a computationally efficient manner. We first apply the well-established simple linear iterative clustering (SLIC) superpixels algorithm (47) directly on the WSI at low resolution (5x magnification) and subsequently classify the superpixels into different tumor region categories using a CNN based on pathologists’ annotations. SuperHistopath particularly capitalizes on:
the use of superpixels which provide visually homogeneous areas of similar size respecting the region boundaries and avoid the potential degradation of classification performance associated with image patches, (no matter how small) spanning over multiple tissue categories.
the use of CNN necessary to accurately classify and map the multiple tissue categories that constitute the rich and complex histological intratumoral heterogeneity.
the computational efficiency, faster processing speed and lower memory requirements associated with processing the WSI at low resolution.
We applied SuperHistopath to H&E-stained images from three different cancer types: clinical cutaneous melanoma, triple-negative breast cancer and tumors arising in genetically-engineered mouse models of high-risk childhood neuroblastoma.
Materials and Methods
Datasets
All digitized whole-slide images (WSI) used in this study were H&E-stained, formalin-fixed and paraffin-embedded (FFPE) sections, and scaled to 5x magnification as presented in Table 1 (image sizes at 5x varied from ~8000x8000 to ~12000x12000 pixels). We applied our framework to clinical patient samples of cutaneous melanoma and triple-negative breast cancer, in addition to tumor samples from transgenic mouse models of childhood neuroblastoma. Both the Th-MYCN and Th-ALKF1174L/MYCN mouse models have been shown to spontaneously develop abdominal tumors, which mirror the major histopathological characteristics of childhood high-risk disease (50, 51).
Table 1.
Cancer type | Number of WSIs | Digital scanner | Pixel resolution (5x magnification) | Dataset |
---|---|---|---|---|
Cutaneous melanoma | 127 | Aperio ImageScope | 2.016 μm | The Cancer Genome Atlas (TCGA) |
Triple-negative breast cancer | 23 | NanoZoomer XR | 2.3 μm | Internal dataset, Collaboration with The Serbian Institute of Oncology |
High-risk neuroblastoma (mouse models) | 73 | NanoZoomer XR | 2.3 μm | Internal dataset Tumors samples coming from established Th-MYCN and Th-ALKF1174L/MYCN transgenic mouse colonies (48, 49) and processed by a clinical histopathological core facility |
Region Classification
First, each dataset was pre-processed using the Reinhard stain normalization (52) to account for stain variabilities that could affect classification. Then, all images were segmented using the simple linear iterative clustering (SLIC) superpixels algorithm, which groups together similar neighboring pixels. With our pathologist’s input, we selected the optimal number of superpixels by visually identifying a superpixel size that capture only homogeneous areas and adhere to image boundaries. This is a critical step for ensuring accurate tissue segmentation, and therefore, classification ( Figure 1 ). The number of superpixels was adapted for each image to ensure a homogenous superpixel size across the datasets and was automatically set based on the image size according to Equation 1 (53).
(1) |
where Ni is the number of superpixels in the ith image, Si is the size of image i in pixels, and U is a constant held across all images that defined the desired superpixels size.
The SLIC algorithm inherently provides a roughly uniform superpixel size. Setting U = 1500, Equation 1 gave a mean superpixels size of 51 × 51 pixels, equivalent to an area of approximately 117 × 117 μm2. Bilinear interpolation was subsequently use to resize each superpixel to a fixed size of 56 x 56 or 75 x 75 pixels (the minimum input size for inception-like network architectures).
Region annotations were provided by a senior pathologist with over 20 years of experience for the melanoma and breast cancer clinical datasets, and a senior pediatric neuropathologist with over 20 years of experience for the neuroblastoma mouse datasets. For training and testing, superpixels were assigned to each category based on their isocenter location within the annotated regions. Note that region annotations for our algorithm do not need to delineate boundaries as illustrated in Figure 1B .
The numbers of clinically relevant tissue categories, number of WSIs and superpixels used for training and testing are summarized for each tumor types in Table 2 . Standard image augmentations, such as rotations (90°, -90°, 180°), flips (horizontal and vertical), and contrast (histogram equalization) were performed in each case to capture more variation and even out the training dataset imbalances.
Table 2.
Cancer type | Number of WSIs used for network training | Regional classification | ||
---|---|---|---|---|
Cutaneous melanoma | Total | 27 | 6 categories | Superpixels for training |
Training | 22 | Tumor tissue | 21940 | |
Testing | 5 | Stroma | 12419 | |
Normal epidermis | 1646 | |||
Lymphocytes cluster | 2367 | |||
Fat | 15484 | |||
Empty/white space | 3412 | |||
Triple-negative breast cancer | Total | 23 | 6 categories | Superpixels for training |
Training | 18 | Tumor tissue | 18873 | |
Testing | 5 | Stroma | 24220 | |
Necrosis | 15102 | |||
Lymphocytes cluster | 3472 | |||
Fat | 10044 | |||
Empty/white space | 16473 | |||
High-risk neuroblastoma (mouse model) | Total | 60 | 8 categories | Superpixels for training |
Training | 44 | Region of undifferentiated neuroblasts | 20512 | |
Testing | 16 | Tissue damage (necrosis/apoptosis) | 17645 | |
Differentiation region | 5740 | |||
Lymphocytes cluster | 4009 | |||
Hemorrhage (blood) | 6124 | |||
Muscle | 6415 | |||
Kidney | 14976 | |||
Empty/white space | 21470 |
Note that the testing datasets consisted of whole-slide images from different patients from the training dataset.
Training of the Convolutional Neural Networks
Our custom-designed CNN for superpixel classification consists of 6 convolutional layers (32, 32, 64, 64, 128, 128 neurons, respectively) of 3 x 3 filter size and 3 max-pooling layers, followed by a “flatten” layer and a dense layer of 256 neurons ( Figure 2 ). A superpixel RGB image (post-interpolation) was used as input into the network and normalized from range 0–255 to range 0–1 using the maximum value. The output of the network was a label assigned to each superpixel based on which region category it belonged to. After empirical experimentation, a ReLU activation function was used in all layers except for the last layer where standard softmax was used for classification. The weights incident to each hidden unit were constrained to have a norm value less than or equal to 3 and a dropout unit of 0.2 was used before every max-pooling operation to avoid overfitting (54). The weights of the layers were randomly initialized using “Glorot uniform” initialization (55), and the network was optimized using the Adam method (56) with a learning rate of 10-3 and a categorical cross-entropy cost function. The number of trainable parameters for our custom-made network is ~1.9 M. The network was implemented in python (v. 3.6.5) using the Keras/Tensorflow libraries (v. 2.2.4/1.12.0, respectively).
To choose the best network for our framework, we tested other known neural network architectures as implemented in the Keras framework, including InceptionV3 (57), Xception (58), InceptionResNetV2 (59), and ResNet (60). We initialized the weights using the pre-trained ImageNet weights. To optimize each network, we excluded the final classification layer, and added three additional layers, i) a global average pooling layer, ii) a dense layer of 256 neurons with ReLu activation, constrained to have a norm value less than or equal to 3, and iii) a dense layer tailored to the number of classes of each cancer type using the softmax function for classification.
For inception-like architectures (Inception v3, InceptionResNetV2, Xception) only superpixels of size 75 x 75 were used. We trained all the networks for 50 epochs using batch sizes of 150 and 256 for superpixels of sizes 75 x 75 and 56 x 56, respectively, and kept the models with the highest validation accuracy.
The Xception and custom-made networks were re-trained from the beginning for each cancer type, without applying any further changes.
Application of SuperHistopath for the Quantification of Clinical Features of Interest
In the melanoma dataset, we calculated the number of pixels belonging to each classified category. For each patient we derived i) the ratio of pixels classified as stroma region to all pixels in tumor compartments, and ii) the ratio of pixels classified as clusters of lymphocytes to all pixels in tumor compartments; we evaluated the prognostic value of these quantitative indices using survival analysis. Patients were divided into high- and low-risk groups based on split at the median value of all scores to ensure both groups were of similar size. Kaplan-Meier estimation was used to compare overall survival in the 127 patients. Differences between survival estimates were assessed with the log-rank test and hazard ratios were calculated using Cox’s proportional-hazard regression.
In the neuroblastoma dataset, we evaluated the differences in phenotype between the Th-ALKF1174L/MYCN (n=7) and Th-MYCN tumors (n=6) by quantifying the proportion of pixels classified by our SuperHistopath as regions rich in undifferentiated neuroblasts, differentiating neuroblasts, tissue damage (necrosis/apoptosis) hemorrhage and clusters of lymphocytes. Note that i) we did not quantify stroma in these tumors as they faithfully mirror the stroma-poor phenotype which define high-risk disease ii) lymphocytes clusters universally correspond to encapsulation of lymph node by the tumor, rather that tumor infiltrates, consistent with the “cold” immune phenotype of high-risk disease. We focus on identifying any significant difference in the ratio of differentiation or the ratio of hemorrhagic regions to all tumor compartments between the two tumor types using the Mann-Whitney U test, with a 5% level of significance.
Results
SuperHistopath Can Accurately Map the Complex Histological Heterogeneity of Tumors
Melanoma
We first developed and evaluated our framework on the H&E-stained, FFPE sections of clinical specimen of cutaneous melanoma scaled to 5x magnification. Figure 1 shows the results of the segmentation using the simple linear iterative clustering (SLIC) superpixels algorithm, which groups together similar neighboring pixels.
The optimized Xception network achieved the highest score and classified the melanoma sample regions into 6 predefined tissue categories of interest: tumor tissue, stroma, cluster of lymphocytes, normal epidermis, fat, and empty/white space with an overall accuracy of 98.8%, an average precision of 96.9%, and an average recall of 98.5% over 14,092 superpixels in a separate test set of five images ( Tables 3 , 4 ). Our custom CNN also achieved comparable performance to the state-of-the-art networks with an overall accuracy of 96.7%, an average precision of 93.6%, and an average recall of 93.6% ( Figure 2 , Supplementary Table 1 ). The confusion matrices for the XCeption and our custom CNN networks are presented in Table 4 and Supplementary Table 1 , respectively. Figure 3 shows qualitative results of our approach’s regional classification in representative melanoma WSIs using the optimized Xception network.
Table 3.
Network | Accuracy (%) | Precision (%) | Recall (%) | Parameters (in millions) |
---|---|---|---|---|
InceptionV3 | 97.5 | 94.2 | 96.7 | ~22.4 |
InceptionResNetV2 | 97.7 | 94.1 | 97.3 | ~54.8 |
ResNet50 | 93.8 | 92.2 | 88.9 | ~24.2 |
Xception | 98.8 | 96.9 | 98.5 | ~21.4 |
Our custom-made CNN | 96.7 | 93.6 | 93.6 | ~1.9 |
The bold values in the Accuracy (%), Precision (%) and Recall (%) fields indicate the highest value i.e. the best performance achieved amongst the networks under comparison. The bold value in the Parameters (in millions) field indicate the network with the fewer parameters used amongst the networks under comparison.
Table 4.
Tumor | Stroma | Epidermis | Lym | Fat | Empty space | |
---|---|---|---|---|---|---|
Tumor | 5286 | 10 | 7 | 8 | 0 | 0 |
Stroma | 9 | 986 | 0 | 0 | 2 | 0 |
Epidermis | 22 | 0 | 545 | 0 | 1 | 0 |
Lym | 0 | 0 | 1 | 821 | 0 | 0 |
Fat | 0 | 9 | 0 | 0 | 5603 | 3 |
Empty space | 0 | 0 | 0 | 0 | 98 | 681 |
Overall accuracy = 98.8%, average precision = 96.9%, average recall = 98.5%.
The bold values indicate the correct predictions of the network.
Breast Cancer
SuperHistopath classified sample regions into 6 predefined tissue categories of interest: tumor, necrosis, stroma, cluster of lymphocytes, fat, and lumen/empty space with an overall accuracy of 93.1%, an average precision of 93.9%, and an average recall of 93.6% using Xception and 91.7%, 92.5%, 91.8% respectively using our custom-made CNN over 10,349 superpixels in the independent test set of five images. The confusion matrices for the XCeption and our custom CNN networks are presented in Table 5 and Supplementary Table 2 , respectively. Figure 4 shows qualitative results our approach’s regional classification in representative triple-negative breast cancer WSIs.
Table 5.
Tumor | Necrosis | Lym | Stroma | Fat | Empty space | |
---|---|---|---|---|---|---|
Tumor | 1830 | 13 | 15 | 42 | 0 | 0 |
Necrosis | 50 | 1446 | 2 | 320 | 0 | 0 |
Lym | 4 | 2 | 705 | 10 | 0 | 0 |
Stroma | 42 | 120 | 20 | 3836 | 0 | 1 |
Fat | 0 | 0 | 0 | 0 | 562 | 5 |
Empty space | 0 | 0 | 0 | 0 | 67 | 1257 |
Overall accuracy = 93.1%, average precision = 93.9%, average recall = 93.6%.
The bold values indicate the correct predictions of the network.
Neuroblastoma
SuperHistopath classified the tumor regions into eight predefined tissue categories of interest: undifferentiated neuroblasts, tissue damage (necrosis/apoptosis), areas of differentiation, cluster of lymphocytes, hemorrhage, muscle, kidney, and empty/white space with an overall accuracy of 98.3%, an average precision of 98.5%, and an average recall of 98.4% using Xception and 96.8%, 97.1%, 97.2% respectively using our custom-made CNN over 9,868 superpixels in the independent test set of 16 images. The confusion matrices for the XCeption and our custom CNN networks are presented in Table 6 and Supplementary Table 3 , respectively. Figure 5 shows qualitative results of our approach’s regional classification in representative WSIs of neuroblastoma arising in the Th-MYCN mouse model.
Table 6.
Undifferentiated region | Necrosis | Lym | Differentiation | Blood | Empty space | Muscle | Kidney | |
---|---|---|---|---|---|---|---|---|
Undifferentiated region | 1403 | 3 | 0 | 14 | 1 | 0 | 0 | 0 |
Necrosis | 13 | 1642 | 1 | 26 | 49 | 2 | 5 | 18 |
Lym | 6 | 5 | 1150 | 0 | 0 | 0 | 0 | 3 |
Differentiation | 0 | 0 | 0 | 1261 | 0 | 0 | 0 | 0 |
Blood | 1 | 7 | 0 | 0 | 1327 | 0 | 9 | 0 |
Empty space | 0 | 2 | 0 | 0 | 0 | 560 | 3 | 2 |
Muscle | 0 | 2 | 0 | 0 | 1 | 0 | 1176 | 0 |
Kidney | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1176 |
Overall accuracy = 98.3%, average precision = 98.5%, average recall = 98.4%.
The bold values indicate the correct predictions of the network.
SuperHistopath Pipeline for the Analysis of Low-Resolution WSI Affords Significant Speed Advantages
The average time for the SLIC superpixels algorithm to segment a WSI in 5x magnification was < 2 min using a 3.5 GHz Intel core i7 processor. The average time for both the Xception and our custom-made CNN network to classify every superpixel in the images was 1–2 min using the same processor. A quick convergence of the networks (around epoch 30) was observed in all cases, which needed ~3 h for Xception and only ~30 min for our custom-made CNN using a Tesla P100-PCIE-16GB GPU card, and therefore the latter was used for experimenting.
SuperHistopath Can Provide Robust Quantification of Clinically Relevant Features
Stroma-to-Tumor Ratio and Clusters of Lymphocytes Abundance as Predictive Markers of Survival in Melanoma
We first use SuperHistopath to quantify both the stroma-to-tumor ratio and the immune infiltrate, which have both shown to provide prognostic and predictive information in patient with solid tumors, including melanoma (20, 21, 23). The important role of immune hotspots has been established based on density analysis of single cell classification of lymphocytes in high-resolution images (61, 62). Here, we demonstrate in our melanoma dataset of 127 WSIs i) that a high stromal ratio as identified in low resolution WSIs is a predictor of poor prognosis (SuperHistopath: p = 0.028, Coxph-Regression [discretized by median]: HR = 2.1, p = 0.0315; Figure 6A ) and ii) that clusters of lymphocytes hold predictive information in our melanoma dataset, with a high lymphocyte ratio being an indicator of favorable prognosis [SuperHistopath: p = 0.015, Coxph-Regression (discretized by median): HR = 0.4, p = 0.018; Figure 6B ]. Pearson’s correlation showed no significant correlation between stromal ratio and clusters of lymphocytes ratio (r = -0.13, p = 0.13), and between absolute sizes of stroma and clusters of lymphocytes (r = 0.13, p=0.11). Taken together, our data, captured from low resolution (5x) WSIs, are consistent with those extracted from single-cell analysis in high-resolution WSIs (53).
Necrosis Quantification
We use the SuperHistopath to quantity tumor necrosis in our breast cancer and childhood neuroblastoma preclinical datasets. Tumor necrosis, defined as confluent cell death or large area of tissue damage hold predictive and prognostic information, both at diagnosis and after chemotherapy, in many solid tumors including breast cancer and childhood malignancies (24–26, 63, 64). While visible at 5x objective lens magnification, its quantification can often be a challenging task even for experienced pathologists. Here, we show that SuperHistopath can provide satisfactory quantification of necrosis in clinical breast cancer samples by distinguishing from stroma with high specificity (91.5%) and satisfactory precision (79.5%) and in the high-risk neuroblastoma mouse models with high precision and specificity (93.5% and 98.9% respectively).
Quantification of Neuroblastoma Differentiation
We used SuperHistopath to quantify the phenotype of MYCN-driven transgenic mouse models of high-risk stroma-poor neuroblastoma. We show that SuperHistopath can identify areas of differentiation, a critical feature for the stratification of children neuroblastoma, with both high precision and specificity (100% and 96.9% respectively). SuperHistopath also showed that expression of ALKF1174L mutation significantly shift the MYCN-driven phenotype from poorly-differentiated and hemorrhagic phenotype (Th-MYCN: 1.8 ± 1.3% differentiating area and 29.2 ± 6.7% hemorrhage, Figure 6C ) into a differentiating phenotype also characterized by the almost complete abrogation of the hemorrhagic phenotype (Th-ALKF1174L/MYCN: 20.3 ± 3.1% differentiating area and 0.2 ± 0.1% hemorrhage, p=0.0003 and p=0.0008 respectively, Figure 6D ) as previously demonstrated (51, 65).
Discussion
In this study, we implemented SuperHistopath: a digital pathology pipeline for the classification of tumor regions and the mapping of tumor heterogeneity from low-resolution H&E-stained WSIs, which we demonstrated to be highly accurate in three types of cancer. Combining the application of the SLIC superpixels algorithm directly on low magnification WSIs (5x) with a CNN architecture for the classification of superpixels, contributes to SuperHistopath computational efficiency allowing for fast processing, whilst affording the quantification of robust and easily interpretable clinically-relevant markers.
Applying our computational approach on low-resolution images leads to markedly increased processing speed, for both the classification of new samples and network training. Here, we chose the (5x) magnification as a compromise between tumor structures visibility and computational cost. Specific metrics such as stroma-to-tumor ratio could potentially be derived from images at even lower magnifications (e.g. 1.25x) as recently shown (53). Digital histology images are conventionally processed at 40x (or 20x) magnification where cell morphology is most visible. At those resolutions, WSIs are large (representative size at 20x: 60000 x 60000 pixels), requiring of a lot of memory and images to be divided into patches (tiles) for processing. Under these conditions, the training of new networks for cell segmentation and classification typically requires days and the application to new WSI samples can take hours prior to code optimization. In contrast, the training of our neural network until acceptable convergence needed as little as ~30 min and application on new samples ~5 min (for both superpixel segmentation and classification) in our study. High-resolution images are essential when studying cell-to-cell interactions, however we show that the processing of low resolution images is appropriate for the extraction of specific global context features.
Furthermore, SuperHistopath combines the main advantages of regional classification and segmentation approaches. On one hand, classification approaches applied on smaller patches resulting from splitting WSIs allow the use of CNN for the robust classification of many categories necessary to capture intratumor heterogeneity (39), yet at the expense of higher risk of misclassification, especially close to regional boundaries where an image patch, regardless of its size, may contain multiple tumor components. Overlapping (sliding) window approaches can improve the issue, yet at an increased computational cost. On the other hand, segmentation approaches such as U-Net-like architectures can resolve the regional boundaries issue but appear to work better for few classes, typically two. SuperHistopath efficiently combines the use of a segmentation approach using superpixels to adhere to region boundaries with CNN classification to cover the rich tumor histological heterogeneity (here 6-8 region categories depending on the cancer type).
Our method also markedly simplifies and accelerates the process of preparing ground-truth (annotations) datasets as i) the use of superpixels alleviate the need for careful boundary delineation of the tumor components of interest ( Figure 1B ), a cumbersome and time-consuming process necessary for using U-Net-like architectures and ii) each annotated region contains large numbers of superpixels facilitating the collection of the large datasets traditionally required by deep learning methods.
The appropriate choice of superpixel size is crucial to warrant both accurate tissue segmentation and classification. Equation 1 ensured a uniform superpixel size for every whole-slide image regardless of their original size. The main considerations for choosing superpixels size (i.e. setting the constant U) is to ensure that they only contain a single tissue type, while being large enough to contain sufficient tissue information. In our study, we found that classification is not sensitive to small changes of U. However larger superpixels (U > 1750) did not adhere well to the tissue boundaries, whereas smaller superpixels (U < 1250) indeed led to a slight decrease in classification performance.
Many promising computational pathology-derived biomarkers ultimately fail to translate in the clinic due to their inherent complexity and the difficulty for pathologists to evaluate them in new datasets. In this proof-of-concept study, we showed that SuperHistopath can quantify well-understood features/markers already used, albeit only qualitatively or semi-quantitatively, by pathologists, including the stroma-to-tumor ratio, lymphocyte infiltration, tumor necrosis, and neuroblastoma differentiation. We also show that SuperHistopath-derived results corroborated those obtained from single-cell analysis on high-resolution samples (53). The computational efficiency of SuperHistopath, combined with the simple superpixels-enabled data collection, could facilitate its adoption in the clinic to accelerate pathologist workflow, could assist in intra-operative pathological diagnosis and should facilitate working with large datasets in clinical research.
Moving forward, we plan to expand the types of global context features extractable from SuperHistopath in more cancer types. We will also evaluate the accuracy of SuperHistopath on digitized frozen tissue sections to demonstrate its potential to assist in the rapid intra-operative pathological diagnostic. We will also update our previous framework (SuperCRF) which incorporates region classification information to improve cell classification (53) using SuperHistopath. Together both SuperHistopath and SuperCRF would provide invaluable tools to study spatial interactions across length scales to provide a deeper understanding of the cancer-immune-stroma interface, key to further unlock the potential of cancer immunotherapy (17).
In this proof-of-concept study, we applied our method to three cancer types with disparate histology without any changes (just retraining). While the approach could thus be virtually extended to any type of cancer, improvement could be made tailored to a specific global feature, cancer type or dataset and could include further exploring i) the use of SVM to combine the CNN-extracted features with handcrafted ones, ii) the use of other image color spaces which has been shown to improve classification in certain cases (66) and iii) alternative superpixel algorithms such as the efficient topology preserving segmentation (ETPS) algorithm (67). Additionally, further improvement of this proof-of-concept framework could be sought via experimentation with hyperparameter tuning, or the use of other custom and well-established architectures (59, 68). Since superpixels only capture small homogeneous areas, combination with other approaches such as classification of larger image patches with a deepCNN or U-net-like architectures might be more appropriate for the single purpose of segmenting some large and multi-component tumor structures, e.g. certain types of glands (16).
To conclude, our novel pipeline, SuperHistopath can accurately classify and map the complex tumor heterogeneity from low-resolution H&E-stained histology images. The resulting enhanced speed for both training and application (~5 min for classifying a WSI and as low as ~30 min for network training) and the efficient and simple collection of ground-truth datasets make SuperHistopath particularly attractive for research in rich datasets and would facilitate its adoption in the clinic to accelerate pathologist workflow in the quantification of predictive/prognosis markers derived from global features of interest.
Data Availability Statement
The melanoma dataset comes the publicly available TCGA dataset. The neuroblastoma dataset is available from the corresponding authors upon reasonable request. The images from the triple-negative breast cancer dataset cannot be released yet due to ongoing clinical studies. The codes that support the findings of this study are available from the corresponding authors upon reasonable request.
Ethics Statement
The breast cancer clinical dataset was generated from diagnostic H&E images provided anonymised to the researchers by the Serbian Institute of Oncology. The neuroblastoma preclinical dataset was built from H&E images collected during previous in vivo studies approved by The Institute of Cancer Research Animal Welfare and Ethical Review Body and performed in accordance with the UK Home Office Animals (Scientific Procedures) Act 1986. The melanoma clinical samples come from the publicly available TCGA dataset ( Table 1 ).
Author Contributions
Conception and design: KZ-P, IR, YJ, YY. Development of methodology: KZ-P Analysis and interpretation of data: KZ-P, RN, IR, YJ, YY. Administrative and/or material support: RN, DK, IR, YJ. Writing and review of the manuscript: KZ-P, IR, YJ, YY. IR, YJ, and YY are co-senior authors of this study. All authors contributed to the article and approved the submitted version.
Funding
We acknowledge financial support by The Rosetrees Trust (KZ-P, M593). RN acknowledges funding from ISCIII (FIS) and FEDER (European Regional Development Fund) PI17/01558 and CB16/12/00484. YY acknowledges funding from Cancer Research UK Career Establishment Award (C45982/A21808), Breast Cancer Now (2015NovPR638), Children’s Cancer and Leukaemia Group (CCLGA201906), NIH U54 CA217376, and R01 CA185138, CDMRP Breast Cancer Research Program Award BC132057, CRUK Brain Tumor Awards (TARGET-GBM), European Commission ITN (H2020-MSCA-ITN-2019), Wellcome Trust (105104/Z/14/Z), and The Royal Marsden/ICR National Institute of Health Research Biomedical Research Centre. YJ is a Children with Cancer UK Research Fellow (2014/176). We thank Breast Cancer Now for funding IR as part of Programme Funding to the Breast Cancer Now Toby Robins Research Centre.
Conflict of Interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Acknowledgments
We acknowledge Dr Snezana Susnjar’s and Dr Natasa Medic Milijic’s contribution to the creation and curation of the triple negative breast cancer dataset at The Serbian Institute of Oncology, which has made this study possible.
Supplementary Material
The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fonc.2020.586292/full#supplementary-material
References
- 1. Tabesh A, Teverovskiy M, Pang H-Y, Kumar VP, Verbel D, Kotsianti A, et al. Multifeature prostate cancer diagnosis and Gleason grading of histological images. IEEE Trans Med Imaging (2007) 26(10):1366–78. 10.1109/TMI.2007.898536 [DOI] [PubMed] [Google Scholar]
- 2. Madabhushi A. Digital pathology image analysis: opportunities and challenges. Imaging Med (2009) 1(1):7. 10.2217/iim.09.9 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 3. Kumar R, Srivastava R, Srivastava S. Detection and classification of cancer from microscopic biopsy images using clinically significant and biologically interpretable features. J Med Eng (2015) 2015. 10.1155/2015/457906 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 4. Allard FD, Goldsmith JD, Ayata G, Challies TL, Najarian RM, Nasser IA, et al. Intraobserver and interobserver variability in the assessment of dysplasia in ampullary mucosal biopsies. Am J Surg Pathol (2018) 42(8):1095–100. 10.1097/PAS.0000000000001079 [DOI] [PubMed] [Google Scholar]
- 5. Gomes DS, Porto SS, Balabram D, Gobbi H. Inter-observer variability between general pathologists and a specialist in breast pathology in the diagnosis of lobular neoplasia, columnar cell lesions, atypical ductal hyperplasia and ductal carcinoma in situ of the breast. Diagn Pathol (2014) 9(1):121. 10.1186/1746-1596-9-121 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 6. Krupinski EA, Tillack AA, Richter L, Henderson JT, Bhattacharyya AK, Scott KM, et al. Eye-movement study and human performance using telepathology virtual slides. Implications for medical education and differences with experience. Hum Pathol (2006) 37(12):1543–56. 10.1016/j.humpath.2006.08.024 [DOI] [PubMed] [Google Scholar]
- 7. Mukhopadhyay S, Feldman MD, Abels E, Ashfaq R, Beltaifa S, Cacciabeve NG, et al. Whole slide imaging versus microscopy for primary diagnosis in surgical pathology: a multicenter blinded randomized noninferiority study of 1992 cases (pivotal study). Am J Surg Pathol (2018) 42(1):39. 10.1097/PAS.0000000000000948 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8. Kothari S, Phan JH, Stokes TH, Wang MD. Pathology imaging informatics for quantitative analysis of whole-slide images. J Am Med Inform Assoc (2013) 20(6):1099–108. 10.1136/amiajnl-2012-001540 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 9. Campanella G, Hanna MG, Geneslaw L, Miraflor A, Silva VWK, Busam KJ, et al. Clinical-grade computational pathology using weakly supervised deep learning on whole slide images. Nat Med (2019) 25(8):1301–9. 10.1038/s41591-019-0508-1 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 10. Jones TR, Kang IH, Wheeler DB, Lindquist RA, Papallo A, Sabatini DM, et al. CellProfiler Analyst: data exploration and analysis software for complex image-based screens. BMC Bioinf (2008) 9(1):482. 10.1186/1471-2105-9-482 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 11. Yuan Y, Failmezger H, Rueda OM, Ali HR, Gräf S, Chin S-F, et al. Quantitative image analysis of cellular heterogeneity in breast tumors complements genomic profiling. Sci Transl Med (2012) 4(157):157ra43–ra43. 10.1126/scitranslmed.3004330 [DOI] [PubMed] [Google Scholar]
- 12. Chen CL, Mahjoubfar A, Tai L-C, Blaby IK, Huang A, Niazi KR, et al. Deep learning in label-free cell classification. Sci Rep (2016) 6:21471. 10.1038/srep21471 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 13. Sirinukunwattana K, Raza SEA, Tsang Y-W, Snead DR, Cree IA, Rajpoot NM. Locality sensitive deep learning for detection and classification of nuclei in routine colon cancer histology images. IEEE Trans Med Imaging (2016) 35(5):1196–206. 10.1109/TMI.2016.2525803 [DOI] [PubMed] [Google Scholar]
- 14. Bankhead P, Loughrey MB, Fernández JA, Dombrowski Y, McArt DG, Dunne PD, et al. QuPath: Open source software for digital pathology image analysis. Sci Rep (2017) 7(1):16878. 10.1038/s41598-017-17204-5 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15. Khoshdeli M, Cong R, Parvin B. eds. “Detection of nuclei in H&E stained sections using convolutional neural networks. Biomedical & Health Informatics (BHI)”. In: 2017 IEEE EMBS International Conference on. New York, US: IEEE. 10.1109/BHI.2017.7897216 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 16. Raza SEA, Cheung L, Shaban M, Graham S, Epstein D, Pelengaris S, et al. Micro-Net: A unified model for segmentation of various objects in microscopy images. Med Image Anal (2019) 52:160–73. 10.1016/j.media.2018.12.003 [DOI] [PubMed] [Google Scholar]
- 17. Komura D, Ishikawa S. Machine learning methods for histopathological image analysis. Comput Struct Biotechnol J (2018) 16:34–42. 10.1016/j.csbj.2018.01.001 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 18. Humphrey PA, Moch H, Cubilla AL, Ulbright TM, Reuter VE. The 2016 WHO classification of tumors of the urinary system and male genital organs—part B: prostate and bladder tumors. Eur Urol (2016) 70(1):106–19. 10.1016/j.eururo.2016.02.028 [DOI] [PubMed] [Google Scholar]
- 19. Rakha EA, Reis-Filho JS, Baehner F, Dabbs DJ, Decker T, Eusebi V, et al. Breast cancer prognostic classification in the molecular era: the role of histological grade. Breast Cancer Res (2010) 12(4):207. 10.1186/bcr2607 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 20. Ma W, Wang J, Yu L, Zhang X, Wang Z, Tan B, et al. Tumor-stroma ratio is an independent predictor for survival in esophageal squamous cell carcinoma. J Thorac Oncol (2012) 7(9):1457–61. 10.1097/JTO.0b013e318260dfe8 [DOI] [PubMed] [Google Scholar]
- 21. Chen Y, Zhang L, Liu W, Liu X. Prognostic significance of the tumor-stroma ratio in epithelial ovarian cancer. BioMed Res Int (2015) 2015. 10.1155/2015/589301 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 22. Ruan M, Tian T, Rao J, Xu X, Yu B, Yang W, et al. Predictive value of tumor-infiltrating lymphocytes to pathological complete response in neoadjuvant treated triple-negative breast cancers. Diagn Pathol (2018) 13(1):66. 10.1186/s13000-018-0743-7 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 23. Barnes TA, Amir E. HYPE or HOPE: the prognostic value of infiltrating immune cells in cancer. Br J Cancer (2017) 117(4):451–60. 10.1038/bjc.2017.220 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 24. Renshaw AA, Cheville JC. Quantitative tumor necrosis is an independent predictor of overall survival in clear cell renal cell carcinoma. Pathology (2015) 47(1):34–7. 10.1097/PAT.0000000000000193 [DOI] [PubMed] [Google Scholar]
- 25. Pichler M, Hutterer GC, Chromecki TF, Jesche J, Kampel-Kettner K, Rehak P, et al. Histologic tumor necrosis is an independent prognostic indicator for clear cell and papillary renal cell carcinoma. Am J Clin Pathol (2012) 137(2):283–9. 10.1309/AJCPLBK9L9KDYQZP [DOI] [PubMed] [Google Scholar]
- 26. Bredholt G, Mannelqvist M, Stefansson IM, Birkeland E, Bø TH, Øyan AM, et al. Tumor necrosis is an important hallmark of aggressive endometrial cancer and associates with hypoxia, angiogenesis and inflammation responses. Oncotarget (2015) 6(37):39676. 10.18632/oncotarget.5344 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 27. Ronneberger O, Fischer P, Brox T. eds. “U-net: Convolutional networks for biomedical image segmentation”. In: International Conference on Medical image computing and computer-assisted intervention. New York, US: Springer. [Google Scholar]
- 28. Wang D, Khosla A, Gargeya R, Irshad H, Beck AH. “Deep learning for identifying metastatic breast cancer”. arXiv preprint (2016) arXiv:160605718. [Google Scholar]
- 29. Nahid A-A, Mehrabi MA, Kong Y. Histopathological breast cancer image classification by deep neural network techniques guided by local clustering. BioMed Res Int (2018) 2018. 10.1155/2018/2362108 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 30. Bándi P, van de Loo R, Intezar M, Geijs D, Ciompi F, van Ginneken B, et al. eds. “Comparison of different methods for tissue segmentation in histopathological whole-slide images”. In: 2017 IEEE 14th International Symposium on Biomedical Imaging (ISBI 2017). New York, US: IEEE; (2017). 10.1109/ISBI.2017.7950590 [DOI] [Google Scholar]
- 31. Bejnordi BE, Zuidhof G, Balkenhol M, Hermsen M, Bult P, van Ginneken B, et al. Context-aware stacked convolutional neural networks for classification of breast carcinomas in whole-slide histopathology images. J Med Imaging (2017) 4(4):044504. 10.1117/1.JMI.4.4.044504 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 32. Vu QD, Graham S, Kurc T, To MNN, Shaban M, Qaiser T, et al. Methods for segmentation and classification of digital microscopy tissue images. Front Bioeng Biotechnol (2019) 7:53. 10.3389/fbioe.2019.00053 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 33. Araújo T, Aresta G, Castro E, Rouco J, Aguiar P, Eloy C, et al. Classification of breast cancer histology images using convolutional neural networks. PLoS One (2017) 12(6):e0177544. 10.1371/journal.pone.0177544 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 34. Wetteland R, Engan K, Eftestøl T, Kvikstad V, Janssen EA. eds. “Multiclass tissue classification of whole-slide histological images using convolutional neural networks”. In: Proceedings of the 8th International Conference on Pattern Recognition Applications and Methods. New York, US: Springer; (2019). [Google Scholar]
- 35. Xu Z, Moro CF, Kuznyecov D, Bozóky B, Dong L, Zhang Q. eds. “Tissue Region Growing for Hispathology Image Segmentation”. In: Proceedings of the 2018 3rd International Conference on Biomedical Imaging, Signal Processing. New York, US: Association for Computing Machinery; (2018). 10.1145/3288200.3288213 [DOI] [Google Scholar]
- 36. Chan L, Hosseini MS, Rowsell C, Plataniotis KN, Damaskinos S. eds. “Histosegnet: Semantic segmentation of histological tissue type in whole slide images”. In: Proceedings of the IEEE International Conference on Computer Vision. New York, US: IEEE; (2019). [Google Scholar]
- 37. Chen L-C, Papandreou G, Kokkinos I, Murphy K, Yuille AL. Deeplab: Semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected crfs. IEEE Trans Pattern Anal Mach Intelligence (2018) 40(4):834–48. 10.1109/TPAMI.2017.2699184 [DOI] [PubMed] [Google Scholar]
- 38. Xu Y, Jia Z, Wang L-B, Ai Y, Zhang F, Lai M, et al. Large scale tissue histopathology image classification, segmentation, and visualization via deep convolutional activation features. BMC Bioinf (2017) 18(1):281. 10.1186/s12859-017-1685-x [DOI] [PMC free article] [PubMed] [Google Scholar]
- 39. Krizhevsky A, Sutskever I, Hinton GE. eds. “Imagenet classification with deep convolutional neural networks”. In: Adv Neural Inf Process Syst. Cambridge, Massachusetts, US: MIT Press; (2012). [Google Scholar]
- 40. Song Y, Zhang L, Chen S, Ni D, Lei B, Wang T. Accurate segmentation of cervical cytoplasm and nuclei based on multiscale convolutional network and graph partitioning. IEEE Trans BioMed Eng (2015) 62(10):2421–33. 10.1109/TBME.2015.2430895 [DOI] [PubMed] [Google Scholar]
- 41. Romo D, García-Arteaga JD, Arbeláez P, Romero E. eds. “A discriminant multi-scale histopathology descriptor using dictionary learning”. In: Medical Imaging 2014: Digital Pathology; 2014: International Society for Optics and Photonics. Bellingham, Washington USA: SPIE; (2014). 10.1117/12.2043935 [DOI] [Google Scholar]
- 42. Qin P, Chen J, Zeng J, Chai R, Wang L. Large-scale tissue histopathology image segmentation based on feature pyramid. EURASIP J Image Video Processing (2018) 2018(1):75. 10.1186/s13640-018-0320-8 [DOI] [Google Scholar]
- 43. Xu Z, Zhang Q. eds. “Multi-scale context-aware networks for quantitative assessment of colorectal liver metastases”. In: 2018 IEEE EMBS International Conference on Biomedical & Health Informatics (BHI). New York, US: IEEE; (2018). 10.1109/BHI.2018.8333445 [DOI] [Google Scholar]
- 44. Beck AH, Sangoi AR, Leung S, Marinelli RJ, Nielsen TO, Van De Vijver MJ, et al. Systematic analysis of breast cancer morphology uncovers stromal features associated with survival. Sci Transl Med (2011) 3(108):108ra13–ra13. 10.1126/scitranslmed.3002564 [DOI] [PubMed] [Google Scholar]
- 45. Xu J, Luo X, Wang G, Gilmore H, Madabhushi A. A deep convolutional neural network for segmenting and classifying epithelial and stromal regions in histopathological images. Neurocomputing (2016) 191:214–23. 10.1016/j.neucom.2016.01.034 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 46. Bejnordi BE, Litjens G, Hermsen M, Karssemeijer N, van der Laak JA. eds. “A multi-scale superpixel classification approach to the detection of regions of interest in whole slide histopathology images”. In: Medical Imaging 2015: Digital Pathology; 2015: International Society for Optics and Photonics. Bellingham, Washington USA: SPIE; (2015). 10.1117/12.2081768 [DOI] [Google Scholar]
- 47. Achanta R, Shaji A, Smith K, Lucchi A, Fua P, Süsstrunk S. SLIC superpixels compared to state-of-the-art superpixel methods. IEEE Trans Pattern Anal Mach Intelligence (2012) 34(11):2274–82. 10.1109/TPAMI.2012.120 [DOI] [PubMed] [Google Scholar]
- 48. Brockmann M, Poon E, Berry T, Carstensen A, Deubzer HE, Rycak L, et al. Small molecule inhibitors of aurora-a induce proteasomal degradation of N-myc in childhood neuroblastoma. Cancer Cell (2013) 24(1):75–89. 10.1016/j.ccr.2013.05.005 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 49. Berry T, Luther W, Bhatnagar N, Jamin Y, Poon E, Sanda T, et al. The ALK(F1174L) mutation potentiates the oncogenic activity of MYCN in neuroblastoma. Cancer Cell (2012) 22(1):117–30. 10.1016/j.ccr.2012.06.001 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 50. Moore HC, Wood KM, Jackson MS, Lastowska MA, Hall D, Imrie H, et al. Histological profile of tumors from MYCN transgenic mice. J Clin Pathol (2008) 61(10):1098–103. 10.1136/jcp.2007.054627 [DOI] [PubMed] [Google Scholar]
- 51. Jamin Y, Glass L, Hallsworth A, George R, Koh DM, Pearson AD, et al. Intrinsic susceptibility MRI identifies tumors with ALKF1174L mutation in genetically-engineered murine models of high-risk neuroblastoma. PLoS One (2014) 9(3):e92886. 10.1371/journal.pone.0092886 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 52. Reinhard E, Adhikhmin M, Gooch B, Shirley P. Color transfer between images. IEEE Comput Graphics Applications (2001) 21(5):34–41. 10.1109/38.946629 [DOI] [Google Scholar]
- 53. Zormpas-Petridis K, Failmezger H, Raza SEA, Roxanis I, Jamin Y, Yuan Y. Superpixel-based Conditional Random Fields (SuperCRF): Incorporating global and local context for enhanced deep learning in melanoma histopathology. Front Oncol (2019) 9:1045. 10.3389/fonc.2019.01045 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 54. Srivastava N, Hinton G, Krizhevsky A, Sutskever I, Salakhutdinov R. Dropout: a simple way to prevent neural networks from overfitting. J Mach Learn Res (2014) 15(1):1929–58. 10.5555/2627435.2670313 [DOI] [Google Scholar]
- 55. Glorot X, Bengio Y. eds. “Understanding the difficulty of training deep feedforward neural networks”. In: Proceedings of the thirteenth international conference on artificial intelligence and statistics. Proceedings of Machine Learning Research (PMLR) (2010). [Google Scholar]
- 56. Kingma DP, Ba J. Adam: A method for stochastic optimization. arXiv preprint (2014) arXiv:14126980. [Google Scholar]
- 57. Szegedy C, Vanhoucke V, Ioffe S, Shlens J, Wojna Z. eds. “Rethinking the inception architecture for computer vision”. In: Proceedings of the IEEE conference on computer vision and pattern recognition. New York, US: IEEE; (2015). [Google Scholar]
- 58. Chollet F. ed. “Xception: Deep learning with depthwise separable convolutions”. In: Proceedings of the IEEE conference on computer vision and pattern recognition. New York, US: IEEE; (2016). [Google Scholar]
- 59. Szegedy C, Ioffe S, Vanhoucke V, Alemi AA. eds. “Inception-v4, inception-resnet and the impact of residual connections on learning”. In: Thirty-First AAAI Conference on Artificial Intelligence. California, US: AAAI; (2016). [Google Scholar]
- 60. He K, Zhang X, Ren S, Sun J. eds. “Deep residual learning for image recognition”. In: Proceedings of the IEEE conference on computer vision and pattern recognition. New York, US: IEEE; (2015). [Google Scholar]
- 61. Heindl A, Sestak I, Naidoo K, Cuzick J, Dowsett M, Yuan Y. Relevance of spatial heterogeneity of immune infiltration for predicting risk of recurrence after endocrine therapy of ER+ breast cancer. JNCI: J Natl Cancer Institute (2018) 110(2):166–75. 10.1093/jnci/djx137 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 62. Nawaz S, Heindl A, Koelble K, Yuan Y. Beyond immune density: critical role of spatial heterogeneity in estrogen receptor-negative breast cancer. Mod Pathol (2015) 28(6):766. 10.1038/modpathol.2015.37 [DOI] [PubMed] [Google Scholar]
- 63. Gilchrist KW, Gray R, Fowble B, Tormey DC, Taylor 4th S. Tumor necrosis is a prognostic predictor for early recurrence and death in lymph node-positive breast cancer: a 10-year follow-up study of 728 Eastern Cooperative Oncology Group patients. J Clin Oncol (1993) 11(10):1929–35. 10.1200/JCO.1993.11.10.1929 [DOI] [PubMed] [Google Scholar]
- 64. Hanafy E, Al Jabri A, Gadelkarim G, Dasaq A, Nazim F, Al Pakrah M. Tumor histopathological response to neoadjuvant chemotherapy in childhood solid malignancies: is it still impressive? J Invest Med (2018) 66(2):289–97. 10.1136/jim-2017-000531 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 65. Lambertz I, Kumps C, Claeys S, Lindner S, Beckers A, Janssens E, et al. Upregulation of MAPK Negative Feedback Regulators and RET in Mutant ALK Neuroblastoma: Implications for Targeted Treatment. Clin Cancer Res (2015) 21(14):3327–39. 10.1158/1078-0432.CCR-14-2024 [DOI] [PubMed] [Google Scholar]
- 66. Gowda SN, Yuan C. eds. “ColorNet: Investigating the importance of color spaces for image classification”. In: Asian Conference on Computer Vision. New York, US: Springer. [Google Scholar]
- 67. Yao J, Boben M, Fidler S, Urtasun R. eds. “Real-time coarse-to-fine topologically preserving segmentation”. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. New York, US: IEEE; (2015). [Google Scholar]
- 68. Tan M, Le QV. Efficientnet: Rethinking model scaling for convolutional neural networks. arXiv preprint (2019) arXiv:190511946. [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Supplementary Materials
Data Availability Statement
The melanoma dataset comes the publicly available TCGA dataset. The neuroblastoma dataset is available from the corresponding authors upon reasonable request. The images from the triple-negative breast cancer dataset cannot be released yet due to ongoing clinical studies. The codes that support the findings of this study are available from the corresponding authors upon reasonable request.