Abstract
Background and Objective:
Gleason grading system is currently the clinical gold standard for determining prostate cancer aggressiveness. Prostate cancer is typically classified into one of 5 different categories with 1 representing the most indolent disease and 5 reflecting the most aggressive disease. Grades 3 and 4 are the most common and difficult patterns to be discriminated in clinical practice. Even though the degree of gland differentiation is the strongest determinant of Gleason grade, manual grading is subjective and is hampered by substantial inter-reader disagreement, especially with regard to intermediate grade groups.
Methods:
To capture the topological characteristics and the degree of connectivity between nuclei around the gland, the concept of Homology Profile (HP) for prostate cancer grading is presented in this paper. HP is an algebraic tool, whereby, certain algebraic invariants are computed based on the structure of a topological space. We utilized the Statistical Representation of Homology Profile (SRHP) features to quantify the extent of glandular differentiation. The quantitative characteristics which represent the image patch are fed into a supervised classifier model for discrimination of grade patterns 3 and 4.
Results:
On the basis of the novel homology profile, we evaluated 43 digitized images of prostate biopsy slides annotated for regions corresponding to Grades 3 and 4. The quantitative patch-level evaluation results showed that our approach achieved an Area Under Curve (AUC) of 0.96 and an accuracy of 0.89 in terms of discriminating Grade 3 and 4 patches. Our approach was found to be superior to comparative methods including handcrafted cellular features, Stacked Sparse Autoencoder (SSAE) algorithm and end-to-end supervised learning method (DLGg). Also, slide-level quantitative and qualitative evaluation results reflect the ability of our approach in discriminating Gleason Grade 3 from 4 patterns on H&E tissue images.
Conclusions:
We presented a novel Statistical Representation of Homology Profile (SRHP) approach for automated Gleason grading on prostate biopsy slides. The most discriminating topological descriptions of cancerous regions for grade 3 and 4 in prostate cancer were identified. Moreover, these characteristics of homology profile are interpretable, visually meaningful and highly consistent with the rubric employed by pathologists for the task of Gleason grading.
Keywords: Prostate cancer, Gleason grading, Digitized needle biopsy samples, Homology Profile, Statistical representation
1. Introduction
Prostate cancer is the second leading cause of cancer deaths among males, accounting for 7.1% and 3.8% for cancer incidence and mortality, respectively [1]. The Gleason grading system [2] is the most widely used grading scheme for diagnosis and risk stratification of prostate cancer. The Gleason grade is the strongest marker of disease aggressiveness and longer term outcome for prostate cancer. It is significant in that it can help clinicians make appropriate treatment management decisions for prostate cancer patients, i.e. help decide between more aggressive therapy versus active surveillance. In the Gleason grading schema, the regions in the tissue slides are assigned pathological grades from 1 to 5 (worst) based off the observed gland differentiation patterns. Typically the dominant and the secondary most aggressive Gleason patterns are identified and combined to provide the Gleason score. Note that while this can generate scores from 2 to 10, the newest ISUP scheme categorizes the different Gleason score tumors into 5 categories corresponding to the most commonly occurring Gleason scores (3+3, 3+4, 4+3, 4+4 and > 4+4) [3]. The most challenging set of categories to distinguish are grade 3 ( G3 ) and grade 4 ( G4 ). This is a task that is time-consuming and typically has the highest degree of inter-reader variability among pathologists. There is thus a need for developing computerized image analysis systems for distinguishing G3 and G4.
In recent years, with the rapid development of digital pathology [4], many computational pathology algorithms have been presented for studying digitized histological slides and aiding in problems relating to disease diagnosis, prognosis and prediction of treatment response [5–8]. Bera et al. [8] provided an excellent review of the various categories of feature based approaches in computational pathology, including hand-crafted and deep learning based approaches. In the context of prostate cancer, many approaches have been proposed for both prostate cancer diagnosis and grading [9–15]. Doyle et al. [16] identified cancerous regions on biopsy slides by a boosted Bayesian multi-resolution system. Monaco and Madabhushi [17] used a Markov-random field (MRF) based approach with a class specific weighting approach to identify regions of cancer on surgically resected specimens. In the field of deep learning, Li et al. [11] presented a region-based convolutional neural network framework for simultaneous multi-task epithelial cell detection and Gleason grading. Poojitha and Sharma [15] presented a hybrid architecture with unified deep learning techniques to perform Gleason grading. Two papers in The Lancet Oncology by Ström et al. [18], Bulten et al. [19] used deep learning methods for detection and Gleason grading of prostate cancer in digital images of biopsies.
Although deep learning based approaches have become very popular in the context of computational pathology [14,20], the black-box nature of these approaches raises questions about interpretability and clinical adoption [21]. Handcrafted features, however, usually leverage intrinsic domain knowledge of pathologists [8]. For pathologists, the degree of glandular differentiation is critical for determining the grade of prostate cancer (see Fig. 1). Mosquera-Lopez et al. [9] in their paper discussed multiple different textural based approaches for detection and grading of prostate cancer. Ali et al. [22] discriminated different grade patterns by nuclear shape-based, architectural and textural features. Farooq et al. [10] presented a computer aided system by combining Gabor and local binary pattern (LBP) based texture features for the grading of prostate cancer. Leo et al. [23] studied stability and discriminability for a wide spectrum of gland texture and shape features for Gleason grading in prostatectomy specimens. Lee et al. [24] presented a co-occurring gland angularity approach to model the extent of gland architectural disorder for the problem of discriminating benign and malignant glands on prostate tissue sections. The features were also shown to be prognostic of the risk of biochemical recurrence on post-prostatectomy specimens. Niazi et al. [25] evaluated luminal and architectural features to discriminate low-grade (Gleason score ≤ 6) and high-grade (Gleason score ≥ 8) prostate cancers. Interestingly, a number of studies have shown that the Gleason grading system is based on architectural patterns [26,27], which suggests that the topological arrangement of the nuclei with respect to the individual glands could harbor grade discriminating information. None of the approaches presented to date, that we are aware of, have explicitly looked at the interplay of cellular and glandular features for the problem of prostate cancer grading.
Fig. 1.

An example of a digital whole slide image with pathologists’ annotations and its samples, which illustrate the degree of gland differentiation among Gleason G3 and G4. Left and right are samples of G3 and G4 with 160 × 160 pixel at 10x magnification, respectively.
The Homology Profile (HP) algorithm is an algebraic tool for measuring topological features of objects [28]. Given a topological space, the HP algorithm computes numbers of connected components and holes using the structure of that space based on continuous thresholds. These topological features represent the sustained variance in arrangement of objects such as nuclei of glands in different grades of prostate cancer. Nakane et al. [42,43] Recently, Qaiser et al. [29–31] employed the HP algorithm for tumor segmentation by focusing on the connectivity between nuclei.
In this work, we present novel Statistical Representation of Homology Profile (SRHP) features for automated discrimination of Gleason grade patterns G3 and G4 from digital pathological tissue images. We employed the HP features to describe topological arrangements of the nuclei with respect to the gland lumen and quantify the degree of gland differentiation in needle biopsy slides of prostate cancer. A supervised machine classifier algorithm was employed in conjunction with the HP features to evaluate the discrimination of the G3 and G4 tissue pathology images. In this study, 43 whole slide images of prostate biopsy slides were chosen and we obtained 9597 patches for G3 and 8679 patches for G4 grade regions from the slides. A 10-fold cross validation scheme across 18,276 was then utilized for evaluation.
2. Materials and methods
Fig. 2 illustrates the overall flowchart of our approach. The flowchart delineates the image processing, HP feature extraction and classifier modeling modules.
Fig. 2.

(A) The flowchart for our approach for discrimination of G3 and G4 prostate cancerous regions. Hematoxylin staining component of the patch is leveraged for HP computation and statistical representation, followed by classifier modeling. (B) The procedure of HP computation. An image is converted to binary images for computing Betti numbers [29], which form variable distributions.
2.1. Data set and images preprocessing
43 digitized images of prostate biopsy slides stained with Hematoxylin&Eosin (H&E) were used in this study. All these slides including regions of interest with G3 and G4 were obtained from UPenn (the Hospital at the University of Pennsylvania). All slides were digitized at 40x magnification and manually annotated by expert pathologists. From the regions of interest annotated by pathologists, image patches of sizes 640 × 640 pixels at 40x magnification were extracted. The grades of these patches were determined according to pathologist based annotation of the regions they were derived from. Since the focus was on evaluating features relating to the topology and connectivity among nuclei rather than features relating to individual nuclei, the patch size was 160 × 160 pixels at 10x magnification. The window slides across each region of interest image row by row from upper left corner to lower right (the step size was fixed at 80 pixels). Then the overlapping image patches of 160 × 160 pixels were employed for feature extraction. We thus obtained 9597 patches for G3 and 8679 patches for G4 grade regions from a total of 43 biopsy slides.
H&E staining is the most commonly used method for paraffin sections. The hematoxylin staining makes the chromatin in the nucleus and ribosome in cytoplasm purple-blue and eosin makes the components of cytoplasm and extracellular matrix red. We utilized color deconvolution [32] for separation of the hematoxylin stain. However, on account of high variability of color stain, the fixed stain matrix adopted in Ruifrok et al. [32] is not appropriate. Since we needed an adaptive approach, we employed the color stain estimation approach proposed in Macenko et al. [33] for computing the stain matrix. The estimation method yielded a unique matrix for each image patch, representing the stain matrix for hematoxylin separation. The corresponding stain matrices are shown as follows:
| (1) |
| (2) |
where (1) is the original fixed Ruifrok stain matrix [32] and (2) is the average stain matrix obtained via the adaptive estimation method [33]. The first and the second rows in the matrices (1) and (2) refer to hematoxylin and eosin stains, respectively. Separation of hematoxylin stain channel can emphasize the precise location of the individual nuclei and facilitate the computation of the HP features.
2.2. Statistics representation of homology profile (SRHP)
The HP algorithm provides an algebraic tool for measuring and quantifying geometric and topological features of objects and shapes in the topological space [28]. For a two-dimensional image, the HP method involves the calculation of the number of connected components and voids, denoted as b0 and b1 to represent the topological features. In this work we employ b0 and b1 for measuring the degree of connectivity among the nuclei around glands.
Prior to feature extraction, the H stained image is binarized from to by the application of threshold t, replacing any value in X more than t by 1 and all other values by 0. The resulting image can be thought of as filtered representations of the original image X. Betti numbers [29] b0 and b1 are then computed based off connected components in and , respectively, where is the complement of in and is the binarized image from X with a threshold 255. As t gradually changes with respect to the value from 0 to K, the filtered representations change dynamically. This makes b0 and b1 values increase or decrease. Fig. 3 shows the process of image evolution and the calculation of Betti numbers. In this work, the threshold t range was set from 1 to 255 with an increment of 1 for HP computation. We thus obtain a 510-dimensional (255 × 2) low-level Betti sequence vector which represents the image patch.
Fig. 3.

The procedure of binary image evolution and calculation of the Betti numbers. The H staining component image is binarized using different threshold values (40, 100 and 150). Connected components b0 and white voids b1 are indicated via blue dots and red stars, respectively.
There are many methods to characterize and evaluate the 510-dimensional feature vector corresponding to an image patch. One approach is to measure the distance between the testing patch and an exemplar patch identified via a Convolutional Neural Network using the symmetric Kullback-Leibler Divergence [29]. However, this approach tends to rely heavily on the choice of exemplars. On account of the significant morphologic heterogeneity within individual grades, let alone between G3 and G4 [34], identifying exemplars to represent G3 and G4 is non-trivial. We utilized the classical statistical method [35] to re-characterize the 510-dimensional feature vector and use it to discover the regularity of data.
The two 255 successive Betti numbers can be viewed as 2 discrete probability distributions Y0 , Y1. Y0, Y1 can be viewed in terms of (1) the centralized trend of distribution, (2) the degree of dispersion and (3) the shape [36]. The central trend is a typical description of a probability distribution. The most common measures of central tendency are the mean, median and mode. Dispersion is the extent to which a distribution is stretched or squeezed. Examples of dispersion measures include standard deviation, range, mean absolute deviation, and coefficient of variation. Moreover, quantitative measures such as skewness and kurtosis are descriptive of the shape of a distribution. We utilized the mean and median statistics for describing central tendency and, standard deviation, range and coefficient of variation to represent the dispersion of distributions Y0, Y1. Consequently, a total of 7 measures of statistics are employed for Y0, Y1 and ratio of Y1 to Y0. The 510-dimensional feature of each image patch were expressed in terms of 21 different statistical measures.
2.3. Classifier modeling
A machine learning classifier was built to evaluate the SRHP features. We utilized the weighted k nearest neighbor (wKNN) classifier with an exhaustive search algorithm and the Euclidean distance metric for sample grading and feature modeling. The wKNN algorithm assigns weight vectors for the K nearest neighbors. The jth weight value for the ith sample is defined as:
| (3) |
where is the nth feature value of the ith sample for testing. denotes the nth dimension feature value of the jth sample amongst the K nearest neighbors from the training set. N represents the number of features, which is 21 in this study. Prior to the application of wKNN, the features are normalized by the mean and standard deviation of samples in the training set.
3. Experimental designs and comparative strategies
In order to demonstrate the effectiveness of our SRHP approach for discriminating G3 and G4 image patterns, the experiments described below were conducted.
3.1. Evaluation of SRHP approach
The aim of this experiment was to evaluate the effectiveness of our SRHP approach for discriminating G3 and G4 patterns at both the image patch- and slide-levels, respectively. We adopted a 10-fold cross validation scheme involving 18,276 representative patches from 43 biopsy slides for evaluation.
In these experiments, we compared the performance of the SRHP approach with other extant state-of-the-art methods [12,22,37]. A brief description of these approaches is summarized below:
End-to-end supervised learning method, MobileNet [38], which Arvaniti et al. [12] adopted for 4-class Gleason grading (benign, grade 3, 4 and 5) directly on prostate cancer tissue micro-array images. We abbreviate this Deep Learning for Gleason Grading approach as DLGg. It is worth noting that we employed the pre-trained model on our dataset.
End-to-end unsupervised learning method, stacked sparse autoencoder algorithm (SSAE) presented by Xu et al. [37]. The method was used for nuclei detection on pathological images. The stacked sparse autoencoder was employed for classifying individual candidate locations as nuclei or not via a sliding window scheme.
Traditional handcrafted features including nuclear morphological, architectural and textural features, denoted collectively as MATF. These features were employed by Ali et al. [22] for distinguishing G3 and G4 patterns of prostate cancer.
The values of hyper-parameters were set based off the specifications in the original publications where the methods were first introduced.
3.2. Statistical features analysis
The aim of this experiment was to analyze the effectiveness of all 21 statistical features for discriminating G3 and G4 patterns. We also compared different feature permutation schemes in terms of connected components (b0 ), voids (b1 ) and the ratio of b1 to b0 (denoted as b ).
3.3. Sensitivity analysis
The aim of this experiment was to evaluate the sensitivity of our model to the H stain separation scheme and hyper-parameters K values. Specifically we sought to evaluate (1)how the HP features extracted from the H channel alone compared to the features extracted from the original RGB images and also to assess (2) how the classifier performed across K different values.
3.4. Performance evaluation
The performance of our SRHP approach and comparative strategies were evaluated in terms of Area Under Curve value (AUC), Accuracy (ACC), Recall (REC), Precision (PRC), Specificity (SPC) and F1 score [37]. We employed a 10-fold cross validation scheme and reported the average values for the performance metrics on the models.
4. Results
4.1. Evaluation of SRHP at the image patch-level
We compared the performance of our approach and the comparative strategies listed in Table 1 in terms of AUC, Accuracy, Recall, Precision, Specificity, and F1 score measurements. The reported performance measures were averaged across 10-fold cross validation and the results are reported in Table 1. The results show that the SRHP approach outperforms three other comparative methods and achieves the highest AUC of 0.96, accuracy of 89.02% and an F1 score of 0.89. The results suggest that the SRHP approach is more accurate at differentiating G3 and G4 tissue as compared to traditional handcrafted cellular features. Also, our approach outperforms the end-to-end DLGg and SSAE methods. This further reinforces the effectiveness of our SRHP approach for distinguishing G3 and G4 grades.
Table 1.
Quantitative evaluation results of our SRHP algorithm against comparative methods at the patch-level in terms of AUC, Accuracy, Recall, Precision, Specificity, and F1 score. The results are averaged across multiple runs of 10 fold cross validation (The best results are shown in bold).
| Patch-level | AUC | Accuracy | Recall | Precision | Specificity | F1 Score |
|---|---|---|---|---|---|---|
| DLGg | 0.91 | 85.04% | 0.72 | 0.90 | 0.94 | 0.80 |
| SSAE | 0.79 | 72.07% | 0.65 | 0.73 | 0.79 | 0.69 |
| MATF | 0.94 | 86.41% | 0.84 | 0.87 | 0.89 | 0.85 |
| SRHP | 0.96 | 89.02% | 0.94 | 0.84 | 0.84 | 0.89 |
4.2. Evaluation of SRHP at the image slide-level
Table 2 illustrates the results of quantitative evaluation of SRHP and comparative strategies at the slide-level in terms of AUC, Accuracy, Recall, Precision, Specificity, and F1 score. The SRHP approach is evaluated on a per slide basis. Noting that only 20 slides include both G3 and G4 regions, the evaluation results reported were averaged across these 20 slides. We can see from Table 2 that our SRHP still achieves better performance compared to the other methods at the slide-level in terms of all performance metrics. This again reinforces the effectiveness of our HP features. Standard deviation in the performance metrics across slides are also reported. The relatively small standard deviation in the performance metrics appears to suggest robustness of our approach across different slide images.
Table 2.
Results of quantitative evaluation of features robustness of our SRHP algorithm against comparative methods at the slide-level in terms of AUC, Accuracy, Recall, Precision, Specificity, and F1 score. The results are averaged across 20 slides, which includes both G3 and G4 regions on the same slide. We also report standard deviations in performance metrics across slides (The best results are indicated in bold).
| Slide-level | AUC | Accuracy | Recall | Precision | Specificity | F1 Score |
|---|---|---|---|---|---|---|
| DLGg | 0.93 ± 0.13 | 89.35% ± 0.14 | 0.80 ± 0.25 | 0.82 ± 0.20 | 0.91 ± 0.13 | 0.78 ± 0.21 |
| SSAE | 0.62 ± 0.18 | 71.31% ± 0.12 | 0.42 ± 0.30 | 0.48 ± 0.36 | 0.73 ± 0.21 | 0.53 ± 0.25 |
| MATF | 0.94 ± 0.01 | 86.45% ± 0.02 | 0.84 ± 0.03 | 0.87 ± 0.05 | 0.89 ± 0.03 | 0.85 ± 0.03 |
| SRHP | 0.99 ± 0.01 | 98.75% ± 0.01 | 0.97 ± 0.03 | 0.93 ± 0.12 | 0.98 ± 0.02 | 0.95 ± 0.09 |
Fig. 4 shows the scatter plot for each slide in terms of AUC and accuracy for the 4 different methods considered in our experimental design. The performance metrics for each method is shown in a different color. Each point represents the results on a per slide basis, and the DLGg, SSAE, MATF and the SRHP approaches are represented via the blue, black, green and red colors, respectively. It is clear that our SRHP method achieves more accurate and robust results compared to the other approaches, note that the red elements aggregate towards the extreme north-east corner, reflecting an accuracy and AUC of almost 100%. It not only reinforces the effectiveness of HP features, but also illustrates the robustness of our approach.
Fig. 4.

Scatter plot of each slide in terms of AUC and accuracy measurements across different approaches. Each point represents the evaluation result of a slide, where blue, black, green and red colors represent DLGg, SSAE, MATF, and the SRHP methods, respectively. (For interpretation of the references to color in this figure legend, the reader is referred to the web version of this article.)
4.3. Statistical features analysis
Box-plots of statistical features are illustrated in Fig. 5(a), and the feature indices corresponding to b (b1/b0), b0 and b1 are shown in Fig. 5(b). The box-plots of 21 features illustrate their strong discriminability for the grading task.
Fig. 5.

(a) Box-plots of all 21 statistical features for the G3 and G4 grading task. Each box represents one HP feature which has been normalized. (b) Illustration of the feature indices corresponding to statistical features in (a) in terms of b (b1/b0), b0 and b1 .
Fig. 6 shows results of different combinations of features in terms of connected components (b0 ), voids ( b1 ) and the ratio of b1 to b0 (b). It is apparent that the combination of all features achieves the best result averaged across 10-fold cross validation, reinforcing the effectiveness of the HP features. We can also see that the b1 statistic performs better compared to the b0 features. This trend is in concordance with the clinical grading criteria that the degree of gland differentiation plays a more critical role in Gleason grading, compared to variation of the number of nuclei. The results suggest that HP features can quantitatively capture the topological arrangement and degree of differentiation around prostate cancer glands.
Fig. 6.

Results of quantitative evaluation of different feature combinations in terms of connected components (b0) , voids (b1), the ratio of b1 to b0 (b), and all.
4.4. Sensitivity analysis
The results of sensitivity analysis of our SRHP approach as a function of stain separation and the K value hyper-parameter are illustrated in Table 3 and Fig. 7, respectively. From Table 3 we can observe that the features perform better on the H stain compared to the original RGB image.
Table 3.
Quantitative evaluation results of sensitivity analysis for our SRHP approach in terms of H stain separation as compared to that of original RGB images. The results of RGB and H stain show the performances based on HP features extracted from raw RGB images and hematoxylin channels, respectively.
| Channel | AUC | Accuracy | Recall | Precision | Specificity | F1 Score |
|---|---|---|---|---|---|---|
| RGB | 0.9479 | 87.33% | 0.9011 | 0.8443 | 0.8478 | 0.8718 |
| H stain | 0.9612 | 89.02% | 0.9442 | 0.8435 | 0.8413 | 0.8910 |
Fig. 7.

Sensitivity analysis of the SRHP approach across different hyper-parameter K values as assessed in terms of AUC, Accuracy, Recall, Precision, Specificity and F1 Score of a wKNN classifier, where K ranges from 3 to 20.
Fig. 7 shows performance results for the HP features as a function of the hyper-parameter K in conjunction with a wKNN classifier. We can observe from these results that the model yields stable performance across a spectrum of K values (3 to 20). This in turn appears to suggest that our SRHP algorithm is relatively robust to the spectrum of hyper-parameter K values.
5. Discussion
Although our SRHP approach achieved excellent results in discriminating G3 and G4 pattern tissue biopsy images, the representative regions of interest were manually selected via an expert’s annotations of biopsy slides. To demonstrate the feasability of automated grading on the prostate whole biopsy slide, we performed the automated grading on a slide-level.
Firstly, we tiled one whole biopsy slide for detection of the regions of cancer. The patch size was 160 × 160 pixels at 10x magnification. The window slides across the whole biopsy slide row by row from upper left corner to lower right (the step size was fixed at 80 pixels). Then the overlapping image patches of 160 × 160 pixels were employed for feature extraction. We adopted the OTSU algorithm [39] for background filtering and AlexNet [40] for identifying the cancerous patches. This AlexNet classifier was trained explicitly to distinguish benign from malignant image regions. Patches identified as being malignant by the AlexNet classifier were then utilized for discrimination of G3 and G4 image patches via our SRHP algorithm. Eventually, we stitched together the patch-level prediction assignments into a single contiguous whole image, which included background, benign, G3 and G4 regions.
Fig. 8 shows the probability map for regions of cancer on a single biopsy slide, obtained by stitching together the AlexNet prediction results of multiple individual image patches. All malignant regions identified were also qualitatively evaluated by comparing the results with the interpretations of expert pathologists. As can be observed from Fig. 8(b), two region of interests have been magnified to better illustrate the regions of malignancy identified by the classifier. Interestingly, the classifier was also able to identify those regions as malignant that were initially missed by the expert pathologist on the initial inspection of the slide image. We also found that marginal glandular areas with a proliferation of nuclei were more likely to be detected as being cancerous. Fig. 9 illustrates the process of automated discrimination of G3 and G4 image patches using our SRHP approach on the whole slide biopsy images.
Fig. 8.

(a) Original whole slide image where pathologist’s annotations of G3 and G4 patterns are delineated by green and blue contours, respectively. (b) The corresponding probability map generated by the SRHP approach reveals the regions identified as being malignant on whole biopsy slide images. (For interpretation of the references to color in this figure legend, the reader is referred to the web version of this article.)
Fig. 9.

Illustration of SRHP approach for discriminating G3 from G4 patterns on a needle biopsy core image. (a) Original biopsy whole slide image with a pathologist’s annotations of G3 and G4 regions. (b) The prediction results for G3 and G4 regions and the results of the SRHP approach are shown in false-color, where white, blue, green and red colors represent assignments of background, benign, G3 and G4 , respectively. (For interpretation of the references to color in this figure legend, the reader is referred to the web version of this article.)
The confusion matrix of errors for the G3 and G4 grading for our SRHP approach is presented in Table 4, where specificity, sensitivity, and accuracy are also listed. The SRHP approach can identify G4 patterns better as compared to G3 regions, which is consistent with the qualitative results. As illustrated in Fig. 9, our approach fails to identify some G3 regions which are atypical [34].
Table 4.
Confusion matrix of G3 and G4 grading for mis-classification analysis for the SRHP approach, where specificity, sensitivity and accuracy are also reported.
| Confusion Matrix | (Target) G3 | (Target) G4 | - |
|---|---|---|---|
| (Output) G3 | 8117 | 476 | 0.9446 |
| (Output) G4 | 1480 | 8203 | 0.8472 |
| - | (Specificity) 0.8458 | (Sensitivity) 0.9452 | (Accuracy) 0.8930 |
6. Conclusions
In this work, we presented a Statistical Representation of Homology Profile (SRHP) approach to capture the topological arrangement and degree of connectivity among nuclei around gland lumen in prostate cancer. The SRHP approach was evaluated at both patch-level as well as at the slide image level. Quantitative evaluation results show that our SRHP approach performed compellingly compared to other supervised / unsupervised learning algorithms (DLGg, SSAE) and handcrafted cellular features approaches (MATF). Extensive experimental results appear to suggest that the SRHP approach could potentially serve as a decision support aid for discriminating G3 from G4 prostate cancer grade regions. One limitation of our approach was that we have also not evaluated the ability of this approach to deal with co-occurring cribriform patterns. In the updated grading guidelines, these cribriform patterns occupy a significant role [41].
In summary, we presented a new feature representation approach (SRHP: Statistical Representation of Homology Profile) for the task of discrimination of G3 from G4 grade regions on biopsy slides. We presented a new class of computational histological image features which capture the topological arrangement of nuclei with respect to the gland lumen. Thus unlike black box approaches, the new computational pathology features are not just discriminating but also interpretable and intuitive.
Supplementary Material
Acknowledgements
This work was supported by National Natural Science Foundation of China (Nos. U1809205, 61771249, 91959207, 81871352); Natural Science Foundation of Jiangsu Province of China (No. BK20181411); Special Foundation by Jiangsu Collaborative Innovation Center of Atmospheric Environment and Equipment Technology (CICAEET) and Jiangsu Key Laboratory of Big Data Analysis Technology (B-DAT) (No. 2020xtzx005); Qing Lan Project of Jiangsu Province; The National Cancer Institute of the National Institutes of Health under award numbers 1U24CA199374-01, U01CA239055; R01 CA202752-01A1, R01 CA208236-01A1, R01 CA216579-01A1, R01 CA220581-01A1, National Center for Research Resources under award number 1 C06 RR12463-01, Merit Review Award VA IBX004121A from the United States (U.S.) Department of Veterans Affairs Biomedical Laboratory Research and Development Service, the DOD Prostate Cancer Idea Development Award (W81XWH-15-1-0558), the DOD Lung Cancer Investigator-Initiated Translational Research Award (W81XWH-18-1-0440), the DOD Peer Reviewed Cancer Research Program (W81XWH-16-1-0329), the Ohio Third Frontier Technology Validation Fund, the Wallace H. Coulter Foundation Program in the Department of Biomedical Engineering and the Clinical and Translational Science Award Program (CTSA) at Case Western Reserve University. The content is solely the responsibility of the authors and does not necessarily represent the official views of the National Institutes of Health or U.S. Department of Veterans Affairs or the United States Government.
Footnotes
Ethical approval
For the cases considered in this study, Institutional Review Board (IRB) approval had been previously obtained.
Declaration of Competing Interest
A.M. is an equity holder in Elucid Bioimaging and Inspirata Inc. He also serves as an advisory board member for Aiforia Inc.
The authors declared that they have no conflicts of interest to this work. We declare that we all do not have any commercial or associative interest that represents a conflict of interest in connection with the work submitted.
Supplementary material
Supplementary material associated with this article can be found, in the online version, at doi:10.1016/j.cmpb.2020.105528.
References
- [1].Bray F, Ferlay J, Soerjomataram I, Siegel RL, Torre LA, Jemal A, Global cancer statistics 2018: globocan estimates of incidence and mortality worldwide for 36 cancers in 185 countries, CA Cancer J. Clin. 68 (6) (2018) 394–424, doi: 10.3322/caac.21492. [DOI] [PubMed] [Google Scholar]
- [2].Gleason DF, Mellinger GT, Prediction of prognosis for prostatic adenocarcinoma by combined histological grading and clinical staging, J. Urol. 111 (1) (1974) 58–64, doi: 10.1016/S0022-5347(17)59889-4. [DOI] [PubMed] [Google Scholar]
- [3].Epstein JI, Humphrey PA, Amin MB, Reuter, Contemporary Gleason grading of prostatic carcinoma an update with discussion on practical issues to implement the 2014 international society of urological pathology (isup) consensus conference on Gleason grading of prostatic carcinoma, Am. J. Surg. Pathol. (2017). [DOI] [PubMed] [Google Scholar]
- [4].Pantanowitz L, Valenstein PN, Evans AJ, Kaplan KJ, Pfeifer JD, Wilbur DC, Collins LC, Colgan TJ, Review of the current state of whole slide imaging in pathology, J. Pathol. Inform. 2 (2011), doi: 10.4103/2153-3539.83746. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [5].Hamilton PW, Bankhead P, Wang Y, Hutchinson R, Kieran D, McArt DG, James J, Salto-Tellez M, Digital pathology and image analysis in tissue biomarker research, Methods 70 (1) (2014) 59–73, doi: 10.1016/j.ymeth.2014.06.015. [DOI] [PubMed] [Google Scholar]
- [6].Xing F, Yang L, Robust nucleus/cell detection and segmentation in digital pathology and microscopy images: a comprehensive review, IEEE Rev. Biomed. Eng. 9 (2016) 234–263, doi: 10.1109/RBME.2016.2515127. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [7].Bankhead P, Loughrey MB, Fernández JA, Dombrowski Y, McArt DG, Dunne PD, McQuaid S, Gray RT, Murray LJ, Coleman HG, et al. , Qupath: open source software for digital pathology image analysis, Sci. Rep. 7 (1) (2017) 16878, doi: 10.1038/s41598-017-17204-5. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [8].Bera K, Schalper KA, Rimm DL, Velcheti V, Madabhushi A, Artificial intelligence in digital pathology new tools for diagnosis and precision oncology, Nat. Rev. Clin. Oncol. 16 (11) (2019) 703–715, doi: 10.1038/s41571-019-0252-y. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [9].Mosquera-Lopez C, Agaian S, Velez-Hoyos A, Thompson I, Computer-aided prostate cancer diagnosis from digitized histopathology: a review on texture-based systems, IEEE Rev. Biomed. Eng. 8 (2014) 98–113, doi: 10.1109/RBME.2014.2340401. [DOI] [PubMed] [Google Scholar]
- [10].Farooq MT, Shaukat A, Akram U, Waqas O, Ahmad M, Automatic Gleason grading of prostate cancer using gabor filter and local binary patterns, in: 2017 40th International Conference on Telecommunications and Signal Processing (TSP), IEEE, 2017, pp. 642–645, doi: 10.1109/tsp.2017.8076065. [DOI] [Google Scholar]
- [11].Li W, Li J, Sarma KV, Ho KC, Shen S, Knudsen BS, Gertych A, Arnold CW, Path r-cnn for prostate cancer diagnosis and Gleasongrading of histological images, IEEE Trans. Med. Imaging 38 (4) (2018) 945–954, doi: 10.1109/tmi.2018.2875868. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [12].Arvaniti E, Fricker KS, Moret M, Rupp NJ, Hermanns T, Fankhauser C, Wey N, Wild PJ, Rueschoff JH, Claassen M, Automated Gleason grading of prostate cancer tissue microarrays via deep learning, Sci. Rep. 8 (2018), doi: 10.1038/s41598-018-30535-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [13].Nagpal K, Foote D, Liu Y, Chen P-HC, Wulczyn E, Tan F, Olson N, Smith JL, Mohtashamian A, Wren JH, et al. , Development and validation of a deep learning algorithm for improving Gleason scoring of prostate cancer, NPJ Digit. Med. 2 (1) (2019) 48, doi: 10.1038/s41746-019-0112-2. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [14].Campanella G, Hanna MG, Geneslaw L, Miraflor A, Werneck Krauss Silva V, Busam KJ, Brogi E, Reuter VE, Klimstra DS, Fuchs TJ, Clinical-grade computational pathology using weakly supervised deep learning on whole slide images, Nat. Med. 25 (8) (2019) 1301–1309, doi: 10.1038/s41591-019-0508-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [15].Poojitha UP, Sharma SL, Hybrid unified deep learning network for highly precise Gleason grading of prostate cancer, in: 2019 41st Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), IEEE, 2019, pp. 899–903, doi: 10.1109/EMBC.2019.8856912. [DOI] [PubMed] [Google Scholar]
- [16].Doyle S, Feldman M, Tomaszewski J, Madabhushi A, A boosted bayesian multi-resolution classifier for prostate cancer detection from digitized needle biopsies, IEEE Trans. Biomed. Eng. 59 (5) (2010) 1205–1218, doi: 10.1109/TBME.2010.2053540. [DOI] [PubMed] [Google Scholar]
- [17].Monaco JP, Madabhushi A, Class-specific weighting for markov random field estimation: application to medical image segmentation, Med. Image Anal. 16 (8) (2012) 1477–1489, doi: 10.1016/j.media.2012.06.007. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [18].Ström P, Kartasalo K, Olsson H, Solorzano L, Delahunt B, Berney DM, Bostwick DG, Evans AJ, Grignon DJ, Humphrey PA, et al. , Artificial intelligence for diagnosis and grading of prostate cancer in biopsies: a population-based, diagnostic study, Lancet Oncol. (2020), doi: 10.1016/S1470-2045(19)30738-7. [DOI] [PubMed] [Google Scholar]
- [19].Bulten W, Pinckaers H, van Boven H, Vink R, de Bel T, van Ginneken B, van der Laak J, Hulsbergen-van de Kaa C, Litjens G, Automated deep-learning system for Gleason grading of prostate cancer using biopsies: a diagnostic study, Lancet Oncol. (2020), doi: 10.1016/S1470-2045(19)30739-9. [DOI] [PubMed] [Google Scholar]
- [20].Fuchs TJ, Buhmann JM, Computational pathology: Challenges and promises for tissue analysis, Computerized Medical Imaging and Graphics 35 (7) (2011) 515–530, doi: 10.1016/j.compmedimag.2011.02.006. Whole Slide Image Process [DOI] [PubMed] [Google Scholar]
- [21].Madabhushi A, Feldman MD, Leo P, Deep-learning approaches for Gleason grading of prostate biopsies, Lancet Oncol. 21 (2) (2020) 187–189, doi: 10.1016/S1470-2045(19)30793-4. [DOI] [PubMed] [Google Scholar]
- [22].Ali S, Veltri R, Epstein JI, Christudass C, Madabhushi A, Selective invocation of shape priors for deformable segmentation and morphologic classification of prostate cancer tissue microarrays, Comput. Med. Imaging Graph. 41 (2015) 3–13, doi: 10.1016/j.compmedimag.2014.11.001. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [23].Leo P, Elliott R, Shih NNC, Gupta S, Feldman M, Madabhushi A, Stable and discriminating features are predictive of cancer presence and Gleason grade in radical prostatectomy specimens: a multi-site study, Sci. Rep. 8 (1) (2018) 14918, doi: 10.1038/s41598-018-33026-5. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [24].Lee G, Sparks R, Ali S, Shih NN, Feldman MD, Spangler E, Rebbeck T, Tomaszewski JE, Madabhushi A, Co-occurring gland angularity in localized subgraphs: predicting biochemical recurrence in intermediate-risk prostate cancer patients, PLoS ONE 9 (5) (2014), doi: 10.1371/journal.pone.0097954. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [25].Niazi MKK, Yao K, Zynger DL, Clinton SK, Chen J, Koyutürk M, LaFramboise T, Gurcan M, Visually meaningful histopathological features for automatic grading of prostate cancer, IEEE J. Biomed. Health Inform. 21 (4) (2017) 1027–1038, doi: 10.1109/JBHI.2016.2565515. [DOI] [PubMed] [Google Scholar]
- [26].Humphrey PA, Gleason grading and prognostic factors in carcinoma of the prostate, Mod. Pathol. 17 (3) (2004) 292, doi: 10.1038/modpathol.3800054. [DOI] [PubMed] [Google Scholar]
- [27].Chen N, Zhou Q, The evolving Gleason grading system, Chin. J. Cancer Res. 28 (1) (2016) 58, doi: 10.3978/j.issn.1000-9604.2016.02.04. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [28].Edelsbrunner H, Harer J, Persistent homology-a survey, Contemp. Math. 453 (2008) 257–282. [Google Scholar]
- [29].Qaiser T, Sirinukunwattana K, Nakane K, Tsang Y-W, Epstein D, Rajpoot N, Persistent homology for fast tumor segmentation in whole slide histology images, Procedia Comput. Sci. 90 (2016) 119–124, doi: 10.1016/j.procs.2016.07.033. [DOI] [Google Scholar]
- [30].Qaiser T, Tsang Y-W, Epstein D, Rajpoot N, Tumor segmentation in whole slide images using persistent homology and deep convolutional features, in: Annual Conference on Medical Image Understanding and Analysis, Springer, 2017, pp. 320–329, doi: 10.1007/978-3-319-60964-5_28. [DOI] [PubMed] [Google Scholar]
- [31].Qaiser T, Tsang Y-W, Taniyama D, Sakamoto N, Nakane K, Epstein D, Rajpoot N, Fast and accurate tumor segmentation of histology images using persistent homology and deep convolutional features, Med. Image Anal. (2019), doi: 10.1016/j.media.2019.03.014. [DOI] [PubMed] [Google Scholar]
- [32].Ruifrok AC, Johnston DA, et al. , Quantification of histochemical staining by color deconvolution, Anal. Quant. Cytol. Histol. 23 (4) (2001) 291–299. [PubMed] [Google Scholar]
- [33].Macenko M, Niethammer M, Marron JS, Borland D, Woosley JT, Guan X, Schmitt C, Thomas NE, A method for normalizing histology slides for quantitative analysis, in: 2009 IEEE International Symposium on Biomedical Imaging: From Nano to Macro, IEEE, 2009, pp. 1107–1110, doi: 10.1109/ISBI.2009.5193250. [DOI] [Google Scholar]
- [34].Epstein JI, A new contemporary prostate cancer grading system: message to the italian pathologists, Pathologica 107 (3–4) (2015) 205–207. [PubMed] [Google Scholar]
- [35].DeGroot MH, Schervish MJ, Probability and Statistics, Pearson Education, 2012. [Google Scholar]
- [36].Wonnacott TH, Wonnacott RJ, Introductory Statistics, 5, Wiley New York, 1990. [Google Scholar]
- [37].Xu J, Xiang L, Liu Q, Gilmore H, Wu J, Tang J, Madabhushi A, Stacked sparse autoencoder (ssae) for nuclei detection on breast cancer histopathology images, IEEE Trans. Med. Imaging 35 (1) (2016) 119–130, doi: 10.1109/TMI.2015.2458702. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [38].Howard AG, Zhu M, Chen B, Kalenichenko D, Wang W, Weyand T, Andreetto M, Adam H, Mobilenets: efficient convolutional neural networks for mobile vision applications, arXiv:1704.04861 (2017). [Google Scholar]
- [39].Otsu N, A threshold selection method from gray-level histograms, IEEE Trans. Syst. Man Cybern. 9 (1) (1979) 62–66, doi: 10.1109/tsmc.1979.4310076. [DOI] [Google Scholar]
- [40].Krizhevsky A, Sutskever I, Hinton GE, Imagenet classification with deep convolutional neural networks, in: Advances in Neural Information Processing Systems, 2012, pp. 1097–1105. [Google Scholar]
- [41].van Leenders GJ, Kweldam CF, Hollemans E, Kümmerlin IP, Nieboer D, Verhoef EI, Remmers S, Incrocci L, Bangma CH, van der Kwast TH, et al. , Improved prostate cancer biopsy grading by incorporation of invasive cribriform and intraductal carcinoma in the 2014 grade groups, Eur. Urol. 77 (2) (2020) 191–198. [DOI] [PubMed] [Google Scholar]
- [42].Nakane K, Takiyama A, Mori S, Matsuura N, Homology-based method for detecting regions of interest in colonic digital images, Diagnostic pathology 10 (1) (2015) 36. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [43].Nakane K, Tsuchihashi Y, Matsuura N, A simple mathematical model utilizing topological invariants for automatic detection of tumor areas in digital tissue images, Diagnostic pathology, volume 8, BioMed Central, 2013, p. S27. [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
