Skip to main content
Cancers logoLink to Cancers
. 2020 Sep 21;12(9):2708. doi: 10.3390/cancers12092708

Computer Extracted Features from Initial H&E Tissue Biopsies Predict Disease Progression for Prostate Cancer Patients on Active Surveillance

Sacheth Chandramouli 1, Patrick Leo 1, George Lee 1, Robin Elliott 2, Christine Davis 3, Guangjing Zhu 3, Pingfu Fu 4, Jonathan I Epstein 3, Robert Veltri 5, Anant Madabhushi 1,6,*
PMCID: PMC7563653  PMID: 32967377

Abstract

Simple Summary

Active surveillance (AS) prostate cancer patients suffer from a lower quality of life, increased risk of anxiety and depression, and an increased risk of disease progression compared to patients who opt for curative treatment. The current inclusion criteria for AS patients is unable to accurately identify patients with increased risk of progression, and therefore there is a need for a risk stratification technique that can identify patients with a higher risk of disease progression. In this work, we leverage quantitative histomorphometric (QH) features describing nuclear position, shape, orientation, and clustering from initial H&E biopsy images to accurately identify AS-eligible patients who are at high risk for disease progression. Our findings indicate that QH features were correlated with the risk of clinical progression in AS-eligible patients and was able to out-perform judgements based on clinical variables such as Gleason score and pro-PSA.

Abstract

In this work, we assessed the ability of computerized features of nuclear morphology from diagnostic biopsy images to predict prostate cancer (CaP) progression in active surveillance (AS) patients. Improved risk characterization of AS patients could reduce over-testing of low-risk patients while directing high-risk patients to therapy. A total of 191 (125 progressors, 66 non-progressors) AS patients from a single site were identified using The Johns Hopkins University’s (JHU) AS-eligibility criteria. Progression was determined by pathologists at JHU. 30 progressors and 30 non-progressors were randomly selected to create the training cohort D1 (n = 60). The remaining patients comprised the validation cohort D2 (n = 131). Digitized Hematoxylin & Eosin (H&E) biopsies were annotated by a pathologist for CaP regions. Nuclei within the cancer regions were segmented using a watershed method and 216 nuclear features describing position, shape, orientation, and clustering were extracted. Six features associated with disease progression were identified using D1 and then used to train a machine learning classifier. The classifier was validated on D2. The classifier was further compared on a subset of D2 (n = 47) against pro-PSA, an isoform of prostate specific antigen (PSA) more linked with CaP, in predicting progression. Performance was evaluated with area under the curve (AUC). A combination of nuclear spatial arrangement, shape, and disorder features were associated with progression. The classifier using these features yielded an AUC of 0.75 in D2. On the 47 patient subset with pro-PSA measurements, the classifier yielded an AUC of 0.79 compared to an AUC of 0.42 for pro-PSA. Nuclear morphometric features from digitized H&E biopsies predicted progression in AS patients. This may be useful for identifying AS-eligible patients who could benefit from immediate curative therapy. However, additional multi-site validation is needed.

Keywords: prostate cancer, active surveillance, machine learning, pathology

1. Introduction

For low grade prostate cancer (CaP) patients, active surveillance (AS) is an increasingly favored management strategy that involves repeat blood work, biopsies, and digital rectal exams [1]. However, AS patients express a significant decrease in their quality of life and an increased risk of anxiety and depression [1]. Repeat biopsies also increases the risk of infection [2]. Delaying curative treatment also introduces added complications of disease progression and metastasis [1]. As high as 43% of AS patients eventually opt for curative therapy due to biopsy progression or anxiety [3]. Physicians use many clinical risk factors to balance the risk of CaP progression and overtreatment.

These risk factors include prostate specific antigen (PSA) density, clinical stage, number of positive cores, and biopsy Gleason grade. Various institutions have defined different thresholds for these risk factors to define AS-eligibility criteria [4,5,6,7]. However, because no perfect threshold exists, these limited number of clinical factors are unable to accurately ascertain the likelihood that a patient will eventually need curative intervention [8,9]. Depending on different combinations of thresholds, between 12% to 31% of AS-eligible demonstrate clinical levels of progression and require curative treatment [8]. The difficulty in identifying patients for whom AS is appropriate is further compounded by the low specificity of PSA elevation [4,10,11], and low inter-reviewer agreement in Gleason grading [11].

Recently, there has been an increased interest in determining, within a cohort of AS patients, features that may be associated with a greater risk of clinically significant progression. Several groups identified serum, urinary, and tissue biomarkers linked to aggressive CaP morphology in AS-eligible patients [12]. However, these markers have yet to be directly correlated with patients who progress or are unable to determine the risk when the patient is initially placed in AS [4,13,14,15,16]. Therefore, there is a need for an accurate, quantifiable method that can determine early whether a patient is at an increased risk of significant progression in CaP.

With the increase in computing power, features extracted from initial biopsies can directly characterize the tumor morphology and predict outcome instantly and accurately. Quantitative histomorphometric (QH) features are computationally derived descriptors of morphology extracted from digitized pathology slides. QH approaches have shown to be useful in characterizing CaP in a number of studies [17,18,19,20,21,22], including through the use of deep learning, in which features are automatically generated via a cascade of neural networks [23,24]. While QH features have been employed for CaP detection, grading, and biochemical recurrence prediction, no previous study, to our knowledge, has attempted to associate QH features with the risk of clinical progression in AS patients.

In this study, we explore the potential role of QH features including spatial arrangement, shape, and disorder of nuclei from within cancer defined regions on initial H&E biopsies to differentiate AS patients who are at low and high risk of clinical progression. A total of n = 191 AS patients were identified from a single institution with a corresponding initial Hematoxylin & Eosin (H&E) biopsy. Supervised machine learning classification approaches were employed to evaluate the ability of nuclear QH features extracted from digitized H&E images of baseline biopsies in predicting risk of progression. The ability of these QH features to predict likelihood of progression was also further compared against a known CaP biomarker, pro-PSA, in n = 47 patients. Finally, the QH model was compared to a deep learning network-based model.

2. Materials and Methods

2.1. Dataset

Digitized H&E needle-core biopsy images from 191 AS patients (125 progressors, 66 non-progressors) from Johns Hopkins University (JHU) who met the criteria of stage T1c, PSA ≤ 10 ng/mL, Gleason sum (GS) ≤ 7, ≤ 2 positive cores, ≤ 50% core involvement, PSA density ≤ 0.15 ng/mL and life expectancy ≤ 20 years were gathered. All cases used for TMAs and AS biopsies were consented under an Institutional Review Board-approved protocol at Johns Hopkins University School of Medicine. De-identified images with no patient header information (PHI) was provided to the CWRU team and hence from an analysis perspective the research we deemed to be human subjects exempt. Histopathologic confirmation of upgrading and upstaging was determined by pathologists at JHU to differentiate progressors and non-progressors. Specimens were digitized at 40× magnification (0.25 microns per pixel). All studies were conducted in a singled representative CaP region per patient annotated by a pathologist. 30 progressors and 30 non-progressors were randomly selected to constitute the training set D1 (n = 60) and the remaining set of 131 (95 progressors, 36 non-progressors) patients comprised the test set D2. A subset (D3, n = 47) of D2 also had pro-PSA measurements.

2.2. Nuclear Segmentation

Each H&E stained image was first normalized to the same color intensity level by adjusting the color levels of the image using the maximum stain vector as a template [25]. Nuclei were detected by convoluting a bank of directional Gaussian filter kernels with the normalized H&E images to generate a response map [24].

Areas of local maximums in the response map indicated high probability of a nucleus. Each nucleus was then automatically segmented by a watershed-based nuclear segmentation method [25]. From these segmentations, the boundary points and centroid of every nucleus was extracted. An example segmentation result is shown in Figure 1.

Figure 1.

Figure 1

Feature maps of a patient that did not progresses (left) and one who progressed (right). Nuclear segmentation results representing the nuclear shape (a,b). Nuclear spatial arrangement represented through the Delaunay triangulation (c,e) Voronoi tessellation (d,f), and Minimum Spanning Trees (g,i). Disorder in localized nuclear orientation (h,j).

2.3. Feature Extraction

Subsequent to nuclear segmentation, a total of 216 nuclear morphology features were extracted from the segmentation results. These features belonged to four categories: graph, shape, nuclear disorder, and cluster graph.

Graph features were constructed based on nuclear centroids. Voronoi tessellation, Delaunay triangulation, minimum spanning trees, and k-nearest neighbor plots [17,18] were constructed over the entire annotated region of each image. Statistics such as mean and standard deviation were derived from the areas and perimeters of Voronoi polygons, Delaunay maps, minimum spanning tree edge lengths, and distances to nearest 3, 5, and 7 nuclear centroids [17,18].

The nuclear shape-based features were derived from the boundary of the segmented nuclei. These measures included nuclear area, perimeter, smoothness, fractal dimension, and invariant moments, and Fourier descriptors [19,20,21,22]. For each measure, the mean, median, standard deviation, and minimum to maximum ratio statistics were extracted.

The nuclear disorder features measure the disorder in the orientation of the nuclei within a local neighborhood [18,26]. The orientation of each nucleus was determined using the first principal component of the boundary points. This orientation vector roughly corresponds to the direction of the major axis of the nucleus. Local subgraphs of the nuclear arrangement were created with a probabilistic decay model based on Euclidean distance and an empirically determined decaying factor [18]. A co-occurrence matrix was created using the relative orientation of each nucleus in each subgraph. Measures of entropy, energy, and contrast were extracted from this matrix [18].

The cell cluster graph features use clusters of nuclei as nodes and capture the local and global spatial relationships between each cluster [27,28]. These spatial relationships include edge length, connectivity coefficients, and clustering coefficients. These features are shown in Figure 1 and are summarized in Table 1.

Table 1.

Summary of QH based features to describe tumor morphology.

Feature Family Description Features
Graph Voronoi, Delaunay, minimum spanning trees, k-NN graphs 51
Shape Area, perimeter ratio, smoothness, distance, etc. 100
Nuclear Disorder Orientation entropy, energy, contrast 39
Cluster Graphs Clustering coefficient, edge length, connected components 26

2.4. Feature Identification

In order to identify the optimal model and feature set, the 216 features from Table 1 were evaluated on D1 for stability and accuracy. D1 was randomly divided into two cohorts, cohort D1,A and cohort D1,B, each with 30 patients (15 progressors, 15 non-progressors). Discriminating features were first identified on cohort D1,A. Two sets of six discriminating feature were selected, one by Wilcoxon rank-sum (WLCX), the other by t-test (TT). Each of these sets were then used to train two classifiers on cohort D1,A, one using linear discriminant analysis (LDA), the other using quadratic discriminant analysis (QDA). The patients in cohort D1,B were then classified using each of these four models (TT + QDA, TT + LDA, WLCX + QDA, WLCX + LDA). The AUC of each model on cohort D1,B was recorded. Cohorts D1,A and D1,B were then randomly redrawn from the training set and the feature identification and classifier training process was repeated 30 times.

From the 30 iterations, only feature combinations yielding an AUC ≥ 0.65 were retained. This threshold was chosen empirically to maximize predictive potential and minimize overfitting. The six most recurring features among the retained feature combinations were deemed discriminatory and used for classifier construction. These six most recurring features are listed in Table 2.

Table 2.

The six features identified from D1 and used in the final model.

Feature
Voronoi: Min/max polygon perimeter
Shape: Min/max standard deviation of distance of contour point from centroid of nuclei
Shape: Min/max nuclei perimeter
Shape: Standard deviation of Fourier descriptor 6
Orientation: Tensor contrast inverse moment range
Voronoi: Standard deviation of polygon area

Hierarchical clustering of the top six features in D1 was performed using spearman distance to further evaluate the top six features’ ability in differentiating outcome [28]. Clustering results were visualized on D1 using a heatmap where green indicates higher expression of the feature values and brown indicates the converse. These unsupervised clustering results were compared against patient outcome to evaluate associations between feature expression and patient outcome.

2.5. Classifier Construction

The 6 features identified in the previous step were used to create three models for progression prediction. These models were constructed using LDA, QDA, and random forest (RF) classifiers. 3-fold cross validation was used in 100 iterations to identify the model with the highest average AUC. The optimal operating point for the best model was determined on D1. The best model and the operating point was then evaluated on D2 to determine AUC and accuracy. A confusion matrix was constructed using the best model and the operating point on D2 and displayed in Table 3. To establish a comparison point for QH features, the AUC of a two clinical predictors (pro-PSA, and biopsy Gleason sum) was investigated.

Table 3.

A confusion matrix created on D2 using the RF model using an operating point threshold of 0.69. The positive predictive value (PPV), negative predictive value (NPV), sensitivity and specificity was also calculated on D2 using the same operating point.

Predicted
Actual Non-Progressor Progressor
Non-Progressor 22 14 NPV: 44%
Progressor 28 67 PPV: 83%
Specificity: 61% Sensitivity: 71%

2.6. Deep Learning Model

A deep learning model using the DenseNet architecture [29] was trained to compare to the QH models. The deep learning model takes images as input, splits them into 224 × 224 pixel patches at 1 MPP (equivalent to 10× magnification), and renders a progressor/non-progressor decision on each patch. Model hyperparameters were chosen to produce a smaller version of the DenseNet121 model more suitable for a small dataset, resulting in a model with a 64 features in the first layer and an additional 12 in each layer with a feature multiplier of 4 in the bottleneck layers, two layers in each pooling block, and no dropout, for a total of 79,784 parameters. This model was trained on D1, split into a 50 patient training set and 10 patient testing set, for 500 epochs, with the final model being the one with the lowest loss on the test set. This model was then applied to D2, on which per-image AUC, based on fraction of positive patches, and accuracy, based on the majority vote of the patches, were calculated.

3. Results

3.1. Experiment 1: Identifying Features Associated with Risk of Progression in AS Patients

The top six nuclear QH features obtained during feature identification within D1 are listed in Table 2. Three nuclear shape features were represented in the six feature set, two nuclear spatial arrangement features, and a single feature representing the localized nuclear disorder.

As can be seen in Figure 2, there exists good separation between patient outcome and feature values. For all the shape features (columns 1, 2, 4), an increasing feature value is associated with a better prognosis. For the Voronoi polygon area feature (column 6), an increasing feature value is closely associated with clinical progression. Similarly, a high nuclear disorder feature (column 5) seems to be strongly associated with an increased risk of clinical progression. A high Voronoi polygon perimeter feature (column 3) value appears to associate with a better disease prognosis.

Figure 2.

Figure 2

The 6 features used to create a clustergram on D1. Features and patients were clustered hierarchically. The shading of each cell shows the relative over- or under-expression of that feature in a patient. From left to right, the features are: (1) standard deviation of minimum/maximum nuclear radius, (2) minimum/maximum nuclear perimeter, (3) minimum/maximum ratio of Voronoi polygon perimeter, (4) standard deviation of nuclear shape’s Fourier descriptor 6, (5) tensor contrast inverse moment range, and (6) standard deviation of Voronoi polygon area.

3.2. Experiment 2: QH based Classifier Construction to Predict Progression in AS Patients

Results from the three-fold cross validation with LDA, QDA, and RF models using the top six features are shown in Table 4. The RF model yielded an AUC of 0.75 in D2. Within D3, the RF model yielded an AUC of 0.79, surpassing both pro-PSA (AUC = 0.42) and Gleason (AUC = 0.64).

Table 4.

The average AUC, accuracy (ACC), sensitivity (SENS), and specificity (SPEC) across 100 iterations for each classification scheme on D1 using the 6 discriminating features from Experiment 1. These values were calculated using 3-fold cross validation. Each fold had an equal number of progressors (n = 10) and non-progressors (n = 10).

Model Type AUC ACC SENS SPEC
LDA 0.68 ± 0.12 0.63 ± 0.11 0.53 ± 0.16 0.26 ± 0.13
QDA 0.69 ± 0.11 0.62 ± 0.10 0.58 ± 0.17 0.67 ± 0.16
RF 0.73 ± 0.10 0.75 ± 0.09 0.70 ± 0.17 0.81 ± 0.16

3.3. Experiment 3: Deep Learning Classifier for Prediction of Progression in AS Patients

The deep learning model had an AUC of 0.64 based on fraction of positive patches and accuracy of 0.73 based on patch majority voting in D2, performing worse than the random forest model. While the deep learning model had a specificity of 0.96, its sensitivity was just 0.14. This bias arose despite the deep learning model’s training set consisting of an equal number of progressor and non-progressor cases.

4. Discussion

Patient inclusion criteria for prostate cancer (CaP) patients on active surveillance is not an exact science. Many studies have suggested that the current inclusion criteria fails to accurately exclude patients at a high-risk of CaP progression [1,8,9]. Therefore, there has been an increased effort to determine factors correlated with an increased risk of progression in AS patients. Multivariate nomograms have been tested with AS patients to track CaP progression, however average AUCs reported were 0.59 [30]. Some groups have also attempted to identify serum, and urinary clinical biomarkers to determine the risk of progression in disease using protein expression, and tumor morphology [12,15,16]. For example, PCA3 and pro-PSA have been studied as predictors of more aggressive disease. However, PCA3 possesses low predictive power (AUC = 0.66) and provides no increased independent predictive value compared to Gleason score [4,12]. Pro-PSA has been reported to predict histologic upgrading and upstaging with a c-index of 0.62 for AS-eligible patients, however lacks a comparative analysis to other clinical variables such as Gleason score [4,31].

In this work, we evaluated the ability of QH features relating to nuclear spatial arrangement, shape, and disorder to quantitatively describe tumor morphology from an initial H&E needle core biopsy and predict the likelihood of progression in AS patients. We validated this model on an independent validation set (n = 131) and compared it against pro-PSA, a form of PSA that is more linked with CaP, and Gleason sum on a subset (n = 47) using AUC. Additionally, we compared our model to a deep learning-based method.

Several studies have evaluated the use of QH features to characterize nuclear and glandular morphology for CaP detection, Gleason grading, and BCR prediction. Doyle et al. used 103 features describing nuclear and glandular spatial arrangement, and texture to classify H&E biopsies into Gleason grade 3 and 4 with an accuracy of 76.9%. Other studies have exceeded the limitations of Gleason grade by relating the nuclear and glandular architecture directly with patient outcome. Lee et al. derived 39 features describing entropy in nuclear orientation from CaP tissue microarray (TMA) to predict biochemical recurrence post-radical prostatectomy (RP) [26]. The localized cell clustering has been used to predict 5-year BCR using TMAs with an accuracy of 83.1% [32].

To our knowledge, the work presented in this study, represents the first attempt to use computerized histologic image analysis and machine learning approaches to predict likelihood of CaP progression in AS patients based off routine H&E biopsy tissue images.

We found that the smaller the variance in the roundness of the nuclear perimeter, the better the prognosis. This suggests that a greater degree of nuclear shape homogeneity is associated with a lower risk of clinical progression, a finding consistent with prior work in CaP and breast cancer [11,18,19,32,33,34,35]. The Voronoi polygons constructed using the nuclear centroids showed a relationship between density of nuclear packing and patient outcome. Patients that exhibited greater variability in density of nuclear packing on the initial biopsy were associated with an increased likelihood of clinically significant CaP progression, a conclusion that is consistent with other studies that found decreasing organization of nuclear arrangement closely related to more aggressive disease [36]. Finally, the nuclear orientation feature is a second order nuclear angle statistic representing the range of the orientations of the nuclei in local subgraphs on an image. In this dataset, a smaller range of nuclear orientations present on a biopsy slide was associated with a better prognosis. Lee et al. has reported similar results with glandular [18] and nuclear arrangements in predicting BCR [26].

The combination of the six QH features yielded a higher predictive power (AUC = 0.75) than the average reported predictive capability of standard nomograms (average AUC = 0.59) [30]. Due to the low specificity of PSA, groups have investigated the use of pro-PSA, a variant of PSA, in differentiating benign and aggressive disease [31,37,38,39,40]. While pro-PSA has been reported to be significantly correlated with upgrade in disease, in this dataset the pro-PSA had very low discriminating ability (AUC = 0.42) when compared to the QH model (AUC = 0.79) [31,38].

The deep learning method (AUC = 0.64) performed worse than the QH model, with a sensitivity (0.14) far too low to be clinically useful. While it is possible that further refinement of the model could improve these results, it is clear that deep learning does not provide a simple solution to prediction of progression from AS.

Our study did however have its limitations. First, in the 191 patient dataset, only 47 patients in our study had pro-PSA information for us to compare to. Second, we acknowledge that one of the limitations of our study was the lack of detailed manual annotations of individual nuclei that precluded us from comprehensively evaluating the sensitivity of the feature extraction process on fidelity of the nuclear contours. Our dataset also lacked time-to-progression information required to conduct a survival analysis of our model. Our dataset was missing clinical information needed to compare the performance of our QH model with that of nomograms. Our entire process was only conducted in a small region of interest identified by a pathologist rather than the entire whole slide image of the biopsy specimen. Our study was also conducted using patients from a single institution. Finally, while we did not explicitly address the issue of tumor heterogeneity, it was interesting that the QH features from a single tumor focus was associated with likelihood of progression. A direction of future work would be to explicitly invoke multiple different tumor foci and evaluate whether the associated QH measurements could even better prognosticate likelihood of disease progression. Future work will also need to involve a larger, multi-site, independent validation of these findings.

5. Conclusions

In this study, we demonstrated the ability of QH features extracted from initial needle core biopsies to identify AS-eligible patients at high risk of progression. While our QH model was found to be accurate on a separate test set and appeared to outperform a clinically employed CaP biomarker, additional multi-site testing is needed prior to use of this approach for clinical management of CaP patients on AS.

The deep learning method (AUC = 0.64) performed worse than the QH model, with a sensitivity (0.14) far too low to be clinically useful. While it is possible that further refinement of the model could improve these results, it is clear that deep learning does not provide a simple solution to prediction of progression from AS based off digital biopsy scans. Furthermore, there is no straightforward way to understand the rationale for the decisions of the deep learning model, obfuscating model interpretability. It may be that much larger training sets might be needed to develop more robust deep learning based prognostic models for prostate cancer in the active surveillance setting.

Utilizing QH features to aid in selecting ideal candidates for AS could decrease the rate of overtreatment in low-risk CaP patients with minimal additional costs as it only requires digitization of an initial biopsy and a standard desktop computer. While digital pathology services are not currently routinely performed in every pathology lab, with the recent Food and Drug Administration (FDA) approval of whole slide scanners from two different vendors, it is highly likely that slide digitization will become more routine and prevalent in the coming years [41].

Author Contributions

S.C., P.L., and A.M. conceived the experiments and analyzed the results. S.C. and P.L. performed the experiments. P.F. ensured the statistical analysis and results were appropriate. G.L., C.D., G.Z., J.I.E., and R.V. gathered and digitized the slides from JHU. R.E. annotated the digitized slides for regions of interest. S.C., P.L., and A.M. wrote the manuscript. All authors reviewed the manuscript and contributed to the final manuscript. All authors have read and agreed to the published version of the manuscript

Funding

Research reported in this publication was supported by the National Cancer Institute of the National Institutes of Health under award numbers 1U24CA199374-01, R01CA202752-01A1 R01CA208236-01A1, R01 CA216579-01A1, R01 CA220581-01A1, 1U01 CA239055-01, 1U01CA248226-01, National Heart, Lung and Blood Institute 1R01HL15127701A1, National Institute for Biomedical Imaging and Bioengineering 1R43EB028736-01, National Center for Research Resources under award number 1 C06 RR12463-01, VA Merit Review Award IBX004121A from the United States Department of Veterans Affairs Biomedical Laboratory Research and Development Service, The DoD Breast Cancer Research Program Breakthrough Level 1 Award W81XWH-19-1-0668, The DOD Prostate Cancer Idea Development Award (W81XWH-15-1-0558), The DOD Lung Cancer Investigator-Initiated Translational Research Award (W81XWH-18-1-0440), The DOD Peer Reviewed Cancer Research Program (W81XWH-16-1-0329), The Kidney Precision Medicine Project (KPMP) Glue Grant, 5T32DK747033 CWRU Nephrology Training Grant, Neptune Career Development Award, The Ohio Third Frontier Technology Validation Fund, The Wallace H. Coulter Foundation Program in the Department of Biomedical Engineering and The Clinical and Translational Science Award Program (CTSA) at Case Western Reserve University. The content is solely the responsibility of the authors and does not necessarily represent the official views of the National Institutes of Health, the U.S. Department of Veterans Affairs, the Department of Defense, or the United States Government.

Conflicts of Interest

Madabhushi is an equity holder in Elucid Bioimaging and in Inspirata Inc. He is also a scientific advisory consultant for Inspirata Inc. In addition he has served as a scientific advisory board member for Inspirata Inc, Astrazeneca, Bristol Meyers-Squibb and Merck. He also has sponsored research agreements with Philips and Inspirata Inc. His technology has been licensed to Elucid Bioimaging and Inspirata Inc. He is also involved in a NIH U24 grant with PathCore Inc., and 3 different R01 grants with Inspirata Inc. George Lee is a digital pathology informatics lead and a data scientist at Bristol-Myers Squibb.

References

  • 1.Albertsen P.C. Treatment of localized prostate cancer: When is active surveillance appropriate? Nat. Rev. Clin. Oncol. 2010;7:394–400. doi: 10.1038/nrclinonc.2010.63. [DOI] [PubMed] [Google Scholar]
  • 2.Evans R., Loeb A., Kaye K.S., Cher M.L., Martin E.T. Infection-Related Hospital Admissions After Prostate Biopsy in United States Men. Open Forum Infect. Dis. 2017;4:ofw265. doi: 10.1093/ofid/ofw265. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3.Albertsen P.C. Observational studies and the natural history of screen-detected prostate cancer. Curr. Opin. Urol. 2015;25:232. doi: 10.1097/MOU.0000000000000157. [DOI] [PubMed] [Google Scholar]
  • 4.Briganti A., Fossati N., Catto J.W.F., Cornford P., Montorsi F., Mottet N., Wirth M., Poppel H.V. Active Surveillance for Low-risk Prostate Cancer: The European Association of Urology Position in 2018. Eur. Urol. 2018;74:357–368. doi: 10.1016/j.eururo.2018.06.008. [DOI] [PubMed] [Google Scholar]
  • 5.Bul M., Zhu X., Valdagni R., Pickles T., Kakehi Y., Rannikko A., Bjartell A., van der Schoot D.K., Cornel E.B., Conti G.N., et al. Active Surveillance for Low-Risk Prostate Cancer Worldwide: The PRIAS Study. Eur. Urol. 2013;63:597–603. doi: 10.1016/j.eururo.2012.11.005. [DOI] [PubMed] [Google Scholar]
  • 6.Soloway M.S., Soloway C.T., Eldefrawy A., Acosta K., Kava B., Manoharan M. Careful Selection and Close Monitoring of Low-Risk Prostate Cancer Patients on Active Surveillance Minimizes the Need for Treatment. Eur. Urol. 2010;58:831–835. doi: 10.1016/j.eururo.2010.08.027. [DOI] [PubMed] [Google Scholar]
  • 7.Porten S.P., Whitson J.M., Cowan J.E., Cooperberg M.R., Shinohara K., Perez N., Greene K.L., Meng M.V., Carroll P.R. Changes in prostate cancer grade on serial biopsy in men undergoing active surveillance. J. Clin. Oncol. 2011;29:2795–2800. doi: 10.1200/JCO.2010.33.0134. [DOI] [PubMed] [Google Scholar]
  • 8.Welty C.J., Cooperberg M.R., Carroll P.R. Meaningful End Points and Outcomes in Men on Active Surveillance for Early–Stage Prostate Cancer. Curr. Opin. Urol. 2014;24:288–292. doi: 10.1097/MOU.0000000000000039. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9.Bruinsma S.M., Bangma C.H., Carroll P.R., Leapman M.S., Rannikko A., Petrides N., Weerakoon M., Bokhorst L.P., Roobol M.J. The Movember GAP3 Consortium. Active surveillance for prostate cancer: A narrative review of clinical guidelines. Nat. Rev. Urol. 2016;13:151–167. doi: 10.1038/nrurol.2015.313. [DOI] [PubMed] [Google Scholar]
  • 10.Mottet N., Bellmunt J., Bolla M., Briers E., Cumberbatch M.G., De Santis M., Fossati N., Gross T., Henry A.M., Joniau S., et al. EAU-ESTRO-SIOG Guidelines on Prostate Cancer. Part 1: Screening, Diagnosis, and Local Treatment with Curative Intent. Eur. Urol. 2017;71:618–629. doi: 10.1016/j.eururo.2016.08.003. [DOI] [PubMed] [Google Scholar]
  • 11.McLean M., Srigley J., Banerjee D., Warde P., Hao Y. Interobserver variation in prostate cancer Gleason scoring: Are there implications for the design of clinical trials and treatment strategies? Clin. Oncol. 1997;9:222–225. doi: 10.1016/S0936-6555(97)80005-2. [DOI] [PubMed] [Google Scholar]
  • 12.Loeb S., Tosoian J.J. Biomarkers in active surveillance. Transl. Androl. Urol. 2018;7:155–159. doi: 10.21037/tau.2017.12.26. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.Newcomb L.F., Brooks J.D., Carroll P.R., Feng Z., Gleave M.E., Nelson P.S., Thompson I.M., Lin D.W. Canary Prostate Active Surveillance Study (PASS); Design of a Multi-institutional Active Surveillance Cohort and Biorepository. Urology. 2010;75:407–413. doi: 10.1016/j.urology.2009.05.050. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14.Cooperberg M.R., Carroll P.R., Klotz L. Active Surveillance for Prostate Cancer: Progress and Promise. J. Clin. Oncol. 2011;29:3669–3676. doi: 10.1200/JCO.2011.34.9738. [DOI] [PubMed] [Google Scholar]
  • 15.Lin D.W., Newcomb L.F., Brown E.C., Brooks J.D., Carroll P.R., Feng Z., Gleave M.E., Lance R.S., Sanda M.G., Thompson I.M., et al. Urinary TMPRSS2:ERG and PCA3 in an active surveillance cohort: Results from a baseline analysis in the Canary Prostate Active Surveillance Study. Clin. Cancer Res. Off. J. Am. Assoc. Cancer Res. 2013;19:2442–2450. doi: 10.1158/1078-0432.CCR-12-3283. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16.Cooperberg M.R., Brooks J.D., Faino A.V., Newcomb L.F., Kearns J.T., Carroll P.R., Dash A., Etzioni R., Fabrizio M.D., Gleave M.E., et al. Refined Analysis of Prostate-specific Antigen Kinetics to Predict Prostate Cancer Active Surveillance Outcomes. Eur. Urol. 2018;74:211–217. doi: 10.1016/j.eururo.2018.01.017. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17.Gurcan M.N., Boucheron L.E., Can A., Madabhushi A., Rajpoot N.M., Yener B. Histopathological Image Analysis: A Review. IEEE Rev. Biomed. Eng. 2009;2:147–171. doi: 10.1109/RBME.2009.2034865. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18.Lee G., Sparks R., Ali S., Shih N.N.C., Feldman M.D., Spangler E., Rebbeck T., Tomaszewski J.E., Madabhushi A. Co-occurring gland angularity in localized subgraphs: Predicting biochemical recurrence in intermediate-risk prostate cancer patients. PLoS ONE. 2014;9:e97954. doi: 10.1371/journal.pone.0097954. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19.Veltri R.W., Isharwal S., Miller M.C., Epstein J.I., Partin A.W. Nuclear roundness variance predicts prostate cancer progression, metastasis, and death: A prospective evaluation with up to 25 years of follow-up after radical prostatectomy. Prostate. 2010;70:1333–1339. doi: 10.1002/pros.21168. [DOI] [PubMed] [Google Scholar]
  • 20.Farjam R., Soltanian-Zadeh H., Jafari-Khouzani K., Zoroofi R.A. An image analysis approach for automatic malignancy determination of prostate pathological images. Cytom. B Clin. Cytom. 2007;72:227–240. doi: 10.1002/cyto.b.20162. [DOI] [PubMed] [Google Scholar]
  • 21.Diamond D.A., Berry S.J., Umbricht C., Jewett H.J., Coffey D.S. Computerized image analysis of nuclear shape as a prognostic factor for prostatic cancer. Prostate. 1982;3:321–332. doi: 10.1002/pros.2990030402. [DOI] [PubMed] [Google Scholar]
  • 22.Tabesh A., Teverovskiy M., Pang H.-Y., Kumar V.P., Verbel D., Kotsianti A., Saidi O. Multifeature prostate cancer diagnosis and Gleason grading of histological images. IEEE Trans. Med. Imaging. 2007;26:1366–1378. doi: 10.1109/TMI.2007.898536. [DOI] [PubMed] [Google Scholar]
  • 23.Ström P., Kartasalo K., Olsson H., Solorzano L., Delahunt B., Berney D.M., Bostwick D.G., Evans A.J., Grignon D.J., Humphrey P.A., et al. Artificial intelligence for diagnosis and grading of prostate cancer in biopsies: A population-based, diagnostic study. Lancet Oncol. 2020;21:222–232. doi: 10.1016/S1470-2045(19)30738-7. [DOI] [PubMed] [Google Scholar]
  • 24.Bulten W., Pinckaers H., van Boven H., Vink R., de Bel T., van Ginneken B., van der Laak J., de Kaa C.H., Litjens G. Automated deep-learning system for Gleason grading of prostate cancer using biopsies: A diagnostic study. Lancet Oncol. 2020;21:233–241. doi: 10.1016/S1470-2045(19)30739-9. [DOI] [PubMed] [Google Scholar]
  • 25.Macenko M., Niethammer M., Marron J.S., Borland D., Woosley J.T., Guan X., Schmitt C., Thomas N.E. A method for normalizing histology slides for quantitative analysis; Proceedings of the 2009 IEEE International Symposium on Biomedical Imaging: From Nano to Macro; Boston, MA, USA. 28 June–1 July 2009; pp. 1107–1110. [Google Scholar]
  • 26.Lee G., Ali S., Veltri R., Epstein J.I., Christudass C., Madabhushi A. Cell orientation entropy (COrE): Predicting biochemical recurrence from prostate cancer tissue microarrays; Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention—MICCAI 2013: Medical Image Computing and Computer-Assisted Intervention–MICCAI 2013; Nagoya, Japan. 22–26 September 2013; pp. 396–403. [DOI] [PubMed] [Google Scholar]
  • 27.Ali S., Lewis J., Madabhushi A. Spatially aware cell cluster(spACC1) graphs: Predicting outcome in oropharyngeal pl6+ tumors; Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention—MICCAI 2013: Medical Image Computing and Computer-Assisted Intervention–MICCAI 2013; Nagoya, Japan. 22–26 September 2013; pp. 412–419. [DOI] [PubMed] [Google Scholar]
  • 28.Lee G., Singanamalli A., Wang H., Feldman M.D., Master S.R., Shih N.N.C., Spangler E., Rebbeck T., Tomaszewski J.E., Madabhushi A. Supervised multi-view canonical correlation analysis (sMVCCA): Integrating histologic and proteomic features for predicting recurrent prostate cancer. IEEE Trans. Med. Imaging. 2015;34:284–297. doi: 10.1109/TMI.2014.2355175. [DOI] [PubMed] [Google Scholar]
  • 29.Huang G., Liu Z., van der Maaten L., Weinberger K.Q. Densely Connected Convolutional Networks. arXiv. 2018160806993 Cs [Google Scholar]
  • 30.Wang S.-Y., Cowan J.E., Cary K.C., Chan J.M., Carroll P.R., Cooperberg M.R. Limited ability of existing nomograms to predict outcomes in men undergoing active surveillance for prostate cancer. BJU Int. 2014;114:E18–E24. doi: 10.1111/bju.12554. [DOI] [PubMed] [Google Scholar]
  • 31.Isharwal S., Makarov D.V., Sokoll L.J., Landis P., Marlow C., Epstein J.I., Partin A.W., Carter H.B., Veltri R.W. ProPSA and Diagnostic Biopsy Tissue DNA Content Combination Improves Accuracy to Predict Need for Prostate Cancer Treatment Among Men Enrolled in an Active Surveillance Program. Urology. 2011;77:763.e1–763.e6. doi: 10.1016/j.urology.2010.07.526. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 32.Ali S., Veltri R., Epstein J.A., Christudass C., Madabhushi A. Cell cluster graph for prediction of biochemical recurrence in prostate cancer patients from tissue microarrays; Proceedings of the Medical Imaging 2013: Digital Pathology; International Society for Optics and Photonics; Lake Buena Vista (Orlando Area), FL, USA. 9–14 February 2013; p. 86760H. [Google Scholar]
  • 33.Freedland S.J., Kane C.J., Amling C.L., Aronson W.J., Terris M.K., Presti J.C. Upgrading and Downgrading of Prostate Needle Biopsies: Risk Factors and Clinical Implications. Urology. 2007;69:495–499. doi: 10.1016/j.urology.2006.10.036. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 34.Veltri R.W., Miller M.C., Isharwal S., Marlow C., Makarov D.V., Partin A.W. Prediction of Prostate-Specific Antigen Recurrence in Men with Long-term Follow-up Postprostatectomy Using Quantitative Nuclear Morphometry. Cancer Epidemiol. Prev. Biomark. 2008;17:102–110. doi: 10.1158/1055-9965.EPI-07-0175. [DOI] [PubMed] [Google Scholar]
  • 35.Lu C., Romo-Bucheli D., Wang X., Janowczyk A., Ganesan S., Gilmore H., Rimm D., Madabhushi A. Nuclear shape and orientation features from H&E images predict survival in early-stage estrogen receptor-positive breast cancers. Lab. Investig. J. Tech. Methods Pathol. 2018;98:1438–1448. doi: 10.1038/s41374-018-0095-7. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 36.Christens-Barry W.A., Partin A.W. Quantitative Grading of Tissue and Nuclei in Prostate Cancer for Prognosis Prediction. JOHNS HOPKINS APL Tech. Dig. 1997;18:8. [Google Scholar]
  • 37.Veltri R.W. Serum marker %[-2]proPSA and the Prostate Health Index improve diagnostic accuracy for clinically relevant prostate cancer. BJU Int. 2016;117:12–13. doi: 10.1111/bju.13151. [DOI] [PubMed] [Google Scholar]
  • 38.Tosoian J.J., Loeb S., Feng Z., Isharwal S., Landis P., Elliot D.J., Veltri R., Epstein J.I., Partin A.W., Carter H.B., et al. Association of [-2]proPSA with biopsy reclassification during active surveillance for prostate cancer. J. Urol. 2012;188:1131–1136. doi: 10.1016/j.juro.2012.06.009. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 39.Mikolajczyk S.D., Rittenhouse H.G. Pro PSA: A more cancer specific form of prostate specific antigen for the early detection of prostate cancer. Keio J. Med. 2003;52:86–91. doi: 10.2302/kjm.52.86. [DOI] [PubMed] [Google Scholar]
  • 40.Heidegger I., Klocker H., Pichler R., Pircher A., Prokop W., Steiner E., Ladurner C., Comploj E., Lunacek A., Djordjevic D., et al. ProPSA and the Prostate Health Index as predictive markers for aggressiveness in low-risk prostate cancer—Results from an international multicenter study. Prostate Cancer Prostatic Dis. 2017;20:271–275. doi: 10.1038/pcan.2017.3. [DOI] [PubMed] [Google Scholar]
  • 41.Caccomo S. FDA Allows Marketing of First Whole Slide Imaging System for Digital Pathology. [(accessed on 7 September 2020)]; Available online: https://www.fda.gov/news-events/press-announcements/fda-allows-marketing-first-whole-slide-imaging-system-digital-pathology.

Articles from Cancers are provided here courtesy of Multidisciplinary Digital Publishing Institute (MDPI)

RESOURCES