Skip to main content
Journal of Medical Imaging logoLink to Journal of Medical Imaging
. 2017 Jun 21;4(2):027501. doi: 10.1117/1.JMI.4.2.027501

Gland segmentation in prostate histopathological images

Malay Singh a,b, Emarene Mationg Kalaw b, Danilo Medina Giron c, Kian-Tai Chong d, Chew Lim Tan a, Hwee Kuan Lee a,b,e,*
PMCID: PMC5479152  PMID: 28653016

Abstract.

Glandular structural features are important for the tumor pathologist in the assessment of cancer malignancy of prostate tissue slides. The varying shapes and sizes of glands combined with the tedious manual observation task can result in inaccurate assessment. There are also discrepancies and low-level agreement among pathologists, especially in cases of Gleason pattern 3 and pattern 4 prostate adenocarcinoma. An automated gland segmentation system can highlight various glandular shapes and structures for further analysis by the pathologist. These objective highlighted patterns can help reduce the assessment variability. We propose an automated gland segmentation system. Forty-three hematoxylin and eosin-stained images were acquired from prostate cancer tissue slides and were manually annotated for gland, lumen, periacinar retraction clefting, and stroma regions. Our automated gland segmentation system was trained using these manual annotations. It identifies these regions using a combination of pixel and object-level classifiers by incorporating local and spatial information for consolidating pixel-level classification results into object-level segmentation. Experimental results show that our method outperforms various texture and gland structure-based gland segmentation algorithms in the literature. Our method has good performance and can be a promising tool to help decrease interobserver variability among pathologists.

Keywords: AdaBoost, digital pathology, gland segmentation, prostate cancer, support vector machine

1. Introduction

The prostate gland is a functional conduit that allows urine to pass from the urinary bladder to the urethra and adds nutritional secretions to the sperm to form semen during ejaculation. Microscopically, the prostate is composed of glandular epithelium and fibrous stroma. Prostate cancer (PCa) occurs mostly in gland cells and hence is called prostate adenocarcinoma.

Gleason grade of prostate adenocarcinoma is an established prognostic indicator that has stood the test of time.1,2 The Gleason system is based on the glandular pattern of the tumor. Architectural patterns are identified and assigned a pattern from 1 to 5, with 1 being the most differentiated and 5 being undifferentiated.2,3

Increasing Gleason grade is directly related to a number of histopathological end points, including lympho-vascular space invasion by carcinoma, tumor size, positive surgical margins, and pathological stage, which includes risk of extraprostatic extension and metastasis. Histological grade of prostatic carcinoma is one of the most powerful, if not the dominant, predictor of clinical outcome for patients with this cancer. A number of clinical end points have been linked to histological grade. These clinical end points include clinical stage, response to different therapies, prostatic-specific antigen (biochemical) failure, progression to metastatic disease, PCa-specific survival, and overall survival.2,47

Certain areas of the Gleason system have been modified over the years. Some aspects in the original scoring are interpreted differently in modern practice. Due to the changes in the Gleason scoring over the years, it is now known that a specimen that has received a Gleason score of 1+1=2 in the area of Gleason would today be referred as adenosis (atypical adenomatous hyperplasia).8 It is therefore a grade that should not be diagnosed regardless of the type of specimen, with extremely rare exceptions. Also, a diagnosis of a lower score on needle biopsy is inaccurate because of

  • 1.

    poor reproducibility among experts,

  • 2.

    poor correlation with prostatectomy grade, with the majority of cases exemplifying higher grade at resection, and

  • 3.

    the possibility of misguiding clinicians and patients into believing that the patient has an indolent tumor.

Hence, clinicians are no longer using these Gleason 1 and 2 cancers in managing PCa. As for Gleason 5, it is easily identified by pathologists, and hence in our study, we did not focus on it. Instead, our study concentrates on the difficulties and reliability in identifying Gleason 3 and 4 cancers, which will subsequently affect both pathology and actual clinical management of PCa.

In the International Society of Urological Pathology 2005 Modified Gleason System, Gleason pattern 3 includes discrete glandular units, glands that infiltrate in and among nonneoplastic prostate acini and showing marked variations in size and shape. Pattern 4 exhibits ill-defined glands with poorly formed glandular lumina, cribriform, and fused microacinar glands. Pattern 4 also shows hypernephromatoid structures.9

The Gleason grading system, like all histological grading methods, possesses an inherent degree of subjectivity. Intra-and interobserver variability do exist. Recent data suggest that, for needle biopsy grading, pathologist training and experience can influence the degree of interobserver agreement.1012 Manual diagnosis of PCa is a tedious and time-consuming process where the pathologist examines a prostate tissue under a microscope. Moreover, the diagnostic accuracy depends on the personal skill and experience of the pathologist.

While the Gleason system has been in use for many decades, its dependence on qualitative features leads to a number of issues with standardization and variability in diagnosis. Intermediate Gleason grades (3 or 4) are often under- or overgraded, leading to interobserver variability up to 48% of the time.12 The main reason for undergrading (score 6 versus 7) was not recognizing relevant fusion of glands as a criterion for pattern 4.13

These problems motivate the research and development of assisting the diagnosis and prognosis processes through the use of software.14 Effective assistive diagnosis and prognosis tools have to be built in stages. First, basic tools, such as automated prominent nucleoli detection and gland segmentation tools, need to be developed. Second, these tools need to be put together in an effective way. We developed an effective prominent nucleoli detector in our previous work.15 In this paper, we present an effective method for gland segmentation. Gland segmentation is the first step for the analysis of gland shapes and architecture. We illustrate the main challenges we encountered in gland segmentation in a sample Gleason pattern 3 image in Fig. 1(a). The original hematoxylin and eosin (H&E) image I0 overlaid by annotation image G and the exhibiting potential issues in segmentation are shown in this figure.

Fig. 1.

Fig. 1

(a) Example of Gleason pattern 3 H&E image, I0. We also show the annotations (image G) overlaid on I0. For image G, red: gland, blue: lumen, and green: periacinar retraction clefting. Both the gland section with discontiguous lumina and the gland section with no visible lumen generally result from tangential sectioning of well-formed glands. We illustrate these two cases by the schematic diagram of gland tissue with example section positions in the middle. (b) Overview of our gland segmentation system.

Glands are inherently three-dimensional (3-D) structures, but, when processed and sliced, we can only analyze the cross section of these 3-D structures. Glands are generally composed of a lumen in the center with epithelial nuclei at the boundary and cytoplasm filling in the intermediate space. Lumen can be identified as empty white spaces within the glandular structure. Periacinar retraction clefting resulting in clear spaces around the glandular area may be seen on the histological material as a random phenomenon in tissue fixation. These cleftings arise from tissue handling during the biopsy procedure, undue delay in fixation, and tissue fixation itself. These periacinar retraction clefting are also white and are very similar to lumen. Some of the white areas may be masked by other cellular components, such as blood cells, which may impede automated identification of lumen and periacinar retraction clefting objects. Tangential sectioning of well-formed glands leads to gland sections with no visible lumen and/or with discontiguous lumina in the histological tissue slides. These issues arise when the tangential section plane misses the lumen or passes through the lumen multiple times (see Fig. 1). The varying gland shapes and sizes impede segmentation algorithms, which rely on template shapes and structures. We have illustrated all of these issues in Fig. 1(a).

In this study, we aim to develop an automated system for gland segmentation in PCa tissue images of H&E-stained histology slides. We used a machine learning and image processing-based gland segmentation method. Machine learning methods used pixel neighborhood and object shape information for an empirical estimate of various regions of the image. Image processing methods were used to further refine these approximations to compute the final segmentation result. The workflow of our method is as follows:

  • 1.

    Tissue images are manually extracted from whole slide images. These images are then segmented by our automated gland segmentation system.

  • 2.

    Periacinar retraction clefting and lumen regions are automatically identified in the extracted tissue images.

  • 3.

    All the glands lined by epithelial cells and surrounded by stromal cells are automatically segmented for the final result.

2. Related Work

Existing gland segmentation methods generally use texture and/or gland structure information. Farjam et al.16 used variance and Gaussian filters to extract roughness texture features in prostate histopathological images. They used k-means clustering for pixel-wise classification for nucleus and cytoplasm. This pixel information was then used for gland segmentation. Texture-only based approaches generally do not have access to lumen and epithelial nuclei’s spatial information and may perform badly, as discussed in Sec. 4.

To alleviate this problem in gland segmentation, algorithms have been developed to use gland morphology and lumen–epithelial nuclei relationships. Naik et al.17 used a level set method for gland segmentation. The initial level set is detected by a Bayesian classifier and surrounded the lumen region. The energy functional was developed for attaining minimum energy when the level set is around the epithelial nuclei. Xu et al.18 proposed a faster version of Naik et al.17 after modifying the energy functional in the geodesic general active contour model. These level set methods, when initialized correctly, would result in a reasonable segmentation. This property also defines their main limitation, which is level set initialization. These algorithms would also result in inaccurate segmentation due to incorrect or missing initial level sets.

Gunduz-Demir et al.19 proposed using object graphs for segmentation of colon histopathology images. Their algorithm first detects circular objects in the images and then identifies them as nuclei and lumen using object-graph information. A boundary of the gland was then constructed by connecting the centroids of nuclei objects. The identified lumen objects are then grown until they touch the gland boundary. This method will segment glands with lumen inside them but would fail to properly segment glands with no visible lumen and discontiguous lumina.

Nguyen et al.20 used k-nearest neighbor classification based on CIELab color space information to classify pixels as stroma, lumen, nuclei, cytoplasm, or blue mucin. This method then unified the nuclei and cytoplasm pixels to extract gland boundaries. Similarly, Monaco et al.21 used luminance channel in the CIELab color space in conjunction with Gaussian kernels to find the gland centers. These gland centers were then expanded into glands using a region growing method22 in the RGB color space. This method will segment glands with lumen inside them but would fail to segment glands with no visible lumen and discontiguous lumina.

An improvement of Nguyen et al.20 was suggested in Ref. 23 by finding nuclei and lumen object first and then using a local spatial search to connect the lumen to their nearby nuclei. Various points at the lumen boundary were sampled, and all the nearest nuclei points were connected to these points. The final gland segmentation was defined as the convex hull of various connected components, with each connected component being defined as a gland. Various lumen–nuclei connections were pruned using simple rules for reducing noise in the final result. This method will segment glands with lumen inside them but would fail to segment glands with no visible lumen and discontiguous lumina.

Nguyen et al.24 proposed a nuclei-based gland segmentation method to address both issues of gland sections with no visible lumen and discontiguous lumina by building a graph of nuclei and lumina in an image. This paper then used the normalized cut method25 to partition the graph into different components, each corresponding to a gland. The gland boundaries were then created by connecting nearby nuclei by straight lines. As such, the resultant glands would most likely cover most of the actual gland region. On the other hand, the resultant gland boundaries would be polygonal in shape, which may not follow the exact continuous shape of the gland.

Ren et al.26 also proposed a region-based nuclei segmentation method. This paper used color map normalization27 and subsequent color deconvolution28 to get separate masks for nuclei and gland regions. This paper addressed the issues of gland sections with discontiguous lumina and no visible lumen by not using lumen as prior information. Local maxima points in the distance transform map for the binary gland mask were used to identify individual glands. Nuclei regions were grouped to these local maxima points, and their Delaunay triangulation was used for output gland segmentation. This method relied on local maxima points to identify individual glands objects. The output gland boundaries are polygonal in shape, which like Nguyen et al.24 will not follow the continuous shape of gland.

Sirinukunwattana et al.29 also proposed a solution for a similar problem of gland segmentation in colon histopathological images by suggesting a random polygons model. This model treats each glandular structure as a polygon of a random number of vertices, which in turn represent approximate locations of nuclei. Their algorithm used a reversible-jump Markov chain Monte Carlo (RJMCMC) method to infer all the maximum a posteriori polygons. This method relies on using pixel-based identification of lumen regions to infer the initial random polygons for RJMCMC. It addresses the issue of segmenting glands with discontiguous lumina by filtering the converged maximum a posteriori polygons using a heuristic criterion. The method then proceeds to rerun RJMCMC with the remaining maximum ‘a posteriori’ polygons as initial polygons. The subsequent polygon-shaped glands are then improved upon by cubic spine interpolation such that the segmented glands have a more similar shape to the actual glands. This paper, however, does not address the issue of segmenting glands with no visible lumen explicitly.

Various deep learning-based methods have also been developed for similar segmentation problems in various types of histopathological images.3036 Xu et al.33 proposed a multichannel neural network for gland segmentation in colon histopathological images. Their method comprised three neural network channels. One channel was responsible for segmenting foreground pixels from background pixels. The second channel was responsible for detecting gland boundaries while the third channel was designed for gland detection. The outputs of these three channels were fused together by a convolutional neural network (CNN) for the final gland segmentation result. This paper, however, does not address the issue of segmenting glands with no visible lumen and discontiguous lumina explicitly.

Chen et al.34 proposed a deep contour-aware network (DCAN) for gland segmentation in colon histopathological images. This network was designed to find the gland regions and gland boundaries (contours) simultaneously. DCAN was trained after initializing its weights, which were defined according to a DeepLab37 model trained on the PASCAL VOC 2012 dataset.38 DCAN fused the results for gland regions and their boundaries by a simple pixel-based binary operation. This paper, however, does not address the issue of segmenting glands with no visible lumen and discontiguous lumina explicitly by focusing on detecting gland boundaries along with gland regions.

Ren et al.35 focused on the gland classification task while proposing a CNN for binary gland segmentation in prostate histopathological images. This CNN was designed in an encoding and decoding architecture with 10 layers in both of them. Using this approach, it was able to improve upon Nguyen et al.23 and Ren et al.26 on a dataset of 22 H&E images, each of 2400×1800  pixels, in the gland segmentation task. This paper, however, does not address the issues of segmenting glands with no visible lumen and discontiguous lumina explicitly.

Jia et al.36 proposed a weakly supervised CNN framework with area constraints for segmenting histopathological images. The CNN was trained and tested on two separate datasets of colon histopathological images. Their main contribution is the introduction of area constraints as weak supervision. The dataset was created by annotating cancerous and noncancerous regions along with a rough estimation (in term of percentage area) of the relative size of the cancerous regions in each image. This rough estimate acted as a weak constraint for the proposed CNN. This paper, however, does not address the issues of segmenting glands with no visible lumen and discontiguous lumina explicitly as it was more focused on binary segmentation of cancerous and noncancerous regions.

3. Method

We propose a machine learning and image processing-based method to solve the gland segmentation problem while addressing the issues illustrated in Fig. 1(a). Figure 1(b) shows an overview of our gland segmentation system. We developed a set of pixel classifiers and object classifiers for identification of various regions in an image. Our classifiers are trained on the data extracted from the manual annotations of the pathologist. The trained pixel and object classifiers are tested on a given image for an estimate of various regions. Image processing methods are then used for refining these region estimates into the final segmentation result.

3.1. Obtaining Manual Annotations for Training Pixel and Object Classifiers

We describe our image acquisition and manual annotation steps below.

3.1.1. Annotation of images into gland, lumen, periacinar retraction clefting, and stroma regions by the research pathologist

The whole slide tissue images of prostate biopsies for 10 patients with PCa were downloaded from The Cancer Genome Atlas (TCGA).39 The patients in this data set exhibit Gleason pattern 3 and/or 4. A total of 43 images from these whole slide images were extracted. Out of 43 images, 36 images were of 2720×2048  pixels and 7 images were of 1360×1024  pixels dimensions. All of these images were extracted at 40× magnification with pixel resolution of 0.25  μm per pixel. Scanner and source institution information for these patients was not available on TCGA website. We give details of these TCGA images in Table 1. After image extraction, the annotation workflow of the research pathologist is as follows:

Table 1.

Description of extracted images.

Image Gleason grade Bounding box (xmin,ymin),(xmax,ymax)
TCGA-2A-A8VO-01Z-00-DX1_L72702_T66921_W2720_H2048 3 (72,702, 66,921), (75,422, 68,969)
TCGA-2A-A8VO-01Z-00-DX1_L79545_T35778_W2720_H2048 3 (79,545, 35,778), (82,265, 37,826)
TCGA-2A-A8VO-01Z-00-DX1_L81449_T55250_W2720_H2048 3+4 (81,449, 55,250), (84,169, 57,298)
TCGA-2A-A8VO-01Z-00-DX1_L82247_T40015_W2720_H2048 3 (82,247, 40,015), (84,967, 42,063)
TCGA-2A-A8VO-01Z-00-DX1_L85502_T58665_W2720_H2048 3 (85,502, 58,665), (88,222, 60,713)
TCGA-HC-7075-01Z-00-DX1_L11216_T3583_W2720_H2048 3+4 (11,216, 3583), (13,936, 5631)
TCGA-HC-7075-01Z-00-DX1_L12975_T10686_W2720_H2048 3 (12,975, 10,686), (15,695, 12,734)
TCGA-HC-7075-01Z-00-DX1_L13921_T14013_W2720_H2048 3 (13,921, 14,013), (16,641, 16,061)
TCGA-HC-7075-01Z-00-DX1_L18204_T7968_W2720_H2048 3 (18,204, 7968), (20,924, 10,016)
TCGA-HC-7075-01Z-00-DX1_L25396_T4657_W2720_H2048 3+4 (25,396, 4657), (28,116, 6705)
TCGA-HC-7077-01Z-00-DX1_L31247_T4210_W1360_H1024 3+4 (31,247, 4210), (32,607, 5234)
TCGA-HC-7077-01Z-00-DX1_L33310_T6060_W1360_H1024 3 (33,310, 6060), (34,670, 7084)
TCGA-HC-7077-01Z-00-DX1_L34840_T5090_W1360_H1024 3 (34,840, 5090), (36,200, 6114)
TCGA-HC-7077-01Z-00-DX1_L35090_T6525_W1360_H1024 3 (35,090, 6525), (36,450, 7549)
TCGA-HC-7077-01Z-00-DX1_L36272_T3010_W1360_H1024 3 (36,272, 3010), (37,632, 4034)
TCGA-HC-7211-01Z-00-DX1_L30692_T17413_W1360_H1024_pattern4 4 (30,692, 17,413), (32,052, 18,437)
TCGA-HC-7211-01Z-00-DX1_L31053_T21217_W1360_H1024_pattern4 4 (31,053, 21,217), (32,413, 22,241)
TCGA-HC-7212-01Z-00-DX1_L23123_T10332_W2720_H2048_pattern3 3+4 (23,123, 10,332), (25,843, 12,380)
TCGA-HC-7212-01Z-00-DX1_L17555_T17714_ W2720_H2048_pattern4 4 (17,555, 17,714), (20,275, 19,762)
TCGA-HC-7212-01Z-00-DX1_L2051_T6465_ W2720_H2048_pattern4 4 (2051, 6465), (4771, 8513)
TCGA-HC-7212-01Z-00-DX1_L9643_T2160_ W2720_H2048_pattern4 4 (9643, 2160), (12,363, 4208)
TCGA-HC-7212-01Z-00-DX1_L6955_T4656_ W2720_H2048_pattern4 4 (6955, 4656), (9675, 6704)
TCGA-XJ-A9DQ-01Z-00-DX1_L58543_T10148_ W2720_H2048 3 (58,543, 10,148), (61,263, 12,196)
TCGA-G9-6364-01Z-00-DX1_L36428_T45944_ W2720_H2048 3+4 (36,428, 45,944), (39,148, 47,992)
TCGA-G9-6364-01Z-00-DX1_L43716_T42632_ W2720_H2048 3+4 (43,716, 42,632), (46,436, 44,680)
TCGA-G9-6365-01Z-00-DX1_L83292_T39235_ W2720_H2048 3+4 (83,292, 39,235), (86,012, 41,283)
TCGA-G9-6365-01Z-00-DX1_L84869_T37165_ W2720_H2048 3+4 (84,869, 37,165), (87,589, 39,213)
TCGA-G9-6364_L151748_T35850_W2720_ H2048 3+4 (151,748, 35,850), (154,468, 37,898)
TCGA-G9-6364_L131990_T52007_W2720_ H2048 3+4 (131,990, 52,007), (134,710, 54,055)
TCGA-G9-6364_L122615_T48695_W2720_ H2048 3+4 (122,615, 48,695), (125,335, 50,743)
TCGA-G9-6365_L80439_T8350_W2720_ H2048 3+4 (80,439, 8350), (83,159, 10,398)
TCGA-G9-6365_L83152_T15462_W2720_ H2048 3+4 (83,152, 15,462), (85,872, 17,510)
TCGA-G9-6365_L69172_T28814_W2720_ H2048 3+4 (69,172, 28,814), (71,892, 30,862)
TCGA-G9-6385_L23690_T17439_W2720_ H2048 3+4 (23,690, 17,439), (26,410, 19,487)
TCGA-G9-6385_L29600_T19834_W2720_ H2048 3+4 (29,600, 19,834), (32,320, 21,882)
TCGA-G9-6385_L26699_T26585_W2720_ H2048 3+4 (26,699, 26,585), (29,419, 28,633)
TCGA-G9-6385_L23524_T23766_W2720_ H2048 3+4 (23,524, 23,766), (26,244, 25,814)
TCGA-G9-6385_L26359_T23758_W2720_ H2048 3+4 (26,359, 23,758), (29,079, 25,806)
TCGA-XJ-A83H-01Z_L67058_T32077_ W2720_H2048 3+4 (67,058, 32,077), (69,778, 34,125)
TCGA-XJ-A83H-01Z_L69605_T30039_ W2720_H2048 3+4 (69,605, 30,039), (72,325, 32,087)
TCGA-XJ-A83H-01Z_L50723_T22263_ W2720_H2048 3+4 (50,723, 22,263), (53,443, 24,311)
TCGA-XJ-A83H-01Z_L55225_T17149_ W2720_H2048 3+4 (55,225, 17,149), (57,945, 19,197)
TCGA-XJ-A83H-01Z_L68817_T9948_ W2720_H2048 3+4 (68,817, 9948), (71,537, 11,996)
  • 1.

    Segmentation of nontissue regions that appeared white in color, and these regions were labeled as periacinar retraction clefting.

  • 2.

    Identification and segmentation of the glands lined by epithelial cells. Whether the glands appeared discrete, variable in size and shape, or fused were properly classified and annotated. If somewhat touching (not fused), glands have a discernible boundary, then they are annotated as separate glands.

  • 3.

    Identification of the lumen region(s) of the gland. Note that lumen regions also appear white in the image. Special care has to be taken to prevent confusion between periacinar retraction clefting and lumen objects.

All the regions in the images that were not annotated as either parts of lumen or gland or periacinar retraction clefting are considered stroma. We show a sample H&E image and its corresponding manual annotation in Fig. 1(a). We use these manually annotated labels for training various pixel classifiers. Similarly, the annotated lumen and periacinar retraction clefting objects are used to train the periacinar retraction clefting versus lumen object classifier.

3.1.2. Distribution of images for training and subsequent testing

We did a twofold cross-validated study to gauge the performance of our gland segmentation system. The first fold has 24 images (from six patients) for training and 19 images (from four patients) for testing. The second fold has 19 images (from four patients) for training and 24 images (from six patients) for testing. For both folds, the training images are used to extract the pixel- and object-level data to train the pixel and object classifiers.

3.2. Pixel Classifiers

The manual annotations define four types of pixels, i.e., gland, stroma, lumen, and periacinar retraction clefting. They are used to extract the training data for the pixel classifiers. Various machine learning methods, such as k-means clustering, k-nearest neighbor classification, random forest, and support vector machine (SVM)40 along with features generated from RGB color space, CIELab color space, texture, and pixel neighborhood, have been proposed in the literature for pixel-level identification.1621,23,24,29 Similarly, Yap et al.15 also proposed using different intensity gradient-based features derived from various color spaces, including RGB and CIELab along with SVM,40 standard logistic regression (LR), and AdaBoost,41 for prominent nucleoli detection; they also showed that usage of various color spaces emphasizes multiple aspects of pixel-level patterns from various regions in the histopathological images. In our pixel classifier, we use a similar approach of LR in various color spaces along with AdaBoost41 for the classifier combination. We discuss the generation of LR classifier ensemble and the subsequent combination for pixel classification later in this section.

3.2.1. Training data extraction from the annotated dataset

An image in our dataset contains at least a million pixels. It is then computationally infeasible to use all the pixels for training the pixel classifiers. Instead, we sampled a subset of pixels to train the pixel classifiers. Each sampled pixel is endowed with a feature that is derived from a variable image patch of size s×s pixels centered at that pixel. We define the periacinar retraction clefting and lumen regions as cavity regions as both of them appear white in the images. We also define gland and stroma regions as noncavity regions for their tissue content.

Sampling cavity pixels and noncavity pixels

All the pixels whose corresponding square patches are in either periacinar retraction clefting or lumen are identified as cavity pixels. All the pixels whose corresponding patches are in either gland or stroma are identified as noncavity pixels. This definition of cavity and noncavity pixels is used for sampling the training dataset. The pixels that were very near the boundary of cavity and noncavity regions were not sampled for training purposes. Figure 2 shows an example of an annotated image G and the corresponding cavity pixels and noncavity pixels in the cavity mask. We sampled 36,000 cavity pixels and noncavity pixel pairs to train the cavity versus noncavity pixel classifier.

Fig. 2.

Fig. 2

Gland and stroma regions are extracted from image G to the gland mask (right). Cavity and noncavity regions are extracted from image G to the cavity mask (left). Lumen edge and periacinar retraction clefting edge regions are extracted from image G to the lumen edge mask (bottom). We also separate the binary groupings in the three masks to show each grouping separately in six images at the bottom. In all of these images, the inactive pixels are shown in a stripe pattern. Active pixels are shown using the RGB values of the original image. In the six pixel grouping images, only the active pixels are used for sampling a particular group (type) of pixels. We also indicate the total number of pixels sampled from each grouping for training of pixel classifiers.

Sampling lumen edge pixels and periacinar retraction clefting edge pixels

We define another pixel classifier to extract neighborhood information of lumen and periacinar retraction clefting objects. Both types of objects have different characteristic surrounding regions. Lumen is always inside the gland and surrounded by cytoplasm. Periacinar retraction clefting may have any type of surrounding region.

All the pixels at the edge of lumen are considered lumen edge pixels, and all the pixels at the edge of periacinar retraction clefting are considered periacinar retraction clefting edge pixels. Figure 2 shows an example of an annotated image G and the corresponding lumen edge pixels and periacinar retraction clefting edge pixels in the lumen edge mask. We sampled 36,000 lumen edge pixels and periacinar retraction clefting edge pixel pairs to train the lumen edge versus periacinar retraction clefting edge pixel classifier.

Sampling gland and stroma pixels

All the pixels with their corresponding square patch inside the gland are considered gland pixels. All the pixels whose corresponding square patch is inside the stroma are considered stroma pixels. This definition of gland and stroma pixels is used for sampling the training dataset. The pixels that were very near the boundary for gland or stroma were not sampled for training purposes. Any pixel/information from the lumen and periacinar retraction clefting regions is not used. Figure 2 shows an example of an annotated image G and the corresponding gland pixels and stroma pixels in the gland mask. We sampled 36,000 gland pixels and stroma pixel pairs to train the gland versus stroma pixel classifier.

3.2.2. Training of pixel classifiers

We designed three pixel classifiers, i.e., cavity versus noncavity pixel classifier, gland versus stroma pixel classifier, and lumen edge versus periacinar retraction clefting edge pixel classifier, as part of our gland segmentation system. They are trained in two stages. In the first training stage, an ensemble of LR classifiers is trained using the training data. In the second stage, this ensemble of LR classifiers is combined using AdaBoost.41 We defined two disjoint subsets of training images (for a given fold) Setweak and SetAdaBoost corresponding to the two training stages.

For a given pixel classifier, the LR pixel classifier ensemble was trained on 20,000 pixels pairs sampled from Setweak, and the AdaBoost41 combination was trained on 16,000 pixel pairs sampled from SetAdaBoost. As such, we sampled a total of 36,000 cavity pixel, noncavity pixel pairs; 36,000 gland pixel, stroma pixel pairs; 36,000 lumen edge pixel, periacinar retraction clefting edge pixel pairs for training. We were able to use a total of 72,000 pixels (=36000×2) within a feasible training time for each of the three pixel-level classifiers.

Ensemble of logistic regression pixel classifiers

We use the standard LR with the L2 penalty (regularized) and the gradient descent method to train our LR pixel classifiers. We use the raw pixel values of pc(x,y) (c{1,2,3}) for a three-channel image patch p of s×s pixels centered at the given pixel as the input to the LR pixel classifier. We have a total of 3s2 features that are the input to the LR. The pixel classification score by LR is within the interval [0, 1] and is linearly mapped to [0, 255] to create a grayscale prediction image from all the pixel-wise scores.

The grayscale LR pixel classification score can be thresholded to create a binary classifier. The binary classifier will predict a positive sample for score value higher than the threshold and a negative sample otherwise. As such, we can generate 256 binary classifiers from a single LR pixel classifier. We term the set of all binary classifiers as weak classifiers.

Yap et al.15 showed that it is effective to extend the RGB color space to other color spaces viz. HLS, HSV, CIELab, CIELuv, XYZ, and YCbCr. This paper also proposed using different image patch sizes to capture information at different resolutions for a given data sample, i.e., pixel. The original images in RGB color space were digitally extended to the other six color spaces indicated above. We also used image patch sizes of 2×2, 4×4, 6×6, 8×8, 12×12, and 16×16  pixels for LR pixel classifier training. The LR training with all 20,000 pixel pairs for patch sizes of 8×8, 12×12, and 16×16  pixels can be time-consuming. We can reduce the training time of LR pixel classifiers with patch sizes of 8×8, 12×12, and 16×16  pixels using subsets of 1000, 2000, 4000, 8000, and 16000 pixel pairs from the total of 20,000 pixel pairs instead of using all pixel pairs. We were able to train the LR pixel classifiers with patch sizes of 2×2, 4×4, and 6×6  pixels using all the 20,000 pixel pairs within feasible time.

A weak binary classifier is considered to be unique according to image patch size (s×s), number of training samples, color space of input images that the underlying LR pixel classifier is trained on, and the threshold for the grayscale LR pixel classification score. We have a total of six image patch sizes (2×2, 4×4, 6×6, 8×8, 12×12, and 16×16), six possible subset cases (1000, 2000, 4000, 8000, 16000, and 20000 pixel pairs), seven color spaces (RGB, HLS, HSV, CIELab, CIELuv, XYZ, and YCbCr), and 256 (0 to 255) thresholds to define a unique weak binary classifier. As such, there are 64,512(=6×6×7×256) possible weak classifiers.

Out of all possible 64,512 weak binary classifiers, we were able to train 35,840 within reasonable computing time. We generated three ensembles of 35,840 weak binary classifiers for each of the three pixel classifiers. These three ensembles correspond to cavity versus noncavity pixel classification, gland versus stroma pixel classification, and lumen edge versus periacinar retraction clefting edge pixel classification.

Logistic regression pixel classifier ensemble combination by boosting

The results of weak classifiers were combined by AdaBoost.41 Each of the three ensembles was combined using the AdaBoost41 trained on the corresponding 16,000 pixel pairs sampled from SetAdaBoost images. The boosted combination result for a pixel is a weighted sum (H) of the votes by all the weak classifiers. A positive H indicates a positive sample prediction while a negative H indicates a negative sample prediction.

Our subsequent steps in the segmentation pipeline require the pixel-level classification output as a probability map, i.e., in the interval [0, 1]. We mapped the weighted sum H to [0, 1] using f(H)=1/1+eH. After mapping, the outputs for all three pixel classifiers are within the same interval of [0, 1]. This mapping also allows usage of image processing methods without any extra pixel value scaling operations.

3.3. Periacinar Retraction Clefting Versus Lumen Object Classifier

The manual annotations define periacinar retraction clefting and lumen objects. The annotated lumen and periacinar retraction clefting objects along with the two pixel classifiers, i.e., gland versus stroma pixel and lumen edge versus periacinar retraction clefting edge, are used to extract the training data for the object classifier.

3.3.1. Training data preparation for periacinar retraction clefting versus lumen object classifier

We used the neighborhood and shape-based information with SVM40 for object classification. Vanderbeck et al.42 proposed using morphological and textural features along with nuclei density for automated white region (regions that look like cavity objects in our dataset) classification in liver biopsy images. Nguyen et al.23 also proposed using morphological features for gland classification in histopathological images. We use similar feature extraction methods for our dataset and the periacinar retraction clefting versus lumen object classification problem.

Our object feature extraction process comprises three independent subprocesses. Two of these subprocesses pertain to the object’s neighborhood information while the third one pertains to shape (morphology) information.

We have defined gland versus stroma pixel to differentiate between gland and stroma pixels and lumen edge versus periacinar retraction clefting edge pixel classifier to differentiate between lumen edge and periacinar retraction clefting edge pixels. Both of these pixel classifiers use information in the square patch surrounding a given pixel. We wish to extract an object’s neighborhood information that comprises the distribution of different types of pixels.

Our above two pixel classifiers after training can aid in approximation of an object surrounding the pixels’ distribution information. The gland versus stroma pixel classifier can quantify the distribution of gland and stroma pixels in the neighborhood. The lumen edge versus periacinar retraction clefting edge pixel classifier does a more sophisticated prediction and provides the distribution information about the surrounding pixels if they belong to a lumen edge or periacinar retraction clefting edge. The lumen objects are more likely to be surrounded by gland (predicted) pixels and lumen edge (predicted) pixels. On the other hand, the periacinar retraction clefting objects will most likely be surrounded by stroma (predicted) pixels and periacinar retraction clefting edge (predicted) pixels. Both of the above pixel classifiers’ responses are hence important for quantification of the given object’s neighborhood information. Our two subprocesses (1) and (2) use these two pixel classifiers for extracting the object’s neighborhood information.

The third subprocess is computation of the given object’s shape descriptors. We illustrate the object feature extraction process in Fig. 3. In panel (d) of this figure, we illustrate the three subprocesses for feature extraction from an annotated lumen object. We marked the subprocess (1) for lumen edge versus periacinar retraction clefting edge classification based-subfeature, (2) for gland versus stroma classification based-subfeature, and (3) for shape descriptor subfeature. In both (1) and (2), a histogram is computed from the classifier responses. These histograms are subsequently L1-normalized such that the resultant bin height is a normalized count and the sum of all the bin heights is unity. In panel Fig. 3(d), for both (1) and (2), we first show the pixels that are tested by the corresponding pixel classifier followed by the classifier response and the L1-normalized histogram.

Fig. 3.

Fig. 3

Object feature extraction from manual annotations. (a) Example of an annotated image G overlaid over the original image. (b) All annotated lumen and periacinar retraction clefting objects in example image. We also show a zoomed-in image of a lumen object. (c) A lumen object (annotation) is overlaid over the original image. (d) Object feature extraction process (three independent subprocesses). In (1) and (2) subprocesses, lumen edge versus periacinar retraction clefting edge classifier and gland versus stroma classifier are used to test the pixels around the given object. All the inactive pixels (not tested by the pixel classifiers) are shown in a stripe pattern. In (3) subprocess, shape descriptors are computed for the object. The final concatenated object feature vector is of 173 (=130+34+9) dimensions. We show the number of dimensions of each computed value by the three subprocesses in parenthesis.

  • 1.

    All the pixels within the 27×27  pixels patch around any edge pixel are tested by a pretrained lumen edge versus periacinar retraction clefting edge pixel classifier. These classifier responses ([0,1]) are binned into a 128-bin L1-normalized histogram. The average value of these classifier responses and proportion of responses that are greater than 0.5 are also computed. This generates 130 feature values for the given object.

  • 2.

    All the pixels within the 9×9  pixels patch around any edge pixel and not within the annotated object are tested by a pretrained gland versus stroma pixel classifier. These classifier responses ([0,1]) are binned into a 32-bin L1-normalized histogram. If a pixel being tested is within the annotated stroma region, then its response is considered fixed at zero. The average value of these classifier responses and proportion of responses that are greater than 0.5 are also computed. This generates 34 feature values for the given object.

  • 3.

    We compute the object shape descriptors, i.e., Hu moments (seven-dimensional),43 solidity (ratio of the object area to that of its convex hull), and circularity (=4π×Area/Perimeter2). This generates nine feature values for the given object.

The subfeatures from (1), (2), and (3) are concatenated to create a 173 (=130+34+9)-dimensional object feature vector. For both (1) and (2), various values of square patch size and number of histogram bins were tried, and the combination with the best object classification performance was chosen for the final system. All the periacinar retraction clefting and lumen objects except the ones being cut at the border of the images were used to train the periacinar retraction clefting versus lumen object classifier.

3.3.2. Training of periacinar retraction clefting versus lumen object classifier

We used the lumen and periacinar retraction clefting objects in all the images of Setweak and SetAdaBoost to train our object classifier and ensure usage of the same images for training both object and pixel classifiers. Both Nguyen et al.23 and Vanderbeck et al.,42 along with proposing morphological features for object classification as discussed, also proposed using a SVM40 classifier. They finalized SVM40 for classification after comparing various classification methods, such as naive Bayes classifier, LR, decision trees, neural networks, etc., and using various subset of features. For our gland classification system, an object classifier SVM40 with radial basis function kernel, C=512, and γ=0.0195312 (found using cross-validation) was trained on the 173-dimensional feature space. The object classifier was trained on 810 lumina, 451 periacinar retraction clefting objects in the first fold, 553 lumina, and 222 periacinar retraction clefting objects in the second fold.

3.4. Testing of the Trained Pixel and Object Classifiers and Subsequent Usage of Image Processing Methods to Compute Final Segmentation from Their Predictions

Once we have trained the pixel and object classifiers, we can use them on the input test image for identifying various regions. This testing provides a rough estimate for various regions. These estimates are refined and improved upon by the image processing methods discussed in the following sections. We illustrate the complete segmentation process in Fig. 4. The input image is denoted by I0, and the subsequent images in the pipeline are denoted by an “I” (for binary images) and “I” (for nonbinary images) symbol with a subscript format. Binary images are single-channel monochrome images, while nonbinary images are single-channel grayscale or three-channel RGB images.

Fig. 4.

Fig. 4

Complete work flow for segmenting an image into gland, lumen, periacinar retraction clefting, and stroma regions. Input image is I0, and final segmentation is Spixel.

We used some of the basic image processing methods, including Gaussian blur, median blur, image thresholding, erosion, and dilation, in our segmentation pipeline. We use Gaussian blur to reduce noise in grayscale images. We used median blur along with erosion and dilation to reduce noise in binary images. A dilation operation was also used to enlarge small objects in our images as part of various algorithms discussed later in this section. For both Gaussian blur and median blur, we fixed the kernel size at 5×5  pixels. Image thresholding methods, including Otsu,44 moments preserving,45 percentile,46 maxentropy,47 and Shanbhag48 methods, were implemented. These image thresholding methods were used to generate binary images from the grayscale images in the segmentation pipeline. We tried various combinations of thresholding methods and Gaussian kernels with different values of σ at various stages of the pipeline. The configuration with the best performance was chosen for the final system.

3.4.1. Image processing methods to compute final segmentations using pixel and object classifiers’ prediction

Given the input test image I0, all the pixels in the image are tested by a cavity versus noncavity pixel classifier. The pixel-wise classification score is within the interval [0, 1], and this is linearly mapped to [0, 255] to create a grayscale prediction image. This prediction image is blurred by a Gaussian kernel of σ=10 and thresholded by Otsu’s method44 to separate the cavity and noncavity regions. The resultant binary mask Icavity indicates cavity regions in white on black noncavity regions (see Fig. 4). The Gaussian blur reduces noise in the prediction (cavity versus noncavity) image and prevents various false positive (FP) (small) cavity objects in Icavity.

The original image I0 is masked by Icavity. All the white pixels of Icavity are set inactive to create Igland_mask. The remaining active pixels in the image are tested by the gland versus stroma pixel classifier. These predictions are also mapped into a grayscale prediction image. This prediction image is blurred by a Gaussian kernel of σ=2 to create I1 (see Fig. 4). In I1, the brighter grayscale regions indicate glands, while darker grayscale regions indicate stroma with inactive pixels being indicated by a stripe pattern. During the method development, we found that a slight amount of Gaussian blur with σ=2 was needed to enhance the gland boundaries in the prediction image and to compute I1. I1 is then thresholded by a moment preserving method45 to separate the gland and stroma regions.

The grayscale values of I1 in the identified gland regions are copied over to another image I2. In I2, all stroma, periacinar retraction clefting/lumen regions are inactive with their pixel intensity set at zero. We indicate these inactive regions by a stripe pattern and the remaining active pixels by a grayscale value (see Fig. 4).

The binary mask Icavity contains all the detected cavity objects. I2 contains the neighborhood information of these cavity objects. Object shape information from Icavity and neighborhood information from I2 are used for the classification of cavity objects as either lumen or periacinar retraction clefting. The object classification result is stored in Ilumen_clefting (see Fig. 4), where lumen and periacinar retraction clefting objects are indicated in blue and green, respectively, on a white background.

We augment I2 by putting all identified lumen pixels as white (255) and all identified periacinar retraction clefting /stroma pixels as black (0). The resultant image is Igland_map (see Fig. 4). This grayscale gland probability map is a rough estimate of glandular regions as the gland boundaries of nearby/touching glands may not be prominent. Igland_map is thresholded by Otsu’s method44 to compute Iglandthreshold, indicating gland regions in white on a black background (see Fig. 4).

So far, we have identified gland, lumen, stroma, and periacinar retraction clefting regions. The next step is to further segment the gland regions into unique gland objects. The gland regions at this stage in the segmentation pipeline have many merged gland objects (as indicated by Iglandthreshold in Fig. 4). The epithelial nuclei generally line the glandular boundary; hence, once identified, they can aid in estimating gland boundaries. We use a seed-based segmentation algorithm to find the unique gland objects in Igland_map. These algorithms use the seed locations as initial objects. These initial objects are grown pixel wise into glands with each seed defining only one gland. Optimal positions for the seed for an object are generally inside its central region. Lumen objects being in this region are hence a natural choice for seeds. There are some glands sections with no visible lumen inside them. Nuclei being at the gland border can aid in finding a seed for these types of glands. There are also glands with discontiguous lumina. These lumen objects need to be merged into one seed to avoid oversegmentation. We use the nuclei objects’ and lumen objects’ spatial distribution to find unique seeds for glands with discontiguous lumina. We first discuss the nuclei object segmentation using H&E stain information.

Stain separation for identification of nuclei objects

Nuclei are stained in blue by the hematoxylin stain in the H&E (RGB) image. This H&E image is separated into hematoxylin (blue) and eosin (pink) channels using ImageJ plugin implementation of Ruifrok et al.49 The separated channels of eosin and hematoxylin are in grayscale, indicating the respective channel’s color intensity distribution. The hematoxylin grayscale image is denoised using Gaussian blur, thresholding, and median blur operations. The final binary mask Inuclei (see Fig. 4) indicates nuclei regions in white on a black background. The NUCLEI-SEGMENTATION algorithm shows this process in detail (Algorithm 1).

Algorithm 1.

Nuclei segmentation algorithm (NUCLEI-SEGMENTATION).

Input:
 - I0 is the original H&E image.
Output:
 - Inuclei is the binary nuclei mask. White regions are nuclei on a black background.
Function and variable description:
 - BLUR(I,σs) is the Gaussian blur function with kernel with σ=σs. Output is a grayscale image.
 - Stain separation method is described by Ruifrok et al.49
 - MOMENTS-THRESHOLD(I) thresholds I using the moments preserving image thresholding method.45 Output is a binary image.
 - Median filter (blur) replaces each pixel intensity by the median of all the pixel intensities in its kernel neighborhood.
Parameters:
 - Select Gaussian kernel with σ=3 for the blur operation. This value for sigma was optimized empirically.
1:  Procedure NUCLEI-SEGMENTATION(I0)
2:   Separate the haemotoxylin and eosin channels of I0 into IHematoxylin and IEosin.       ⊳ Both IHematoxylin and IEosin are grayscale images.
3:   IHematoxylinblur=BLUR(IHematoxylin,3).
4:   IHematoxylinthreshold=MOMENTSTHRESHOLD(IHematoxylin).
5:   Remove noise in IHematoxylinthreshold to compute Inuclei by median blur.
6:   return nuclei mask Inuclei.
7: end procedure
Seed generation for glands with no visible lumen using nuclei objects

The thresholded gland probability map Iglandthreshold and Inuclei are used to find the seeds for the gland with no visible lumen inside them. We illustrate this process in Fig. 5. The nuclei mask Inuclei is dilated 8 times using a 3×3  pixel dilation window to compute Inucleiborder. The nuclei border image Inucleiborder is subtracted from Iglandthreshold to extract nonnuclei gland regions and compute Imask (see Fig. 5). A Euclidean distance map is computed for Imask, and ImageJ’s “Find Maxima” function is used for finding intensity (distance) peaks in this distance map. These peaks approximate the centroids of the glands and hence can be used as seeds for these glands. The MAXIMA-SEEDS (Algorithm 2) algorithm shows this process in detail, and its output is Iseed_maxima (see Fig. 5, also shown in Fig. 4). We identify the seeds in Iseed_maxima as maxima seeds.

Fig. 5.

Fig. 5

Seed generation for glands without any visible lumen by MAXIMA-SEEDS (Algorithm 2).

Algorithm 2.

Maxima-seeds generation algorithm (MAXIMA-SEEDS).

Input:
 - Iglandthreshold is the binary gland mask. White regions are glands on a black background.
 - Inuclei is the binary nuclei-border mask. White regions are nuclei borders on a black background.
Output:
 - Iseed_maxima is the binary maxima-seed mask. White regions are centroid locations of the gland objects.
Function and variable description:
 - DILATE(I,d) conducts “d” dilation operations on binary image I with a 3×3 pixels dilation window. Output image is also binary.
 - DISTANCE-TRANSFORM(I) computes the Euclidean distance map of image I. Output image is grayscale.
 - FIND-MAXIMA(I) computes the local intensity maxima pixels in the I. Output is a binary image indicating the maxima pixels in white.
 - Given a binary image I, I¯ is a binary image with corresponding inverted pixel values.
 - IaIb is a Boolean operation defining pixel-wise AND operation between Ia and Ib.
Parameters:
 - Select number of dilation operations d=8. This value for d was optimized empirically.
1:  Procedure MAXIMA-SEEDS (Iglandthreshold, Inuclei)
2:   Inucleiborder=DILATE(Inuclei,8).
3:   Imask=IglandthresholdInucleiborder¯.    ⊳ Remove nuclei borders from gland regions.
4:   Iseed_maxima=FINDMAXIMA(DISTANCETRANSFORM[Imask)].    ⊳ Find the centers of the glandular objects.
5:   return maxima-seed mask Iseed_maxima.
6: end procedure
Seed generation for glands with discontiguous lumina using nuclei and lumen objects

The maxima seeds in Iseed_maxima and lumen objects in Ilumen_clefting provide different sets of seeds for segmentation. After incorporating both maxima seeds and lumen, some gland objects may have multiple potential seeds inside them. These potential seeds need to be merged into one unique seed to avoid oversegmentation.

We use an electric potential field-based approach for refining the gland seeds. We developed REFINE-SEEDS (Algorithm 3) for this task, illustrated in Fig. 6. REFINE-SEEDS algorithm treats lumen and maxima seeds as positively charged objects and the nuclei as negatively charged objects. We defined COMPUTE-FIELD algorithm (Algorithm 4) to compute the electric potential field distribution in an image according to the charge distribution. All lumen and maxima seeds inside a gland will be in a region with high magnitude of potential field due to their positive charge. Given the field image according to a given charge distribution, it can be thresholded to identify regions with high magnitude of field. The lumen objects and maxima seeds, which are inside one gland, will be inside one of the high-field magnitude regions. This identified region can act as a single unique seed for the corresponding gland and hence help prevent oversegmentation.

Algorithm 3.

Seed-refining algorithm (REFINE-SEEDS).

Inputs:
 - Iglandthreshold is the binary gland mask. White regions are glands on a black background.
 - Inuclei is binary the nuclei mask. White regions are nuclei on a black background.
 - Iseed_maxima is the binary maxima-seed mask. White regions are centroid locations of the gland objects.
 - Ilumen_clefting is the lumen-clefting RGB image. Blue objects are lumen and green objects are periacinar retraction clefting on a white background.
Output:
 - Iseed is the binary seed mask. White regions are seed objects on a black background.
Function and variable description:
 - PIXEL-IN(I) returns all the white pixels in the binary image I. DILATE(I,d) and DISTANCE-TRANSFORM(I) are the same functions as defined in Algorithm 2.
 - Given a binary image I, I¯ defines a binary image with corresponding inverted pixel values. IaIb and IaIb are the Boolean operations defining pixel-wise AND and OR operations between Ia and Ib, respectively.
 - pixels+ and pixels are two sets containing positive and negative charged particles (pixels).
 - MAX-ENTROPY-THRESHOLD(I) thresholds I using the entropy-based image thresholding method.47 Output is a binary image.
Parameters:
 - Select number of dilation operations d=8. This value for d was optimized empirically.
1:  Procedure REFINE-SEEDS (Iglandthreshold, Inuclei, Iseed_maxima, Ilumen_clefting)
2:   Inucleiborder=DILATE(Inuclei,8).
3:   Create binary lumen mask Ilumen from Ilumen_clefting.    ⊳ Ilumen indicate lumen in white.
4:   Sample some pixels along the edges of lumen objects (Ilumen) into pixels+.
5:   pixels+=pixels+PIXELSIN(Iseed_maxima).    ⊳ All the maxima seeds are added to pixels+.
6:   Extract approximate centroids of nuclei in Inuclei into pixels.    ⊳ Remove nuclei from the gland regions.
7:   Imask=IglandthresholdInucleiborder.
8:   Ifield=COMPUTE-FIELD(Iglandthreshold,pixels+,pixels).
9:   Ifield_map=Ifield×DISTANCETRANSFORM(Imask).    ⊳ Pixel-wise multiplication of two real valued images.
10:   Linearly rescale Ifield_map to values within [0, 255] for further numerical operations.
11:   Ifield_mapthershold=MAXENTROPYTHRESHOLD(Ifield_map).
12:   Fill holes in Ifield_mapthreshold image.
13:   Iseed=Iseed_maxima(Ifield_mapthresholdIlumen).    ⊳ Unify all the seeds.
14:   return seed mask Iseed.
15: end procedure
Fig. 6.

Fig. 6

Seed refinement using REFINE-SEEDS (Algorithm 3). The distribution of positively and negatively charged objects is shown in black box. Nuclei constitute negatively charged particles. Lumen and maxima seeds constitute positively charged objects.

Algorithm 4.

Field computation algorithm (COMPUTE-FIELD).

Inputs:
 - Iglandthreshold is the binary gland mask. White regions are glands on a black background. Field is computed inside gland regions only.
 - pixels+ is a set of pixels with an assigned positive charge.
 - pixels is a set of pixels with an assigned negative charge.
Output:
 - Ifield is the grayscale field image. Pixel intensity is directly proportional to field magnitude at its location. Pixel intensities are within [0, 255].
Parameters:
 - Define +ve and −ve charges as +1 and 1, respectively. Set decay rate for +ve and −ve charges as 1.0 and 1.6, respectively. These charge values and decay rates were optimized empirically.
1:  Procedure COMPUTE-FIELD(Iglandthreshold, pixels+, pixels)
2:   Initialize all pixels in Ifield to zero.
3:   for All pixels (x,y)Iglandthresholddo
4:   Ifield(x,y)=(xj,yj)pixels+(1.0)[(xxj)2+(yyj)2]12+(xk,yk)pixels(1.0)[(xxk)2+(yyk)2]1.62.
5:   end for
6:   Linearly rescale Ifield to values within [0, 255] for further numerical operations.
7:   return field image Ifield.
8: end procedure

Instead of using all pixels inside nuclei, lumen, or maxima seeds, we sample some pixels for computing electric potential field image, Ifield, thereby saving computational time. In REFINE-SEEDS (Algorithm 3), we first sample the set of positive pixels “pixels+” from the lumen edges and union it with the maxima seeds pixels. We extract approximate centroids of nuclei objects to compute the set of negative pixels, “pixels”. COMPUTE-FIELD algorithm (Algorithm 4) uses the pixel sets pixels+, pixels, and the binary gland mask Iglandthreshold to compute the image’s electric potential field image Ifield. Electric potential field for nongland regions is not computed.

We define a binary mask Imask that indicates the nonnuclei-border glandular regions. The Euclidean distance map of Imask will have low pixel values around the nuclei regions compared with high pixel values in the central region of the glands. Field map image Ifield_map is computed by pixel-wise multiplication of Ifield and the Euclidean distance map of Imask. It will have low pixel values around the nuclei regions and high pixel values in the central region of the glands. The multiplication operation enhances the electric field distribution image from Ifield to Ifield_map. The high electric potential field regions are identified by thresholding Ifield_map using Maxentropy image thresholding method.47 All the seeds from thresholded field image Ifield_mapthreshold, lumen object image Ilumen, and maxima-seed image Iseed_maxima are unified by a pixel-wise OR operation to compute Iseed (see Figs. 4 and 6).

Evolving generalized Voronoi diagram-based binary gland segmentation

We use the evolving generalized Voronoi diagram (EGVD)-based segmentation method to define gland boundaries of touching or nearby glands in Igland_map. The EGVD-based binary segmentation uses a seed (binary) image and a grayscale image. The grayscale image defines background regions in black and foreground (gland) regions by a grayscale value. The output binary segmentation is created by growing all the seed objects defined in the seed image until they touch each other or meet the background region as defined in the input grayscale image.50

We show an example of EGVD output image IEGVD when we use seed image Iseed and grayscale image Igland_map in Fig. 7(a). For a good gland segmentation result, the seed growing should stop at the gland boundaries lined by epithelial nuclei. The output glands should also not have their boundary defined within cytoplasm. The seed growing iterations in EGVD treat both cytoplasm and nuclei regions in the foreground equally. The black arrows on IEGVD in Fig. 7(a) show the gland boundaries being defined in cytoplasm. We also observe that some of the seeds in Iseed are extraneous and do not belong to any gland object. The bottom-left black arrow indicates the segmented gland that has been grown across nuclei and hence leads its boundary to be defined in cytoplasm. It also indicates one of the glands grown from the extraneous seed.

Fig. 7.

Fig. 7

(a) Example of EGVD output with Iseed and Igland_map as inputs. Top-right black arrows indicate gland boundaries being predicted within the cytoplasm. Bottom-left black arrow indicates one of the glands that was grown from extraneous seed. (b) Complete workflow for the THREE-LEVEL-EGVD algorithm (Algorithm 5).

We developed a THREE-LEVEL-EGVD method (Algorithm 5) to solve the issues of extraneous seeds and gland boundaries being predicted within cytoplasm. We illustrate the method in Fig. 7(b). We first put all nuclei border pixels indicated by Inucleiborder in Igland_map as black to compute Igland_temp. Iseed is used as the seed image and Igland_temp as the grayscale image to compute the binary EGVD result, IEGVDoutput1. Nuclei border pixels are put as black in Igland_temp to prevent the gland object from growing across the nuclei by EGVD. As black pixels are defined as background, EGVD will stop growing the seed objects once they touch the nuclei border regions. Hence, the grown seed objects will not cover the nuclei regions. IEGVDoutput1 is cleaned up by removing small gland objects and noise to compute IEGVDoutput1denoised. A pixel-wise OR operation is done on Iseed and IEGVDoutput1denoised to retain all the initial seeds in Iseed, for they may have been removed during previous denoising step. The resultant image is Iseed2.

Algorithm 5.

Three level EGVD algorithm (THREE-LEVEL-EGVD).

Inputs:
 - Iseed is the binary seed mask. White regions are seed objects on a black background.
 - Igland_map is the grayscale gland image. Grayscale regions are glands on a black background.
 - Inuclei is the binary nuclei mask. White regions are nuclei on a black background.
Output:
 - Iglandresult is the binary gland mask. White regions depict gland on a black background.
Function and variable description:
 - DILATE(I,d) conducts “d” dilation operations on binary image I with a 3×3 pixels dilation window. Output image is also binary.
 - EGVD(Is,If) is the image segmentation algorithm described by Yu et al.50 where Is is the seed image and If is the grayscale image. It returns the binary segmentation result.
 - IaIb is a Boolean operation defining the pixel-wise OR operation between Ia and Ib.
Parameters:
 - Select number of dilation operations d=8. This value for d was optimized empirically.
1:  Procedure THREE-LEVEL-EGVD(Iseed, Igland_map, Inuclei)
2:    Inucleiborder=DILATE(Inuclei,8).    ⊳ Inucleiborder indicates nuclei border pixels in white.
3:    Put all white pixels indicated by Inucleiborder in Igland_map as black to compute Igland_temp.
4:    IEGVDoutput1=EGVD(Iseed,Igland_temp).
5:    Remove small objects and noise in IEGVDoutput1 and then save it to IEGVDoutput1denoised.
6:    Iseed2=IEGVDoutput1denoisedIseed.
7:    IEGVDoutput2=EGVD(Iseed2,Igland_map).
8:    Remove small objects and noise in IEGVDoutput2 and then save it to IEGVDoutput2denoised.
9:    Erode IEGVDoutput2denoised to compute Iseed3.
10:   IEGVDoutput3=EGVD(Iseed3,Igland_map).
11:   Remove small objects and noise in IEGVDoutput3 and then save it to Iglandresult.
12:   return binary gland mask Iglandresult.
13: end procedure

In the second level of EGVD, Iseed2 and Igland_map are used as the seed image and the grayscale image, respectively. The EGVD result is IEGVDoutput2. As illustrated in Fig. 7(b), we can observe that the objects in IEGVDoutput2 cover a major portion of the respective gland regions. The extraneous seeds are generally in the stroma regions near the gland boundaries. The seed growing iterations for these seeds will lead to small gland objects as they touch the background regions or other gland objects earlier than the seeds, which are in the central regions of actual glands. The gland objects generated by extraneous seeds of Iseed2 are identified by their relatively smaller size and are removed from IEGVDoutput2. The resultant denoised EGVD output, IEGVDoutput2denoised, is eroded to compute the seed image for the third level, Iseed3.

In the third level of EGVD, Iseed3 and Igland_map are used as the seed image and the grayscale image, respectively. The EGVD result is IEGVDoutput3. The seed objects in Iseed3 cover a major portion of the respective gland regions. In many cases, the seed objects’ boundaries are next to the corresponding epithelial nuclei. These type of seed objects will lead the seed growing iterations to terminate within the nuclei regions as the neighboring grown seed objects would be touching each other when they grow into the nuclei regions. This step allows THREE-LEVEL-EGVD to define the gland boundaries in the nuclei regions and not in cytoplasm [see IEGVDoutput3 in Fig. 7(b)]. IEGVDoutput3 is then cleaned up similarly as in the previous steps to compute the final result Iglandresult.

Merging all detected gland, lumen, and periacinar retraction clefting objects for final segmentation

The binary gland output Iglandresult and object classification result Ilumen_clefting are merged to compute the final segmentation result Spixel. We have illustrated the complete segmentation process starting from pixel classifiers testing to the last EGVD-based segmentation step in Fig. 4.

As discussed in this section, the binary nuclei mask, Inuclei, that is computed by NUCLEI-SEGMENTATION (Algorithm 1) is used by algorithms MAXIMA-SEEDS (Algorithm 2), REFINE-SEEDS (Algorithm 3), COMPUTE-FIELD (Algorithm 4), and THREE-LEVEL-EGVD (Algorithm 5) either directly as input or indirectly as pixels. Inuclei may have noisy segmentation result when the nuclei are crowded and/or have margination of chromatin. The sole purpose of Inuclei is to provide an approximate location of gland boundaries and approximate location of negatively charged pixels in our gland segmentation pipeline. This pipeline has been developed such that it dilates the Inuclei to counter noisy nuclei segmentation. Along with number of dilation operations, the value for “−ve” charge and the corresponding decay rate in the COMPUTE-FIELD (Algorithm 4) were also empirically estimated to keep the gland segmentation pipeline robust to noisy nuclei segmentation. Another possible concern in our gland segmentation pipeline is how the color variations of a given image will affect the final result. We have used the images from 10 patients in TCGA dataset. These images were extracted such that they had color variations. Our pixel classifier training sets were sampled randomly from different patient images to ensure heterogeneity. We show four images from our dataset depicting four color variations in Fig. 8. We have also defined the usage of Ruifrok et al.49 in conjunction with tunable blur and threshold operations in NUCLEI-SEGMENTATION (Algorithm 1) to reduce the effect of color variations. The empirical estimation of various parameters in our segmentation pipeline was also done for robust performance.

Fig. 8.

Fig. 8

(a)–(d) Four images from our dataset illustrating different color variations. Our gland segmentation pipeline was developed using 43 images with different color variations for robust and generalizable performance.

4. Experiments and Results

4.1. Dataset and Experiments

H&E-stained images of the PCa were used to demonstrate the effectiveness of our method. This dataset consisted of 43 images extracted from 10 patients’ tissue whole slide images. These whole slide images were hosted by TCGA. We did a twofold cross-validation study of our gland segmentation system. We trained and tested our system as discussed in Sec. 3. We have compared our segmentation method with base-line methods as discussed by Farjam et al.,16 Naik et al.,17 and Nguyen et al.23 We implemented these methods and tuned the parameters to maximize their segmentation performance. We evaluated all the gland segmentation methods’ results according to our manual annotations. As discussed in Sec. 3.1.1, we annotated almost touching nearby glands according to their discernible boundaries. This lead to a design where we have not tried to incorporate any ambiguity with respect to boundaries between different glands.

4.1.1. Evaluation of image segmentation

We evaluate the segmentation performance using pixel-wise metrics, such as precision, recall, accuracy, F-score, and omega index (OI),51 and object-wise metrics, such as Jaccard index (JI), Dice index (DI), and Hausdorff distance (HD). We do image-wise comparison of the segmentation results. We denote a manual annotation image by G and a corresponding segmented image generated using method “t” by St.

We show an example H&E image I0, corresponding annotation G, and various segmentation results St in Figs. 9(a)9(f). Our method as shown in Spixel correctly identifies major portions of various gland objects defined in G. It is also able to segment glands with no visible lumen inside them and glands with discontiguous lumina inside them. The result image SNguyen for Nguyen et al.23 has many segmented lumen objects that are much larger than the actual lumen objects. This method was able to estimate the gland boundaries at the epithelial nuclei. The result image SNaik for Naik et al.17 shows many nearby glands being merged into one. This method also estimated the gland boundaries at the epithelial nuclei. The result image SFarjam for Farjam et al.16 shows that only the nuclei regions were identified as gland objects. This method predicted all the other regions as stroma.

Fig. 9.

Fig. 9

We compare various methods in (a)–(f). (a) Original H&E image I0, (b) annotation image G, (c) image Spixel by our method, (d) image SNguyen by Nguyen et al.,23 (e) image SNaik by Naik et al.,17 and (f) image SFarjam by Farjam et al.16 We also illustrate a failure case for our method in (g)–(i). (g) Another input H&E image, (h) corresponding annotation image, overlaid on image in (g), and (i) our result Spixel. All the annotation/segmented images are shown overlaid on the corresponding original image I0 and red: gland, blue: lumen, green: periacinar retraction clefting, and white: stroma.

Pixel-wise metrics

For each type of classification of gland, lumen, or periacinar retraction clefting, we can compute true positive (TP), FP, true negative (TN), and false negative (FN) pixels predictions. We can then report precision, recall, F-score, and accuracy using TP, FP, FN, and TN values. Higher values of these metrics indicate better performance of a given segmentation method.

The current image segmentation problem can also be viewed as a clustering problem where we want to cluster the pixels into gland, lumen, clefting, or stroma pixels. We can quantify the agreement between G and St by OI, which indicates the ratio of how many pairs of pixels out of all the pixels were put in the same cluster by both G and St. Given L pixels in image G, OI is computed by going through all the corresponding C2L pixel pairs. It is computationally infeasible to calculate OI from all the pixels in our dataset as each image has at least a million pixels. Hence, we approximated OI using a subset of pixels in the images. While sampling pixels, we ensure that we sample the same number of pixels from each pixel type for unbiased estimation. The number of sampled pixels in image G is constrained by the smallest total area spanned by a single pixel type. If the smallest number of pixels of one type in an image is lp, then we can only sample lp pixels from each of the four types in this image. The maximum possible value of lp was chosen as 2217 by going through all the images in the dataset. This was done to ensure OI computation is done using the same number of pixel pairs. We sampled 2217 pixels in each of the four pixel types, namely, gland, lumen, periacinar retraction clefting, and stroma, and then approximated OI by going through C28868 pixel pairs.

Object-wise metrics

We define three types of objects in our segmentations, namely, gland, lumen, and periacinar retraction clefting. We first find all the gland objects in G and St. All the gland objects in G are identified as G1,G2,,Gi,,Gm. We put these objects in OG such that GiOG. Similarly, all the gland objects in St are identified as S1t,S2t,,Sjt,,Skt. We put these objects in OSt such that SjOSt. All the objects Gi are matched to an object MiSjt (say) using the Gale–Shapley algorithm.52 The set of all matched objects is M, i.e., (Gi,Mi)M. We compute the object-wise JI, DI, and HD as

JI(Gi,Mi)=|GiMi||GiMi|, (1)
DI(Gi,Mi)=2|GiMi||Gi|+|Mi|, (2)
HD(Gi,Mi)=max{supxGiinfyMixy,supyMiinfxGixy}, (3)

where |·| denotes the number of pixels of a given object. Both JI and DI evaluate the overlap of two matched objects while HD evaluates the shape similarity between them. Higher values of HD indicate large deviations of shape between the two objects. For evaluating any segmentation algorithm’s result St at the object level, we need to evaluate

  • i.

    how well the segmented objects in OSt overlap the objects in OG and

  • ii.

    how well the ground truth objects in OG overlap the objects in OSt.

We compute weighted averages of JI, DI, and HD for a given image as follows:

JIweighted(OG,OSt)=12[j=1kwjJI(Gj,Sjt)+i=1mwiJI(Gi,Mi)], (4)
DIweighted(OG,OSt)=12[j=1kwjDI(Gj,Sjt)+i=1mwiDI(Gi,Mi)], (5)
HDweighted(OG,OSt)=12[j=1kwjHD(Gj,Sjt)+i=1mwiHD(Gi,Mi)], (6)

where wj=|Sjt|jk|Sjt| and wi=|Gi|im|Gi|. The first and second term in the above equations take care of the evaluation criterions (i) and (ii), respectively, as discussed above. These object-level metrics were derived from the ones discussed by Sirinukunwattana et al.29 Better segmentation algorithms will have higher values of JIweighted, and DIweighted and a lower value of HDweighted.

Some of the objects in OG and OSt will remain unmatched due to very low or no intersection with all the objects in the other set. A higher number of unmatched objects in OG and OSt indicates lower performance of the segmentation algorithm. We evaluate the object matching by computing the F-score, which is a harmonic mean of the percentage of objects matched in OG and the percentage of objects matched in OSt. The higher this object matching F-score is, the better is the segmentation algorithm. We also defined the metrics unmatched object area of G (UG) and unmatched object area of St (US) [see Eqs. (7) and (8)]. We calculate total area of all unmatched objects and then divide it by the area of the image G. Area of G and corresponding St will always be the same. A better segmentation algorithm will result in lower values for UG and US metrics.

UG=lArea of unmatched objectGlOGArea of Image (7)
US=lArea of unmatched objectSltOStArea of Image (8)
Comparison and discussion

The evaluation metrics discussed above were computed on the test results of our method (Spixel), active contour-based method (SNaik17), texture-based method (SFarjam16), k-means clustering, and nuclei–lumen association-based method (SNguyen23). For the active contour-based method by Naik et al.,17 the initial level set was defined from the white regions detected by our method. We compared the pixel-wise metrics of our method with the other three methods. We report image-wise averages of precision, recall, accuracy, and F-score for the three cases of gland, lumen, and clefting classification in Table 2. Image-wise averaged OI is reported in Table 2.

Table 2.

Pixel-wise metrics for testing. The reported values are average value ±standard error over the set of 37 images. Spixel: our method, SNguyen: k-means clustering and nuclei–lumen association-based method by Nguyen et al.,23 SNaik: active contour method by Naik et al.,17 and SFarjam: texture-based method by Farjam et al.16

Pixel type Method Precision Recall F-score Accuracy
(a) Pixel-wise metrics for each pixel type
Gland Spixel 0.64±0.017 0.90±0.004 0.74±0.01 0.72±0.01
SNguyen 0.66±0.015 0.30±0.02 0.39±0.02 0.59±0.01
SNaik 0.61±0.019 0.41±0.02 0.47±0.02 0.59±0.01
SFarjam 0.54±0.01 0.08±0.0003 0.15±0.005 0.54±0.01
Lumen Spixel 0.82±0.02 0.55±0.02 0.64±0.02 0.94±0.004
SNguyen 0.49±0.03 0.69±0.02 0.54±0.02 0.86±0.01
SNaik 0.65±0.02 0.62±0.02 0.60±0.01 0.90±0.01
SFarjam NA
Periacinar retraction clefting
Spixel 0.42±0.05 0.15±0.02 0.18±0.026 0.97±0.003
SNguyen NA
SNaik NA
SFarjam
NA
(b) OI for comparing all pixel types
Method
OI
Spixel 0.48±0.01
SNguyen 0.36±0.01
SNaik 0.36±0.01
SFarjam 0.18±0.01

Note: The best results are shown in bold.

Our method identified all four regions, i.e., gland, lumen, periacinar retraction clefting, and stroma, in the images. Hence, we show our pixel-wise metrics for the three regions in Table 2. Our method used pixel-level information for empirical estimates of the regions and then improved it using object-level information.

Nguyen et al.23 used k-means clustering to identify the lumen regions and nuclei regions first. After identification, the nuclei regions were associated to nearby lumen regions. The output segmented glands were computed using the convex hull of some pixels sampled from the associated lumen and nuclei regions. We observed many instances with identified lumen regions indicating surrounding cytoplasm as lumen region. The output convex hull may not follow the shape of gland as it would connect the adjacent nuclei points by a straight line. This lead to incorrect predictions of both gland and stroma pixels. Nguyen et al.23 also do not differentiate between lumen and periacinar retraction clefting regions, which explains this method’s low pixel-wise accuracy, F-score, and OI as illustrated in Table 2.

Our implementation of Naik et al.17 used the cavity objects provided by our method as the initial level set for the glands. These level sets after energy minimization converged around the epithelial nuclei. In the case of nearby and touching glands, the final level sets merged, leading to one gland instead of two or more adjacent gland segments. This merging lead to incorrect predictions for both gland and stroma pixels and, hence, the low pixel-wise F-score, accuracy, and OI as illustrated in Table 2.

Farjam et al.16 used a texture-based approach to first segment nuclei regions and then segmented stroma, cytoplasm, and lumen regions together. The nuclei regions were used to identify the gland boundaries to obtain gland regions. This method used k-means clustering in the grayscale images for identifying these regions. The nuclei regions were being identified correctly. In many instances, the combined stroma and lumen regions were not being identified properly. The identified stroma, cytoplasm, and lumen regions did not contain the actual lumen, stroma, or cytoplasm completely. This lead to most of the cytoplasm being predicted as background with only nuclei and surrounding pixels being predicted as gland regions. This explains the low pixel-wise F-score, accuracy, and OI as shown in Table 2.

We computed JIweighted, DIweighted, HDweighted, UG, and US for gland objects in each image result. We report the image-wise average of these object-wise metrics for all four methods in Table 3. Figure 10 illustrates the evaluated JI (JIweighted for gland objects), gland object matching F-score, and pixel-wise OI for all the images.

Table 3.

Object-wise metrics for testing (for gland objects). The reported values are average value ±standard error over the set of 37 images. Spixel: our method, SNguyen: k-means clustering and nuclei–lumen association-based method by Nguyen et al.,23 SNaik: active contour method by Naik et al.,17 and SFarjam: texture-based method by Farjam et al.16

Method JI (JIweighted) DI (DIweighted) UG US HD (HDweighted) Object matching F-score
Spixel 0.54±0.01 0.65±0.01 0.04±0.006 0.11±0.02 189.98±13.27 74.54±1.42
SNguyen 0.38±0.01 0.49±0.01 0.10±0.02 0.17±0.01 177.83±13.58 58.24±1.68
SNaik 0.26±0.02 0.37±0.03 0.20±0.02 0.02±0.002 497.04±57.47 51.26±3.14
SFarjam 0.01±0.0008 0.02±0.0002 0.00±0 0.12±0.004 228.46±14.55 1.101±0.078

Note: The best results are shown in bold.

Fig. 10.

Fig. 10

Results for all 37 images. (a) JI (JIweighted) for gland segmentation (object-wise metric), (b) F-score for gland object matching (%age) (object-wise metric), and (c) OI (pixel-wise metric). Object-wise results for texture-based method by Farjam et al.16 are not shown due to low performance.

Our method identified the lumen, periacinar retraction clefting, and nuclei regions based on the object-level information. The spatial distribution of lumen and nuclei regions was used to obtain the gland objects. The usage of lumen and nuclei objects for seed estimation allowed fewer false gland detections; hence, we obtained a high gland object matching F-score as shown in Table 3. Seed-based growing also helped attain high overlap between the actual gland regions and the segmented gland regions and hence the high values of object-wise metrics JIweighted and DIweighted. There were some cases of segmented glands indicating surrounding stroma as gland region and hence having a different shape, leading to a high value of the object-wise metric HDweighted as shown in Table 3.

Nguyen et al.23 obtained gland objects by convex hulls of the same pixels in the detected lumen and nuclei objects. The convex hulls lead to shape differences between the actual glands and detected glands. Also, the detected glands (convex hulls) sometimes included adjoining stroma regions and/or excluded some the glandular region, leading to low values of object-wise metrics JIweighted and DIweighted as shown in Table 3. The shape differences lead to a high value of the object-wise metric HDweighted as shown in Table 3.

Our implementation of Naik et al.,17 as discussed, merged nearby/touching glands into one gland. This lead to low overlap and drastic shape differences between the actual gland regions and detected gland regions. Due to low overlap, we obtained low values of object-wise metrics JIweighted and DIweighted, and huge shape differences lead to a high value of the object-wise metric HDweighted as shown in Table 3.

Farjam et al.,16 as discussed, identified the nuclei and surrounding regions as glands. It essentially identified a lot of false gland objects. These objects did not cover the actual gland regions and led to low overlap and substantial shape difference between the matched objects. This led to very low values of object-wise metrics JIweighted and DIweighted, and huge shape differences led to a high value of the object-wise metric HDweighted as shown in Table 3.

4.2. Failure Cases

In our dataset of 43 images, we have a total of six images with a cribriform pattern. Our method resulted in inaccurate segmentation for these six cribriform pattern cases. We show an example image in Figs. 9(g)9(i). In these cribriform cases, the glands have multiple lumen and they are surrounded by nuclei instead of just cytoplasm. Our method REFINE-SEEDS (Algorithm 3) failed in merging them as it relies on cytoplasm to define the inside region of the gland and on nuclei to define the gland boundary. REFINE-SEEDS assumes the regions with high nuclei density as the border regions and keeps the multiple lumina disconnected. Due to inaccurate seed refining, subsequent THREE-LEVEL-EGVD results in oversegmentation as shown in Fig. 9(i). Other methods, such as Farjam et al.,16 Naik et al.,17 and Nguyen et al.,23 also generated final segmentation similar to that of our method (oversegmented) for these six cribriform pattern cases.

5. Conclusion

We developed an automated method for gland segmentation in prostate histopathological images using machine learning and image processing methods. We used a dataset of Gleason pattern 3 and 4 PCa images for developing this system. Our method outperformed texture- and gland structure-based methods by Farjam et al.,16 Naik et al.,17 and Nguyen et al.23 Our method, however, fails in the images with the cribriform pattern. The cribriform pattern contains a lot of glands fused into one and hence contains a lot of nuclei clusters inside a gland. These clusters, in turn, impede our method during the seed-refining step, resulting in inaccurate segmentation.

A possible extension of our work can be the development of a practical user-friendly interface with efficient image acquisition and dataset preparation systems. Another future research can be the development of automated systems to analyze the segmented glands for grading purposes. The architecture, shape, and size of glands are some of the important features in the diagnostics of PCa. A good quality gland segmentation enables the computer to elucidate these features correctly. As such, a gland segmentation system may work as a base for future expert systems that analyze gland architectures.

Acknowledgments

This work was supported in part by the Biomedical Research Council of A*STAR (Agency of Science, Technology and Research), Singapore, the National University of Singapore, Singapore, and the Departments of Urology and Pathology at Tan Tock Seng Hospital, Singapore. Malay Singh would also like to thank his colleagues Connie Kou, Daniel Holdbrook, and Lakshmi Govindarajan for various fruitful discussions about the methodology and for proofreading the paper.

Biographies

Malay Singh is a PhD student in the Department of Computer Science, School of Computing, National University of Singapore. He received his BTech degree in information technology from the Indian Institute of Information Technology-Allahabad, India, in 2012.

Emarene Mationg Kalaw is a research scientist at Bioinformatics Institute. She received her BSc (Hons) degree in biology in 1997 and her MD degree in 2001 from De La Salle University Philippines. She later obtained her MSc degree in medical research from Griffith University in 2007. She has worked as a research pathologist/scientist at Singapore General Hospital and National Cancer Centre, Singapore. Her research interests include histopathology for prostate cancer (PCa), kidney cancer, and bladder cancer.

Danilo Medina Giron is a senior pathologist at the Tan Tock Seng Hospital, a tertiary hospital in Singapore. He is a medical graduate from the University of the Philippines and finished his pathology residency at the State University of New York at Buffalo. He is board certified by the American Board of Pathology and the Royal College of Physicians and Surgeons of Canada. He practiced pathology in Canada, the Philippines, and Singapore.

Kian-Tai Chong is the consultant, adjunct assistant professor, codirector of urologic oncology, and director of research in the Department of Urology. He also manages the prospective urology cancer clinical databases and tissue repository in Tan Tock Seng Hospital. He completed his full-time clinical urologic-oncology fellowship at Memorial-Sloan Kettering Cancer Center, New York. His key professional interests are the prevention, diagnosis, treatment, and management of PCa, kidney cancer, bladder cancer, testis cancer, and penile cancer.

Chew Lim Tan is a professor in the Department of Computer Science, School of Computing, National University of Singapore. He received his BSc (Hons) degree in physics from the University of Singapore in 1971, his MSc degree in radiation studies from the University of Surrey, UK, in 1973, and his PhD in computer science from the University of Virginia, USA, in 1986. His research interests include document image analysis, text and natural language processing, and neural networks.

Hwee Kuan Lee is a senior principal investigator of the Imaging Informatics Division in Bioinformatics Institute. His current research work involves computer vision algorithms for clinical and biological studies. He obtained his PhD in theoretical physics from Carnegie Mellon University in 2001. He then held a joint postdoctoral position with Oak Ridge National Laboratory and University of Georgia. He worked at Tokyo Metropolitan University and Data Storage Institute before joining Bioinformatics Institute in 2006.

Disclosures

The authors, Malay Singh, Emarene Mationg Kalaw, Danilo Medina Giron, Kian-Tai Chong, Chew Lim Tan, and Hwee Kuan Lee, have no financial disclosures or conflicts of interest.

References

  • 1.Humphrey P. A., “Gleason grading and prognostic factors in carcinoma of the prostate,” Mod. Pathol. 17(3), 292–306 (2004). 10.1038/modpathol.3800054 [DOI] [PubMed] [Google Scholar]
  • 2.Gleason D., “Histological grading and clinical staging of prostatic carcinoma,”in Urol. Pathol.: Prostate, Tannenbaum M., Ed., pp. 171–198, Lea & Feibiger, Philadelphia: (1977). [Google Scholar]
  • 3.Sternberg S. S., Mills S. E., Carter D., Sternberg’s Diagnostic Surgical Pathology, Vol. 1, Lippincott Williams & Wilkins, Philadelphia: (2004). [Google Scholar]
  • 4.Epstein J. I., et al. , “Prediction of progression following radical prostatectomy: a multivariate analysis of 721 men with long-term follow-up,” Am. J. Surg. Pathol. 20(3), 286–292 (1996). 10.1097/00000478-199603000-00004 [DOI] [PubMed] [Google Scholar]
  • 5.Albertsen P. C., et al. , “Competing risk analysis of men aged 55 to 74 years at diagnosis managed conservatively for clinically localized prostate cancer,” J. Am. Med. Assoc. 280(11), 975–980 (1998). 10.1001/jama.280.11.975 [DOI] [PubMed] [Google Scholar]
  • 6.Barry M. J., et al. , “Outcomes for men with clinically nonmetastatic prostate carcinoma managed with radical prostactectomy, external beam radiotherapy, or expectant management,” Cancer 91(12), 2302–2314 (2001). [DOI] [PubMed] [Google Scholar]
  • 7.Egevad L., et al. , “Prognostic value of the Gleason score in prostate cancer,” BJU Int. 89(6), 538–542 (2002). 10.1046/j.1464-410X.2002.02669.x [DOI] [PubMed] [Google Scholar]
  • 8.Epstein J. I., et al. , “The 2005 International Society of Urological Pathology (ISUP) consensus conference on Gleason grading of prostatic carcinoma,” Am. J. Surg. Pathol. 29(9), 1228–1242 (2005). 10.1097/01.pas.0000173646.99337.b1 [DOI] [PubMed] [Google Scholar]
  • 9.Epstein J. I., “An update of the Gleason grading system,” J. Urol. 183(2), 433–440 (2010). 10.1016/j.juro.2009.10.046 [DOI] [PubMed] [Google Scholar]
  • 10.Humphrey P. A., et al. , Prostate Pathology, American Society for Clinical Pathology, Chicago: (2003). [Google Scholar]
  • 11.Allsbrook W. C., Jr., et al. , “Interobserver reproducibility of Gleason grading of prostatic carcinoma: urologic pathologists,” Hum. Pathol. 32(1), 74–80 (2001). 10.1053/hupa.2001.21134 [DOI] [PubMed] [Google Scholar]
  • 12.Allsbrook W. C., Jr., et al. , “Interobserver reproducibility of Gleason grading of prostatic carcinoma: general pathologist,” Hum. Pathol. 32(1), 81–88 (2001). 10.1053/hupa.2001.21135 [DOI] [PubMed] [Google Scholar]
  • 13.Bottke D., et al. , “Phase 3 study of adjuvant radiotherapy versus wait and see in pt3 prostate cancer: impact of pathology review on analysis,” Eur. Urol. 64(2), 193–198 (2013). 10.1016/j.eururo.2013.03.029 [DOI] [PubMed] [Google Scholar]
  • 14.Nguyen K., Sabata B., Jain A. K., “Prostate cancer grading: gland segmentation and structural features,” Pattern Recognit. Lett. 33(7), 951–961 (2012). 10.1016/j.patrec.2011.10.001 [DOI] [Google Scholar]
  • 15.Yap C. K., et al. , “Automated image based prominent nucleoli detection,” J. Pathol. Inf. 6, 39 (2015). 10.4103/2153-3539.159232 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16.Farjam R., et al. , “An image analysis approach for automatic malignancy determination of prostate pathological images,” Cytom. Part B: Clin. Cytom. 72(4), 227–240 (2007). 10.1002/(ISSN)1552-4957 [DOI] [PubMed] [Google Scholar]
  • 17.Naik S., et al. , “Gland segmentation and computerized Gleason grading of prostate histology by integrating low-, high-level and domain specific information,” in MIAAB Workshop, pp. 1–8 (2007). [Google Scholar]
  • 18.Xu J., et al. , “High-throughput prostate cancer gland detection, segmentation, and classification from digitized needle core biopsies,” in Int. Workshop on Prostate Cancer Imaging, pp. 77–88, Springer; (2010). [Google Scholar]
  • 19.Gunduz-Demir C., et al. , “Automatic segmentation of colon glands using object-graphs,” Med. Image Anal. 14(1), 1–12 (2010). 10.1016/j.media.2009.09.001 [DOI] [PubMed] [Google Scholar]
  • 20.Nguyen K., Jain A. K., Allen R. L., “Automated gland segmentation and classification for Gleason grading of prostate tissue images,” in 20th Int. Conf. on Pattern Recognition (ICPR’10), pp. 1497–1500, IEEE; (2010). 10.1109/ICPR.2010.370 [DOI] [Google Scholar]
  • 21.Monaco J. P., et al. , “High-throughput detection of prostate cancer in histological sections using probabilistic pairwise Markov models,” Med. Image Anal. 14(4), 617–629 (2010). 10.1016/j.media.2010.04.007 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22.Hojjatoleslami S., Kittler J., “Region growing: a new approach,” IEEE Trans. Image Process. 7(7), 1079–1084 (1998). 10.1109/83.701170 [DOI] [PubMed] [Google Scholar]
  • 23.Nguyen K., Sarkar A., Jain A. K., “Structure and context in prostatic gland segmentation and classification,” in Int. Conf. on Medical Image Computing and Computer-Assisted Intervention, pp. 115–123, Springer; (2012). [DOI] [PubMed] [Google Scholar]
  • 24.Nguyen K., Sarkar A., Jain A. K., “Prostate cancer grading: use of graph cut and spatial arrangement of nuclei,” IEEE Trans. Med. Imaging 33(12), 2254–2270 (2014). 10.1109/TMI.2014.2336883 [DOI] [PubMed] [Google Scholar]
  • 25.Shi J., Malik J., “Normalized cuts and image segmentation,” IEEE Trans. Pattern Anal. Mach. Intell. 22(8), 888–905 (2000). 10.1109/34.868688 [DOI] [Google Scholar]
  • 26.Ren J., et al. , “Computer aided analysis of prostate histopathology images Gleason grading especially for Gleason score 7,” in 37th Annual Int. Conf. of the IEEE Engineering in Medicine and Biology Society (EMBC’15), pp. 3013–3016, IEEE; (2015). 10.1109/EMBC.2015.7319026 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27.Kothari S., et al. , “Automatic batch-invariant color segmentation of histological cancer images,” in IEEE Int. Symp. on Biomedical Imaging: From Nano to Macro, 2011, pp. 657–660, IEEE; (2011). 10.1109/ISBI.2011.5872492 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 28.Macenko M., et al. , “A method for normalizing histology slides for quantitative analysis,” in IEEE Int. Symp. on Biomedical Imaging: From Nano to Macro (ISBI’09), pp. 1107–1110, IEEE; (2009). 10.1109/ISBI.2009.5193250 [DOI] [Google Scholar]
  • 29.Sirinukunwattana K., Snead D. R., Rajpoot N. M., “A stochastic polygons model for glandular structures in colon histology images,” IEEE Trans. Med. Imaging 34(11), 2366–2378 (2015). 10.1109/TMI.2015.2433900 [DOI] [PubMed] [Google Scholar]
  • 30.Madabhushi A., Lee G., “Image analysis and machine learning in digital pathology: challenges and opportunities,” Med. Image Anal. 33(6), 170–175 (2016). 10.1016/j.media.2016.06.037 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 31.Greenspan H., van Ginneken B., Summers R. M., “Guest editorial deep learning in medical imaging: overview and future promise of an exciting new technique,” IEEE Trans. Med. Imaging 35(5), 1153–1159 (2016). 10.1109/TMI.2016.2553401 [DOI] [Google Scholar]
  • 32.Janowczyk A., Madabhushi A., “Deep learning for digital pathology image analysis: a comprehensive tutorial with selected use cases,” J. Pathol. Inf. 7, 29 (2016). 10.4103/2153-3539.186902 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 33.Xu Y., et al. , “Gland instance segmentation using deep multichannel neural networks,” IEEE Trans. Biomed. Eng. PP(99), 1 (2017). 10.1109/TBME.2017.2686418 [DOI] [PubMed] [Google Scholar]
  • 34.Chen H., et al. , “DCAN: deep contour-aware networks for accurate gland segmentation,” in Proc. of the IEEE Conf. on Computer Vision and Pattern Recognition, 2487–2496 (2016). 10.1109/CVPR.2016.273 [DOI] [Google Scholar]
  • 35.Ren J., et al. , “Computer aided analysis of prostate histopathology images to support a refined Gleason grading system,” Proc. SPIE 10133, 101331V (2017). 10.1117/12.2253887 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 36.Jia Z., et al. , “Constrained deep weak supervision for histopathology image segmentation,” arXiv preprint arXiv:1701.00794, https://arxiv.org/abs/1701.00794 (2017). [DOI] [PubMed]
  • 37.Chen L.-C., et al. , “Semantic image segmentation with deep convolutional nets and fully connected CRFs,” arXiv preprint arXiv:1412.7062, https://arxiv.org/abs/1412.7062 (2014). [DOI] [PubMed]
  • 38.Everingham M., et al. , “The PASCAL visual object classes (VOC) challenge,” Int. J. Comput. Vision 88(2), 303–338 (2010). 10.1007/s11263-009-0275-4 [DOI] [Google Scholar]
  • 39.The cancer genome atlas-data portal, https://tcga-data.nci.nih.gov/tcga (15 January 2014).
  • 40.Cortes C., Vapnik V., “Support-vector networks,” Mach. Learn. 20(3), 273–297 (1995). 10.1007/BF00994018 [DOI] [Google Scholar]
  • 41.Freund Y., Schapire R. E., “A decision-theoretic generalization of on-line learning and an application to boosting,” J. Comput. Syst. Sci. 55(1), 119–139 (1997). 10.1006/jcss.1997.1504 [DOI] [Google Scholar]
  • 42.Vanderbeck S., et al. , “Automatic classification of white regions in liver biopsies by supervised machine learning,” Hum. Pathol. 45(4), 785–792 (2014). 10.1016/j.humpath.2013.11.011 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 43.Hu M.-K., “Visual pattern recognition by moment invariants,” IRE Trans. Inf. Theory 8(2), 179–187 (1962). 10.1109/TIT.1962.1057692 [DOI] [Google Scholar]
  • 44.Otsu N., “A threshold selection method from gray-level histograms,” IEEE Trans. Syst. Man Cybern. 9(1), 62–66 (1979). 10.1109/TSMC.1979.4310076 [DOI] [Google Scholar]
  • 45.Tsai W.-H., “Moment-preserving thresholding: a new approach,” Comput. Vision Graphics Image Process. 29(3), 377–393 (1985). 10.1016/0734-189X(85)90133-1 [DOI] [Google Scholar]
  • 46.Doyle W., “Operations useful for similarity-invariant pattern recognition,” J. ACM 9(2), 259–267 (1962). 10.1145/321119.321123 [DOI] [Google Scholar]
  • 47.Kapur J. N., Sahoo P. K., Wong A. K., “A new method for gray-level picture thresholding using the entropy of the histogram,” Comput. Vision Graphics Image Process. 29(3), 273–285 (1985). 10.1016/0734-189X(85)90125-2 [DOI] [Google Scholar]
  • 48.Shanbhag A. G., “Utilization of information measure as a means of image thresholding,” CVGIP: Graphical Models Image Process. 56(5), 414–419 (1994). 10.1006/cgip.1994.1037 [DOI] [Google Scholar]
  • 49.Ruifrok A. C., et al. , “Quantification of histochemical staining by color deconvolution,” Anal. Quant. Cytol. Histol. 23, 291–299 (2001). [PubMed] [Google Scholar]
  • 50.Yu W., et al. , “Evolving generalized Voronoi diagrams for accurate cellular image segmentation,” Cytometry 77(4), 379–386 (2010). 10.1002/cyto.a.v77a:4 [DOI] [PubMed] [Google Scholar]
  • 51.Collins L. M., Dent C. W., “Omega: a general formulation of the rand index of cluster recovery suitable for non-disjoint solutions,” Multivar. Behav. Res. 23(2), 231–242 (1988). 10.1207/s15327906mbr2302_6 [DOI] [PubMed] [Google Scholar]
  • 52.Gale D., Shapley L. S., “College admissions and the stability of marriage,” Am. Math. Mon. 69(1), 9–15 (1962). 10.2307/2312726 [DOI] [Google Scholar]

Articles from Journal of Medical Imaging are provided here courtesy of Society of Photo-Optical Instrumentation Engineers

RESOURCES