Skip to main content
Computational and Mathematical Methods in Medicine logoLink to Computational and Mathematical Methods in Medicine
. 2015 Feb 24;2015:597475. doi: 10.1155/2015/597475

Adaptive Thresholding Technique for Retinal Vessel Segmentation Based on GLCM-Energy Information

Temitope Mapayi 1, Serestina Viriri 1, Jules-Raymond Tapamo 2,*
PMCID: PMC4354732  PMID: 25802550

Abstract

Although retinal vessel segmentation has been extensively researched, a robust and time efficient segmentation method is highly needed. This paper presents a local adaptive thresholding technique based on gray level cooccurrence matrix- (GLCM-) energy information for retinal vessel segmentation. Different thresholds were computed using GLCM-energy information. An experimental evaluation on DRIVE database using the grayscale intensity and Green Channel of the retinal image demonstrates the high performance of the proposed local adaptive thresholding technique. The maximum average accuracy rates of 0.9511 and 0.9510 with maximum average sensitivity rates of 0.7650 and 0.7641 were achieved on DRIVE and STARE databases, respectively. When compared to the widely previously used techniques on the databases, the proposed adaptive thresholding technique is time efficient with a higher average sensitivity and average accuracy rates in the same range of very good specificity.

1. Introduction

Retinal fundus imaging in ophthalmology is of great use in medical diagnosis and progression monitoring of several diseases like hypertension, diabetes, stroke, and cardiovascular disease [1]. Automatic vessel segmentation has a great potential to assist in the reduction of the time required by physicians or skilled technicians for manual labeling of retinal vessels [2].

Several retinal vessel segmentation techniques have been proposed and evaluated in literatures. Chaudhuri et al. [3] implemented a two-dimensional matched filter using a Gaussian shaped curve. However, the technique proposed in [3] achieved very low average accuracy due to low detection of retinal vessels. Hoover [4] segmented retinal vessels by applying a threshold probing technique combining local vessel attributes with region-based attributes on matched filter response (MFR) image. When compared to [3] where a basic thresholding of an MFR was used, the method proposed by [1] reduced the false positive rate by as much as 15 times. Fraz et al. [5] implemented vessel segmentation technique utilizing extracted center-lines of retinal vessels through first-order derivative of Gaussian filter. The authors used morphological operator with directional structuring elements to enhance the structure of blood vessels. They further generated the shape and orientation map of the blood vessels using the bit planes of a gray-scale image. Chakraborti et al. [6] implemented an unsupervised segmentation technique that combines vesselness filter and matched filter using orientation histogram for the segmentation of retinal vessels. Martinez-Perez et al. [7] used a combination of scale space analysis and region growing to segment the vasculature. The technique proposed in [7] was however unable to segment the thin vessels. Zana and Klein [8] used a general vessel segmentation method based on mathematical morphology. However, the technique proposed in [8] was unable to segment the thinner vessels.

Wang et al. [9] proposed multiwavelet kernels and multiscale hierarchical decomposition. Vessels were enhanced using matched filtering with multiwavelet kernels. Szpak and Tapamo [10] used gradient based approach and level set technique. The proposed technique in [10] was however unable to detect the thinner vessels. Vlachos and Dermatas [11] proposed a multiscale retinal vessel segmentation method. The algorithm is based on multiscale line-tracking procedure and morphological postprocessing. However, the proposed technique in [11] was unable to detect the thinner vessels. Mendonça and Campilho [12] combined differential filters for center-line extraction with morphological operators for filling vessel segments considering intensity and morphological properties. Xiao et al. [13] proposed a Bayesian method with spatial constraint for the segmentation of retinal vessels. The spatial dependence of the posterior probability of each pixel in relation to their neighboring pixels was utilized. An energy function was further defined and a modified level set approach was used for the vessel segmentation. Yin et al. [14] implemented a probabilistic tracking-based method for vessel segmentation. A Bayesian method with maximum a posteriori (MAP) was used for detecting the retinal vessel edge points.

Niemeijer et al. [15] proposed pixel classification using a K-nearest neighbour classifier for the segmentation of vessels. Staal et al. [16] proposed a ridge-based vessel segmentation method. The feature vectors were computed for every pixel and classified using a K-nearest neighbour classifier and sequential forward feature selection. Soares et al. [17] generated a feature vector computed from the measurements at different scales of two-dimensional (2D) Gabor wavelet transform on each pixel. Bayesian classifier with Gaussian mixtures was further used to classify the resulting feature space as either vessel or nonvessel pixel. Ricci and Perfetti [18] proposed automated vessel segmentation based on line operators. Two segmentation methods were considered. One of the segmentation methods used two orthogonal line detectors with the gray level of the target pixel to construct a feature vector for supervised classification using a support vector machine. Another segmentation method used by [18] thresholds the response of a basic line detector to obtain unsupervised pixel classification. Marín et al. [19] computed a 7D vector composed of gray-level and moment invariants-based features for pixel representation and used a neural network classifier for the pixel classification. Although the techniques proposed by [1519] provide high sensitivity and accuracy measures, one major challenge is the requirement of more time for the training phase of the classifiers.

A number of local adaptive thresholding approaches for retinal vessel segmentation such as [2023] have been proposed in different literatures. Jiang and Mojon [22] proposed an adaptive local thresholding framework using a verification-based multithreshold probing scheme. Although the average running time of the proposed technique in [22] is relatively very fast (8 to 36 seconds), it was unable to detect the thinner vessels. Akram and Khan [20] enhanced the vascular pattern using 2D Gabor wavelet and followed it by a multilayered thresholding technique that applied different threshold values iteratively to generate gray-level segmented image. Cornforth et al. [21] applied wavelet analysis, supervised classifier probabilities, and adaptive threshold procedures, as well as morphology-based techniques. Li et al. [23] combined multiscale analysis based on Gabor filters, scale multiplication, and region-based thresholding to achieve adaptive thresholding for vessel segmentation.

Several other works such as [2427] have combined pixel thresholding based on certain neighbourhood with global thresholding technique as an adaptive thresholding technique for different segmentation problems.

Gray-level cooccurrence matrix (GLCM) is popularly known for its usage for texture image segmentation [2831]. Haralick features [28] computed from GLCM have been used for both supervised and unsupervised segmentation. Some known unsupervised gray-level cooccurrence based segmentation techniques have been proposed in some other literatures. Entropy has been one of the few major GLCM features that has often been used for unsupervised segmentation. Different entropy based thresholding such as global, local, joint, and relative entropy has been proposed in [3236]. Chanwimaluang and Fan [37] proposed the combination of matched filter and entropy for the segmentation of retinal vessels. The performance measure of the proposed technique in [37] was only visual. Li et al. [38] used a threshold selection method based on multiscale edge analysis and gray-level cooccurrence matrix to handle severely degraded document images. A multiscale image description was first used to analyse the image edge; then gray-level cooccurrence matrix was further used to compute the edge pixel pair information. A threshold value was computed using the edge pixel pair cooccurrence matrix. Mokji and Abu Bakar [29] proposed a technique based on the cooccurrence matrix where statistical features were defined from the edge information to handle images that have fuzzy boundaries between the object and the background of the image.

Although much has been achieved in the previous works, the performance measurement and visual results obtained from literatures suggest the need for further research work to address the robust segmentation of both large and thin vessels in a timely efficient manner. This paper proposes a local adaptive thresholding technique using gray-level cooccurrence matrix- (GLCM-) energy information for the robust segmentation of both large and thin vessels in a timely efficient manner.

The rest of this paper is organized as follows. Section 2 describes the methods and techniques used in this paper. Section 3 explains the experimental setup and results and discussion, while the conclusion is drawn in Section 4.

2. Methods and Techniques

The proposed segmentation approach is considered because it utilizes the angular second moment feature which captures the dynamics of the textural information needed for the robust segmentation of both large and thin vessels. A brief description of GLCM and the formulation of a multiscale angular second moment feature matrix are given in this section. This is followed by the proposed local adaptive thresholding technique. Morphological postprocessing technique is finally applied to remove the misclassifications in the segmented vessels. Although most of the previous techniques used the Green Channel of the retinal image, the proposed local adaptive thresholding technique uses the grayscale and the Green Channel of the colored retinal fundus image.

  • (I)
    Gray-level cooccurrence matrix: GLCM is usually computed using two key parameters, namely, the relative distance “d” between the pixel pair and their relative orientation “Φ.” The distance “d” is usually measured in pixel number while “Φ” is quantized in four directions (horizontal: 0°, diagonal: 45°, vertical: 90°, and antidiagonal: 135°). Given a grayscale image V of M rows and N columns, the gray-level cooccurrence matrix C(i, j), for distance “d” and direction “Φ,” is defined as
    Ci,j=x=0M1y=0N1PVx,y=i,Vx±dΦ1,y±dΦ2=j, (1)
  • where V(x, y) = i which means that i is the gray-level intensity of pixel (x, y) and P is defined as
    Px=1,if  x  is  true,0,otherwise. (2)
  • Six features of GLCM were considered by Cataldo et al. [24] to be the most relevant. These features considered are energy, entropy, contrast, variance, correlation, and inverse difference moment. Energy, which is also called angular second moment (ASM) or uniformity, measures the textural uniformity. It is a GLCM feature that detects disorders in textures. Information based on energy feature will be applied for an adaptive thresholding process in this work. The angular second moment is defined as
    ASM=ijhi,j2, (3)
  • where h(i, j) is the (i, j)th entry in a normalized gray-tone spatial dependence matrix C (i,j)/R, with 1/R being the normalizing factor.

  • The visibility of vessels in a colored fundus retinal image with its grayscale intensity image and the Green Channel of the retinal image are shown in Figure 1. The grayscale intensity and Green Channel of the retinal image are further processed for the segmentation of the retinal vessels.

  • (II)
    Multiscale ASM-feature measurement: the variation of energy information within the varying distance “d” and relative orientation “Φ” is useful in the design of an adaptive thresholding technique for image segmentation. An ASM-feature matrix across different orientations and distances is computed and defined as
    A=(aij),1i,j4, (4)
  • where
    aij=ASM(di,Φj),1i,j4, (5)
  • such that Φ1 = 0°, Φ2 = 45°, Φ3 = 90°, and Φ4 = 135°, with distances (d i)i=1,…,4. The range measure of A is given as follows:
    RangeΦ=RangeA, (6)
  • such that RangeΦ is a row vector containing the range of each column of A. Three different threshold values are computed from each row vector to segment the retinal vessels. The thresholds from the range measure are
    K=0.5MINRangeΦ, (7)
    K=0.5MAXRangeΦ, (8)
    K=0.5MEANRangeΦ. (9)
  • (III)

    The proposed local adaptive thresholding technique includes the following.

    • (a)
      Image enhancement is as follows. A combination of unsharp filter, average filter, and contrast enhancement is applied on the grayscale and Green Channel of the retinal image.
    • (b)
      Convolution of the result is obtained in (a) through a median filter using local window size ww. This is described as
      Ui,j=Hx,yVww1x,y, (10)
    • where U(i, j) is the convolved retinal image, V 1[x, y] is the result obtained in (a) and the convolution mask H[x, y] is a local median filter.
    • (c)
      The difference image D(x, y) is then computed. This is described as
      Dx,y=Ui,jV1x,y. (11)
    • (d)
      The segmented image S image is obtained as
      Simage(x,y)=0,if  Dx,yTx,y,1,otherwise, (12)
    • where T(x, y) = K.
  • (IV)

    Postprocessing is as follows. A combination of morphological opening with median filtering process is performed on the inverted thresholded image to handle the remaining misclassifications.

Figure 1.

Figure 1

(a) Colored retinal image; (b) grayscale retinal image; (c) Green Channel of the colored retinal image.

3. Experimental Results and Discussion

Experiments were carried out using MATLAB 2010a on an Intel Core i5 2410M CPU, 2.30 GHz, 4 GB of RAM. The proposed method was evaluated using the retinal images on the publicly available DRIVE [39] and STARE [4] databases. DRIVE database is made up of 40 images captured with the use of Canon CR5 camera with 24-bit grayscale resolution and a spatial resolution of 565 × 584 pixels. The 40 images were divided into two. The first group of the DRIVE images is a training set made up of twenty images. The second group is a testing set made up of twenty images. DRIVE database also provides gold standard images as the ground truth for vessel segmentation for the comparative performance evaluation of different vessel segmentation algorithms. STARE database consists of retinal images captured with the use of TopCon TRV-50 fundus camera with 24-bit grayscale resolution and spatial resolution of 700 × 605 pixels. The database provides 20 coloured retinal images and 20 hand-labeled images as the ground truth for the comparative performance evaluation of different vessel segmentation algorithms.

Empirically, we established that window sizes 11 × 11 to 17 × 17 were effective for the segmentation of the retinal vessels. There is however a higher amount of noise and an increase in the computational time when the window size is too large (i.e., larger than 17 × 17). In such a situation, the further postprocessing for removal of noise leads to the removal of the thin vessels as well as some large vessels. This is however caused by the influence of the noneven illumination across the retinal image. In related development, there is a possibility of insufficient data when the window size is too small (i.e., lesser than 11 × 11). This leads to the loss of some large and thin vessels during segmentation. The average time taken for the different window sizes to process each image on DRIVE database ranges from 1.9 to 2.6 seconds.

The performance measures commonly used are sensitivity, specificity, and accuracy. The measures are described in (12)–(14) as follows:

Sensitivity=TP(TP+FN), (13)
Specificity=TN(TN+FP), (14)
Accuracy=(TP+TN)(TP+TN+FP+FN), (15)

where TP = true positive, TN = true negative, FP = false positive, and FN = false negative.

An event is said to be TP if a pixel is rightly segmented as a vessel and TN when a pixel is rightly segmented as background. In related development, an event is said to be FN if a vessel pixel is segmented to be a background and a FP when a background pixel is segmented as a pixel in the vessel. Sensitivity measure indicates the ability of a segmentation technique to detect the vessel pixels while specificity measure indicates the ability of a segmentation technique to detect background pixels. The accuracy measure indicates the degree of conformity of the segmented retinal image to the ground truth.

A receiver operating characteristic (ROC) curve performance measure is a plot of the rightly classified pixels, referred to as true positive rate (TPR) versus the fraction of the wrongly classified pixels as vessels, referred to as false positive rate (FPR). Area under the curve (AUC) is a performance measure computed from the ROC curve.

The thresholds K as computed in (6)–(8) are substituted for T(x, y) in (11) for the segmentation of retinal fundus image.

The different results obtained from the grayscale intensity image and the Green Channel of the colored fundus image using the proposed ASM-based local adaptive thresholding technique are compared with the manually segmented vessel by the second human observer and the DRIVE database ground truth in Figure 4. Figures 2 and 5 also show the results obtained by the proposed local adaptive thresholding technique based on different ASM range information-based threshold values on DRIVE database. Figures 3 and 10 show the visual results obtained by the proposed local adaptive thresholding technique on STARE database.

Figure 4.

Figure 4

(a) DRIVE database gold standard; (b) manually segmented vessel by the second human observer on DRIVE database; (c) segmented vessel of the Green Channel using ASM range information-based threshold value; (d) segmented vessel of the grayscale intensity image using ASM range information-based threshold value.

Figure 2.

Figure 2

(a) DRIVE colored fundus image; (b) DRIVE database gold standard; (c) segmented vessel of the Green Channel using ASM range information-based threshold value.

Figure 5.

Figure 5

Adaptive thresholding using different ASM range information on DRIVE database. (a) DRIVE database colored retinal image. (b) DRIVE database gold standard. (c) Segmented vessel through adaptive thresholding using minimum ASM mid-range threshold value for the grayscale intensity image. (d) Segmented vessel through adaptive thresholding using maximum ASM mid-range threshold value for the grayscale intensity image. (e) Segmented vessel through adaptive thresholding using mean ASM mid-range threshold value for the grayscale intensity image. (f) Segmented vessel through adaptive thresholding using minimum ASM mid-range threshold value for the Green Channel. (g) Segmented vessel through adaptive thresholding using maximum ASM mid-range threshold value for the Green Channel. (h) Segmented vessel through adaptive thresholding using mean ASM mid-range threshold value for the Green Channel.

Figure 3.

Figure 3

(a) STARE database ground-truth; (b) segmented vessel of the grayscale intensity image using ASM range information-based threshold value; (c) segmented vessel of the Green Channel using ASM range information-based threshold value.

Figure 10.

Figure 10

Adaptive thresholding using different ASM range information on STARE database. (a) STARE database ground truth. (b) Segmented retinal vessel by Hoover [4]. (c) Segmented vessel through adaptive thresholding using minimum ASM mid-range threshold value for the grayscale intensity image. (d) Segmented vessel through adaptive thresholding using maximum ASM mid-range threshold value for the grayscale intensity image. (e) Segmented vessel through adaptive thresholding using mean ASM mid-range threshold value for the grayscale intensity image. (f) Segmented vessel through adaptive thresholding using minimum ASM mid-range threshold value for the Green Channel. (g) Segmented vessel through adaptive thresholding using maximum ASM mid-range threshold value for the Green Channel. (h) Segmented vessel through adaptive thresholding using mean ASM mid-range threshold value for the Green Channel.

Table 1 shows the performance of the different GLCM-energy threshold values for the proposed adaptive thresholding technique using DRIVE database. All the grayscale intensity images have slightly lower sensitivity rates of 0.7397, 0.7313, and 0.7375 than the Green Channel of the colored retinal images with average sensitivity rates of 0.7650, 0.7560, and 0.7632. The grayscale intensity images however have slightly higher accuracies of 0.9488, 0.9511, and 0.9503 over the Green Channel of the colored retinal images with average accuracies of 0.9449, 0.9461, and 0.9477. The maximum mid-range threshold value on the grayscale intensity image yields the best average accuracy rate of 0.9511 while the least average accuracy rate of 0.9449 was achieved using the minimum mid-range threshold value on the Green Channel of the colored retinal images. In related development, the minimum mid-range threshold value on the Green Channel of the colored retinal images yields the highest average sensitivity of 0.7650 while the maximum mid-range threshold value on the grayscale intensity image yields the least average sensitivity of 0.7313. The average specificities of all the grayscale intensity images are slightly higher than the average specificities of the Green Channel of the colored retinal images.

Table 1.

Performance of different segmentation methods on DRIVE database.

Method Average accuracy Average sensitivity Average specificity
Human observer 0.9473 0.7761 0.9725
Staal et al. [16] 0.9442 0.7345 0.9773
Niemeijer et al. [15] 0.9416 0.7145 0.9801
Zana and Klein [8] 0.9377 0.6971 0.9769
Jiang and Mojon [22] 0.9212 0.6399 0.9625
Vlachos and Dermatas [11] 0.9285 0.7468 0.9551
Wang et al. [9] 0.9461 N/A N/A
Martinez-Perez et al. [7] 0.9181 0.6389 0.9496
Szpak and Tapamo [10] 0.9299 N/A N/A
Chaudhuri et al. [3] 0.8773 0.3357 0.9794
Soares et al. [17] 0.9466 N/A N/A
Akram and Khan [20] 0.9469 N/A N/A
Mendonça and Campilho [12] 0.9463 0.7315 N/A
Marín et al. [19] 0.9452 0.7067 0.9801
Ricci and Perfetti [18] 0.9595 N/A N/A
Xiao et al. [13] 0.9529 0.7513 0.9792
Yin et al. [14] 0.9267 0.6522 0.9710
Chakraborti et al. [6] 0.9370 0.7205 0.9579
K=0.5MINRangeΦGrayIntensity 0.9488 0.7397 0.9691
K=0.5MAXRangeΦGrayIntensity 0.9511 0.7313 0.9724
K=0.5MEANRangeΦ(GrayIntensity) 0.9503 0.7375 0.9709
K=0.5MINRangeΦGreenChannel 0.9449 0.7650 0.9623
K=0.5MAXRangeΦGreenChannel 0.9477 0.7560 0.9663
K=0.5MEANRangeΦGreenChannel 0.9461 0.7632 0.9634

Table 2 shows the performance of the different GLCM-energy threshold values for the proposed adaptive thresholding technique using STARE database. All the grayscale intensity images have slightly lower sensitivity rates of 0.7458, 0.7428, and 0.7427 than the Green Channel of the colored retinal images with average sensitivity rates of 0.7542, 0.7641, and 0.7626. The grayscale intensity images also achieve average accuracy rates of 0.9485, 0.9500, and 0.9504 while the Green Channel of the colored retinal images achieved average accuracy rates of 0.9457, 0.9500, and 0.9510. The average mid-range threshold value on the Green Channel of the retinal image achieved the best average accuracy rate of 0.9510 while the least average accuracy rate of 0.9457 was achieved using the minimum mid-range threshold value on the Green Channel of the colored retinal images.

Table 2.

Performance of different segmentation methods on STARE database.

Method Average
accuracy
Average
sensitivity
Average
specificity
Human observer 0.9354 0.8949 N/A
Hoover [4] 0.9275 0.6751 0.9567
Staal et al. [16] 0.9516 0.6970 N/A
Jiang and Mojon [22] 0.9009 N/A N/A
Marín et al. [19] 0.9526 0.6944 0.9819
Ricci and Perfetti [18] 0.9584 N/A N/A
Soares et al. [17] 0.9480 N/A N/A
Akram and Khan [20] 0.9502 N/A N/A
Wang et al. [9] 0.9521 N/A N/A
Mendonça and Campilho [12] 0.9479 0.7123 N/A
Xiao et al. [13] 0.9476 0.7147 0.9735
Yin et al. [14] 0.9412 0.7248 0.9666
Chakraborti et al. [6] 0.9379 0.6786 0.9586
K=0.5MINRangeΦGrayIntensity 0.9485 0.7458 0.9649
K=0.5MAXRangeΦGrayIntensity 0.9500 0.7428 0.9668
K=0.5MEANRangeΦ(GrayIntensity) 0.9504 0.7427 0.9672
K=0.5MINRangeΦGreenChannel 0.9457 0.7542 0.9612
K=0.5MAXRangeΦGreenChannel 0.9500 0.7641 0.9651
K=0.5MEANRangeΦGreenChannel 0.9510 0.7626 0.9657

The performance of the adaptive thresholding based on different ASM range information using grayscale image as depicted through ROC curves is shown in Figures 6 and 8. The ROC curves depicting the performance of the adaptive thresholding based on different ASM range information using Green Channel image are also shown in Figures 7 and 9.

Figure 6.

Figure 6

ROC curves showing the performance of each of the adaptive thresholding based on ASM using grayscale on STARE.

Figure 8.

Figure 8

ROC curves showing the performance of each of the adaptive thresholding based on ASM using grayscale on DRIVE.

Figure 7.

Figure 7

ROC curves showing the performance of each of the adaptive thresholding based on ASM using Green Channel on STARE.

Figure 9.

Figure 9

ROC curves showing the performance of each of the adaptive thresholding based on ASM using Green Channel on DRIVE.

Although all the six thresholds performed well, the slightly higher sensitivity rates achieved by the Green Channel indicate the fact that a bit more vessels were detected when compared to the use of grayscale images. This is because the Green Channel provides the best vessel-background contrast.

In order to compare the performance of the proposed technique with the state of the art, comparison is made with the results obtained by different unsupervised and supervised techniques such as Marín et al. [19], Ricci and Perfetti [18], Soares et al. [17], and Staal et al. [16] as shown in Tables 1 to 4.

Table 4.

Comparison of AUC of the proposed techniques with previous works on STARE.

Method AUC
Staal et al. [16] 0.9614
Jiang and Mojon [22] 0.929
Wang et al. [9] 0.9682
Soares et al. [17] 0.9671
Akram and Khan [20] 0.970
Marín et al. [19] 0.9769
Ricci and Perfetti [18] 0.9602
ASMthresh = 0.5 (MINRange) (Gray Intensity) 0.9695
ASMthresh = 0.5 (MAXRange) (Gray Intensity) 0.9681
ASMthresh = 0.5 (MEANRange) (Gray Intensity) 0.9745
ASMthresh = 0.5 (MINRange) (Green Channel) 0.9671
ASMthresh = 0.5 (MAXRange) (Green Channel) 0.9782
ASMthresh = 0.5 (MEANRange) (Green Channel) 0.9781

3.1. Comparison with Existing Segmentation Methods on DRIVE Database

The performance evaluation shows that the works of Chaudhuri et al. [3], Martinez-Perez et al. [7], Vlachos and Dermatas [11], Jiang and Mojon [22], Niemeijer et al. [15], Yin et al. [14], Zana and Klein [8], and Chakraborti et al. [6] present lower average accuracy and lower average sensitivity when compared to all the adaptive thresholding using different ASM range information. Szpak and Tapamo [10] present no average sensitivity but a lower average accuracy when compared to all the adaptive thresholding using different ASM range information. Marín et al. [19] present no average sensitivity but a lower average accuracy when compared to five of the six adaptive thresholding techniques based on ASM information. Soares et al. [17] and Akram and Khan [20] present no average sensitivity but a lower average accuracy when compared to four of the six adaptive thresholding techniques based on ASM information. Mendonça and Campilho [12] present a lower average sensitivity when compared to five of the six proposed thresholds but a lower average accuracy when compared to four of the six thresholds using the investigated techniques.

Staal et al. [16] present a lower average sensitivity when compared to five of the six thresholds but a lower average accuracy when compared to all the threshold values of the investigated technique. Wang et al. [9] present no average sensitivity but a lower average accuracy when compared to four of the six adaptive thresholding techniques based on ASM information. Ricci and Perfetti [18] present no average sensitivity but a higher average accuracy when compared to all the adaptive thresholding techniques based on ASM information. Xiao et al. [13] present a higher average accuracy when compared to all the proposed thresholds but a lower average sensitivity when compared to three of the six proposed thresholds. The average sensitivity of the human observer is higher than all sensitivities of all thresholds of the proposed technique while four of the six average accuracies of the proposed technique are higher when compared to the average accuracy of the human observer.

3.2. Comparison with Existing Segmentation Methods on STARE Database

Hoover [4] and Chakraborti et al. [6] present a lower average accuracy and average sensitivity rates when compared to all the adaptive thresholding techniques using different ASM information. The work of Jiang and Mojon [22] also achieves lower average accuracy rate when compared to all the adaptive thresholding using different ASM information. Staal et al. [16] present a higher average accuracy rate when compared to all the average accuracy rates obtained using adaptive thresholding techniques based on ASM information. The average sensitivity rate presented by Staal et al. [16] was however lower when compared to all the average sensitivity rates obtained using adaptive thresholding techniques based on ASM information. Yin et al. [14] present a lower average accuracy rate when compared to all the average accuracy rates obtained using adaptive thresholding techniques based on ASM information.

Mendonça and Campilho [12] and Xiao et al. [13] present lower average accuracy rates when compared to five of the six average accuracy rates obtained using adaptive thresholding techniques based on ASM information. The average sensitivity rates obtained by Mendonça and Campilho [12], Xiao et al. [13], and Yin et al. [14] are lower when compared to all the average accuracy rates obtained using adaptive thresholding techniques based on ASM information. Marín et al. [19], Ricci and Perfetti [18], and Wang et al. [9] present no sensitivity rate but a higher average accuracy rate when compared to all the average accuracy rates obtained using adaptive thresholding techniques based on ASM information. Akram and Khan [20] present no sensitivity rate but a lower average accuracy rate when compared to two of the average accuracy rates obtained using adaptive thresholding techniques based on ASM information. Soares et al. [17] also present no sensitivity but a lower average accuracy rate when compared to five of the average accuracy rates obtained using adaptive thresholding techniques based on ASM information. The average sensitivity of the human observer is higher than all the average sensitivity rates of all the adaptive thresholding techniques based on ASM information. The average accuracy rate of the second observer is however lower when compared to all the average accuracy rates obtained from adaptive thresholding techniques based on ASM information.

All the AUC obtained from the segmentation results achieved by the proposed adaptive thresholding techniques based on ASM information (see Table 3) are higher when compared to the AUC of the previously proposed techniques on DRIVE, while two of the six AUC obtained by the proposed adaptive thresholding techniques based on ASM information on STARE (see Table 4) are higher than the AUC of all the previously proposed techniques.

Table 3.

Comparison of AUC of the proposed techniques with previous works on DRIVE.

Method AUC
Staal et al. [16] 0.9520
Niemeijer et al. [15] 0.9294
Zana and Klein [8] 0.8984
Jiang and Mojon [22] 0.9114
Wang et al. [9] 0.9543
Chaudhuri et al. [3] 0.7878
Soares et al. [17] 0.9614
Akram and Khan [20] 0.963
Marín et al. [19] 0.9588
Ricci and Perfetti [18] 0.9558
Chakraborti et al. [6] 0.9419
ASMthresh = 0.5 (MINRange) (Gray Intensity) 0.9656
ASMthresh = 0.5 (MAXRange) (Gray Intensity) 0.9711
ASMthresh = 0.5 (MEANRange) (Gray Intensity) 0.9698
ASMthresh = 0.5 (MINRange) (Green Channel) 0.9634
ASMthresh = 0.5 (MAXRange) (Green Channel) 0.9680
ASMthresh = 0.5 (MEANRange) (Green Channel) 0.9658

4. Conclusion and Future Work

This paper proposes a local adaptive thresholding technique based on GLCM-energy information for the segmentation of retinal vessels in retinal fundus images. It is shown through different thresholds that the proposed local adaptive thresholding techniques based on energy information perform a robust segmentation from both grayscale intensity and the Green Channel of retinal images. Furthermore, it is shown that the proposed local adaptive thresholding technique is time efficient and gives higher average sensitivity, average accuracy, and AUC values when compared to a wide range of previously proposed techniques on both DRIVE and STARE databases. Future work will investigate the use of soft computing and mainly the introduction of heuristics to detect more thin vessels.

Acknowledgments

The authors would like to thank Staal et al. [16, 39] and Hoover [4] for making their database publicly available.

Conflict of Interests

The authors declare that there is no conflict of interests regarding the publication of this paper.

References

  • 1.Kanski J. J. Clinical Ophthalmology. A Systematic Approach. London, U.K: Butterworth-Heinemann; 1989. [Google Scholar]
  • 2.Sussman E. J., Tsiaras W. J., Soper K. A. Diagnosis of diabetic eye disease. The Journal of the American Medical Association. 1982;247(23):3231–3234. doi: 10.1001/jama.247.23.3231. [DOI] [PubMed] [Google Scholar]
  • 3.Chaudhuri S., Chatterjee S., Katz N., Nelson M., Goldbaum M. Detection of blood vessels in retinal images using two-dimensional matched filters. IEEE Transactions on Medical Imaging. 1989;8(3):263–269. doi: 10.1109/42.34715. [DOI] [PubMed] [Google Scholar]
  • 4.Hoover A. Locating blood vessels in retinal images by piecewise threshold probing of a matched filter response. IEEE Transactions on Medical Imaging. 2000;19(3):203–210. doi: 10.1109/42.845178. [DOI] [PubMed] [Google Scholar]
  • 5.Fraz M. M., Barman S. A., Remagnino P., et al. An approach to localize the retinal blood vessels using bit planes and centerline detection. Computer Methods and Programs in Biomedicine. 2012;108(2):600–616. doi: 10.1016/j.cmpb.2011.08.009. [DOI] [PubMed] [Google Scholar]
  • 6.Chakraborti T., Jha D. K., Chowdhury A. S., Jiang X. A self-adaptive matched filter for retinal blood vessel detection. Machine Vision and Applications. 2014:1–14. doi: 10.1007/s00138-014-0636-z. [DOI] [Google Scholar]
  • 7.Martinez-Perez M. E., Hughes A., Stanton A., Thom S., Bharath A., Parker K. Scale-space analysis for the characterisation of retinal blood vessels. In: Taylor C., Colchester A., editors. Medical Image Computing and Computer-Assisted Intervention—MICCAI’99. 1999. pp. 90–97. [Google Scholar]
  • 8.Zana F., Klein J.-C. Segmentation of vessel-like patterns using mathematical morphology and curvature evaluation. IEEE Transactions on Image Processing. 2001;10(7):1010–1019. doi: 10.1109/83.931095. [DOI] [PubMed] [Google Scholar]
  • 9.Wang Y., Ji G., Lin P., Trucco E. Retinal vessel segmentation using multiwavelet kernels and multiscale hierarchical decomposition. Pattern Recognition. 2013;46(8):2117–2133. doi: 10.1016/j.patcog.2012.12.014. [DOI] [Google Scholar]
  • 10.Szpak Z. L., Tapamo J. R. Automatic and interactive retinal vessel segmentation. South African Computer Journal. 2008;40:23–30. [Google Scholar]
  • 11.Vlachos M., Dermatas E. Multi-scale retinal vessel segmentation using line tracking. Computerized Medical Imaging and Graphics. 2010;34(3):213–227. doi: 10.1016/j.compmedimag.2009.09.006. [DOI] [PubMed] [Google Scholar]
  • 12.Mendonça A. M., Campilho A. Segmentation of retinal blood vessels by combining the detection of centerlines and morphological reconstruction. IEEE Transactions on Medical Imaging. 2006;25(9):1200–1213. doi: 10.1109/TMI.2006.879955. [DOI] [PubMed] [Google Scholar]
  • 13.Xiao Z., Adel M., Bourennane S. Bayesian method with spatial constraint for retinal vessel segmentation. Computational and Mathematical Methods in Medicine. 2013;2013:9. doi: 10.1155/2013/401413.401413 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14.Yin Y., Adel M., Bourennane S. Automatic segmentation and measurement of vasculature in retinal fundus images using probabilistic formulation. Computational and Mathematical Methods in Medicine. 2013;2013:16. doi: 10.1155/2013/260410.260410 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15.Niemeijer M., Staal J., Van Ginneken B., Loog M., Abramoff M. D. Comparative study of retinal vessel segmentation methods on a new publicly available database. Proceedings of the SPIE Medical Imaging Conference; 2004; pp. 648–656. [Google Scholar]
  • 16.Staal J., Abràmoff M. D., Niemeijer M., Viergever M. A., van Ginneken B. Ridge-based vessel segmentation in color images of the retina. IEEE Transactions on Medical Imaging. 2004;23(4):501–509. doi: 10.1109/tmi.2004.825627. [DOI] [PubMed] [Google Scholar]
  • 17.Soares J. V. B., Leandro J. J. G., Cesar R. M., Jr., Jelinek H. F., Cree M. J. Retinal vessel segmentation using the 2-D Gabor wavelet and supervised classification. IEEE Transactions on Medical Imaging. 2006;25(9):1214–1222. doi: 10.1109/TMI.2006.879967. [DOI] [PubMed] [Google Scholar]
  • 18.Ricci E., Perfetti R. Retinal blood vessel segmentation using line operators and support vector classification. IEEE Transactions on Medical Imaging. 2007;26(10):1357–1365. doi: 10.1109/TMI.2007.898551. [DOI] [PubMed] [Google Scholar]
  • 19.Marín D., Aquino A., Gegúndez-Arias M. E., Bravo J. M. A new supervised method for blood vessel segmentation in retinal images by using gray-level and moment invariants-based features. IEEE Transactions on Medical Imaging. 2011;30(1):146–158. doi: 10.1109/tmi.2010.2064333. [DOI] [PubMed] [Google Scholar]
  • 20.Akram M. U., Khan S. A. Multilayered thresholding-based blood vessel segmentation for screening of diabetic retinopathy. Engineering with Computers. 2013;29(2):165–173. doi: 10.1007/s00366-011-0253-7. [DOI] [Google Scholar]
  • 21.Cornforth D. J., Jelinek H. J., Leandro J. J. G., et al. Development of retinal blood vessel segmentation methodology using wavelet transforms for assessment of diabetic retinopathy. Complexity International. 2005;11:50–61. [Google Scholar]
  • 22.Jiang X., Mojon D. Adaptive local thresholding by verification-based multithreshold probing with application to vessel detection in retinal images. IEEE Transactions on Pattern Analysis and Machine Intelligence. 2003;25(1):131–137. doi: 10.1109/TPAMI.2003.1159954. [DOI] [Google Scholar]
  • 23.Li Q., You J., Zhang L., Bhattacharya P. A multiscale approach to retinal vessel segmentation using gabor filters and scale multiplication. Proceedings of the IEEE International Conference on Systems, Man and Cybernetics (SMC '06); October 2006; Taipei, Taiwan. pp. 3521–3527. [DOI] [Google Scholar]
  • 24.Cataldo S. D., Ficarra E., Acquaviva A., Macii E. Segmentation of nuclei in cancer tissue images: Contrasting active contours with morphology-based approach. Proceedings of the 8th IEEE International Conference on BioInformatics and BioEngineering (BIBE '08); October 2008; Athens, Greece. pp. 1–6. [DOI] [Google Scholar]
  • 25.Zhou H., Mao K. Z. Adaptive successive erosion-based cell image segmentation for p53 immunohistochemistry in bladder inverted papilloma. Proceedings of the 27th Annual International Conference of the Engineering in Medicine and Biology Society (EMBS '05); September 2005; pp. 6484–6487. [DOI] [PubMed] [Google Scholar]
  • 26.Milstein N. Technion Israel Institute of Technology, The Faculty for Computer Sciences; 1998. Image segmantation by adaptive thresholding. [Google Scholar]
  • 27.Stiene S., Lingemann K., Nuchter A., Hertzberg J. Contour-based object detection in range images. Proceedings of the 3rd International Symposium on 3D Data Processing, Visualization, and Transmission; 2006; pp. 168–175. [Google Scholar]
  • 28.Haralick R. M., Shanmugam K., Dinstein I. Textural features for image classification. IEEE Transactions on Systems, Man and Cybernetics. 1973;3(6):610–621. doi: 10.1109/TSMC.1973.4309314. [DOI] [Google Scholar]
  • 29.Mokji M. M., Abu Bakar S. A. R. Adaptive thresholding based on co-occurrence matrix edge information. Proceedings of the 1st Asia International Conference on Modelling & Simulation (AMS '07); March 2007; Phuket, Thailand. pp. 444–450. [DOI] [Google Scholar]
  • 30.Mokji M. M., Bakar S. A. R. A. Gray level co-occurrence matrix computation based on haar wavelet. Proceedings of the Computer Graphics, Imaging and Visualisation: New Advances (CGIV '07); August 2007; Bangkok, Thailand. pp. 273–279. [DOI] [Google Scholar]
  • 31.Weszka J. S., Dyer C. R., Rosenfeld A. Comparative study of texture measures for terrain classification. IEEE Transactions on Systems, Man and Cybernetics. 1976;6(4):269–285. doi: 10.1109/tsmc.1976.5408777. [DOI] [Google Scholar]
  • 32.Chang C. I., Chen K., Wang J., Althouse M. L. G. A relative entropy based approach to image thresholding. Pattern Recognition. 1993;120:215–227. [Google Scholar]
  • 33.Kapur J. N., Sahoo P. K., Wong A. K. C. A new method for gray-level picture thresholding using the entropy of the histogram. Computer Vision, Graphics, & Image Processing. 1985;29(3):273–285. doi: 10.1016/0734-189x8590125-2. [DOI] [Google Scholar]
  • 34.Lee S. S., Horng S. J., Tsai H. R. Entropy thresholding and its parallel algorithm on the reconfigurable array of processors with wider bus networks. IEEE Transactions on Image Processing. 1999;8(9):1229–1242. doi: 10.1109/83.784435. [DOI] [PubMed] [Google Scholar]
  • 35.Pal N. R., Pal S. K. Entropic thresholding. IEEE Transactions on Signal Processing. 1989;16(2):97–108. doi: 10.1016/0165-16848990090-x. [DOI] [Google Scholar]
  • 36.Pun T. A new method for grey-level picture thresholding using the entropy of the histogram. Signal Processing. 1980;2(3):223–237. doi: 10.1016/0165-1684(80)90020-1. [DOI] [Google Scholar]
  • 37.Chanwimaluang T., Fan G. An efficient blood vessel detection algorithm for retinal images using local entropy thresholding. Proceedings of the 2003 International Symposium on Circuits and Systems (ISCAS '03); 2003; pp. 21–24. [Google Scholar]
  • 38.Li Y., Cheriet M., Suen C. Y. A threshlod selection method based on multiscale and graylevel co-occurrence matrix analysis. Proceedings of the 8th International Conference on Document Analysis and Recognition; September 2005; Seoul, Republic of Korea. pp. 575–578. [DOI] [Google Scholar]
  • 39. Research Section, Digital Retinal Image for Vessel Extraction (DRIVE) Database, Utrecht, The Netherlands, University Medical Center Utrecht, Image Sciences Institute, http://www.isi.uu.nl/Research/Databases/DRIVE/

Articles from Computational and Mathematical Methods in Medicine are provided here courtesy of Wiley

RESOURCES