Skip to main content
Journal of Imaging Informatics in Medicine logoLink to Journal of Imaging Informatics in Medicine
. 2024 Aug 8;38(1):520–533. doi: 10.1007/s10278-024-01219-2

Retina Blood Vessels Segmentation and Classification with the Multi-featured Approach

Usharani Bhimavarapu 1,
PMCID: PMC11811302  PMID: 39117940

Abstract

Segmenting retinal blood vessels poses a significant challenge due to the irregularities inherent in small vessels. The complexity arises from the intricate task of effectively merging features at multiple levels, coupled with potential spatial information loss during successive down-sampling steps. This particularly affects the identification of small and faintly contrasting vessels. To address these challenges, we present a model tailored for automated arterial and venous (A/V) classification, complementing blood vessel segmentation. This paper presents an advanced methodology for segmenting and classifying retinal vessels using a series of sophisticated pre-processing and feature extraction techniques. The ensemble filter approach, incorporating Bilateral and Laplacian edge detectors, enhances image contrast and preserves edges. The proposed algorithm further refines the image by generating an orientation map. During the vessel extraction step, a complete convolution network processes the input image to create a detailed vessel map, enhanced by attention operations that improve modeling perception and resilience. The encoder extracts semantic features, while the Attention Module refines blood vessel depiction, resulting in highly accurate segmentation outcomes. The model was verified using the STARE dataset, which includes 400 images; the DRIVE dataset with 40 images; the HRF dataset with 45 images; and the INSPIRE-AVR dataset containing 40 images. The proposed model demonstrated superior performance across all datasets, achieving an accuracy of 97.5% on the DRIVE dataset, 99.25% on the STARE dataset, 98.33% on the INSPIREAVR dataset, and 98.67% on the HRF dataset. These results highlight the method’s effectiveness in accurately segmenting and classifying retinal vessels.

Keywords: A/V classification, Multi-level features, Parallel channel attention, Tiny vessels

Introduction

Common eye diseases, including hypertensive retinopathy (HR), diabetic retinopathy (DR), and glaucoma, predominantly originate from alterations in the blood vessels of the retina—a light-sensitive membrane [1]. Prolonged hypertension can result in arteriovenous nicking and systemic arteriolar stenosis [2]. Diabetes, on the other hand, may lead to the formation of numerous abnormal blood vessels prone to rupture [3]. Additionally, the avascular macular region, typically devoid of blood vessels, can experience the undesirable development of new blood vessels [4].

The implementation of computer-aided diagnosis (CAD) systems in ophthalmology is crucial for enhancing diagnostic accuracy and efficiency, particularly in identifying eye conditions related to vascular alterations and arteriovenous (A/V) states [5]. These systems leverage advanced image processing algorithms and machine learning techniques to automatically detect and delineate retinal blood vessels from fundus images. CAD systems provide quantitative assessments of vessel characteristics such as diameter, tortuosity, and crossing patterns, essential for early detection of diseases like diabetic retinopathy and hypertensive retinopathy.

In clinical settings, there is a strong preference for automated segmentation and classification due to the labor-intensive and subjective nature of manual identification in fundus images [68]. Expert segmentation can vary significantly [912], underscoring the need for automated, highly precise techniques in retinal vascular segmentation. The complex morphological structure and abundance of tiny vessels, often with widths of one pixel or fewer than ten pixels, present challenges in accurately separating the retinal vascular tree from the background. CAD systems alleviate variability and subjectivity, ensuring consistent and reliable results across different clinical settings while enabling timely intervention and treatment planning by monitoring disease progression with greater precision and confidence [13].

Likewise, the presence of uneven illumination and lesion zones introduces challenges in differentiating non-vascular tissues from blood vessels, leading to minimal contrast. The precise segmentation of retinal vessels in fundus images remains a formidable task due to these factors, especially when dealing with small and low-contrast vessels. Microvessels, characterized by a limited number of pixels, play a crucial role in diagnosing neovascular disorders, emphasizing the importance of focusing on small vessels rather than larger ones. The process of identifying and separating distinct retinal structures, referred to as retinal segmentation, proves valuable in the diagnosis and treatment of various eye disorders [14].

Yang et al. [15] proposed a strategy that integrates U-Net with deformable convolution for identifying retinal blood vessels. Their model achieved sensitivity (SE), specificity (SP), accuracy (Acc), and area under the curve (AUC) of 0.8115, 0.9780, 0.9568, and 0.9810 on the DRIVE dataset, and 0.8075, 0.9841, 0.9664, and 0.9872 on CHASE_DB1, respectively. Guo et al. [16] introduced a generative adversarial network with U-Net for end-to-end processing to segment retinal vessels, achieving 0.9772 AUROC, 0.9058 AUPRC, and 0.8215 dice on the DRIVE dataset. Despite these advancements, the challenge of segmenting thin vessels and maintaining vessel connectivity persists due to inefficient use of contextual information. Vascular connections and thin blood vessels are crucial for diagnosing vascular disorders [17, 18]. By combining the ROI feature vector composition method with a neural network, they correctly classified 95.32% of vessel center lines between two crossings. The Swin-Unet model, featuring encoder and decoder with pure transformer modules and a symmetrical architecture similar to U-Net, has been proposed for medical image segmentation [19, 20], achieving accuracy, precision, recall, and F1 values of 98.65, 87.19, 88.63, and 87.20, respectively. Wen et al. [21] presented an approach for segmenting the optic disc and cup in the retina using a combination of CNN and transformer. Features are extracted through convolution and then processed through transformer and multi-scale convolution modules, enhancing segmentation performance. Extensive experiments showed improvements of about 7.6% in overall accuracy, 10.9% in sensitivity, and 9.2% in specificity compared to previous methods. Hong et al. [22] achieved multi-organ segmentation by integrating the transformer with the U-Net model, demonstrating a Dice similarity coefficient (DSC) of 80.68%. The transformer module addresses global interactions after low-level feature extraction using convolutions, enabling more accurate and detailed organ segmentation. Several methods have attempted to address these challenges. Yang et al. proposed integrating U-Net with deformable convolution for retinal blood vessel identification, and Guo et al. introduced a generative adversarial network (GAN) with U-Net for end-to-end retinal vessel segmentation. While these methods have shown improvements in segmentation accuracy, they still struggle with thin vessel segmentation and vessel connection due to inefficient contextual information utilization. Accurately segmenting and maintaining the connectivity of thin blood vessels is vital for diagnosing vascular disorders, highlighting the need for innovative approaches in retinal vascular segmentation.

Despite recent advances and the introduction of numerous techniques aimed at mitigating existing drawbacks, arterial/venous (A/V) classification of retinal vessels remains a challenging issue. Firstly, the shape and architecture of vessels exhibit considerable variation owing to the intricate environment within the human eye. The curvature of blood vessels further intensifies the complexity of accurately segmenting and classifying them. Secondly, the subtle peripheral vascular differences make distinguishing between arteries and veins more challenging, as they exhibit minimal distinctions, particularly towards the extremities of blood vessels. These challenges are compounded by the difficulty in accurately segmenting thin vessels and maintaining vessel connectivity, both of which are crucial for clinical diagnostics.

In response to these challenges, we introduce a unique approach that minimizes feature map inconsistencies and enhances feature map compatibility between different layers through the utilization of a multi-scale features unit. Our proposed deep learning end-to-end technique for A/V classification and segmentation aims to alleviate the aforementioned issues. By leveraging an enhanced convolutional neural network (CNN) architecture combined with attention mechanisms, this approach improves upon the limitations of traditional methods by effectively capturing and utilizing contextual information to distinguish between arteries and veins, even with subtle differences. The encoding section converts input images into feature maps using convolutional layers, while retina vessel extraction focuses on identifying and extracting all blood vessels. A/V extraction goes a step further by classifying these vessels into arteries and veins through additional processing. The integration of convolutional networks with attention mechanisms further refines feature extraction, striking a balance between model complexity and interpretability. This design choice optimizes the model’s efficiency in clinical settings, where both accuracy and practical usability are paramount. By incorporating multi-level attention mechanisms, the model enhances its ability to accurately segment thin vessels and maintain vessel connectivity, thereby advancing the field of retinal vessel segmentation and classification.

Research Contributions

The primary contributions are enumerated as follows:

  1. A novel approach to A/V classification is put forth, breaking the work into three distinct phases: vessel feature extraction, artery, and vein classification, and tuning.

  2. An attention module is presented to selectively emphasize key aspects beneficial for a given activity while suppressing redundant, irrelevant information, therefore fusing multi-level features in an adaptive and effective manner.

  3. We implemented a parallel channel and added it to the up sample to highlight the boundary data of small vessels.

  4. Extensive experimental findings on a variety of public datasets demonstrate the novel approach’s feasibility, accurate A/V categorization, and superior performance when compared to current state-of-the-art methodologies.

Methodology

We used an ensemble filter approach to enhance the retina images and a three-phase multi-scale and multi-feature to segment and classify the blood vessels. The proposed framework is shown in Fig. 1.

Fig. 1.

Fig. 1

Proposed framework

In Fig. 1, the encoding section of the neural network model is located in the top left corner and is labeled “Encoder.” This section involves the initial processing of input images, where the images are converted into feature maps. Convolutional layers with a 1 × 1 filter size (Conv 1 × 1) are then applied to these feature maps to extract relevant features. Vessel extraction focuses on identifying all blood vessels in an image, while A/V extraction specifically distinguishes between arteries and veins. A/V extraction is managed by the “Multi Feature A/V Extractor” component. This component further processes the extracted vessel features to separate and identify arteries and veins. It involves additional convolutional layers and transposition operations that help in differentiating between arteries and veins within the feature maps. The output from this component specifically identifies arteries and veins, which are then used for further processing or classification.

Pre-processing

To improve the contrast of the fundus image, we applied the ensemble filter approach. First, we applied the bilateral filter to remove the noise in the fundus images and it further preserves the edges.

1NfaDSgd(a-b)Sgr(Ia-Ib)Ib 1
Nf=aDSgd(a-b)Sgr(Ia-Ib) 2

where the parameters gd and gr denotes the strength of the filter, Nf represents the factor for normalization, Sgd denotes the spatial gaussian which reduces the distant filter influence, and Sgr denotes the Gaussian range which reduces the pixel influence with the value of the intensity Ia.

We later applied the Laplacian edge detector to smoothen the edges and extracted the region of interest. To enhance the contrast of the image, we applied the SUACE algorithm. The orientation map is represented as:

kp(p,q)=p=m-x2m+x2q=n-y2·n+y22λap,qλbp,q 3
kp(p,q)=p=m-x2m+x2q=n-y2·n+y2λa2p,qλb2p,q 4
(p,q)=12cos-1kpp,qkQp,q 5
exp-12pδ2Sp4+qδ2Sq4Tan2Πfpδ 6
pδ=pcosδ+qsinδ 7
qδ=-pcosδ+qsinδ 8

where δ denotes the orientation for the filter, f is the frequency for the cosine wave, and Sp4,Sq4 represents the skewness for the Gaussian curve.

Methodology

This study employs a convolutional neural network (CNN) with attention mechanisms for retinal vessel segmentation and arteriovenous (A/V) classification. The methodology involves multiple phases, including vessel extraction, semantic feature extraction, decoding, and attention matrix computation, to enhance segmentation accuracy and resilience.

Vessel Extraction

During the vessel extraction step, a complete convolutional network processes the input image first. To create a more accurate vessel map that resembles a tree, groups of tiny kernel convolution layers extract and reorganize the detailed information of the vessels. By incorporating the attention operation, the second phase improves the long-term modeling’s perception capability and increases its resilience to slight variations. The vessel features extracted in the previous step, along with the arteriovenous (A/V) feature maps extracted in this phase, are eventually sent into the subsequent phase.

Semantic Feature Extraction

Semantic features are extracted by the encoder at various levels within the image. We utilized an Attention Module, which improved the depiction of blood vessels and preserved the image’s exact position information to better extract image features. Four encoders that retrieved semantic information at various image levels provided several inputs. Cascaded transposition operations with multiple levels and stages are utilized; the multi-vessel feature extractor effectively gathers comprehensive information across various levels, improving the sensitivity of the vessel region.

Decoder Phase

In the decoder phase, upsampling is employed to progressively recover feature maps. Variations in the degree of multi-scale transposition data are merged with each upsampling step. Finally, segmentation results are generated by the parallel channel attention, which refines the depiction of blood vessels in the retinal image.

Attention Matrix

An attention matrix is created in the third phase to determine the degree of resemblance between every pixel and every other pixel. It is necessary to focus more on vessel pixels than on overall pixel similarity. To achieve dependencies, we divide the attention duration into horizontal and vertical attention phases. We use five 1 × 1 convolutional layers for both the horizontal and vertical attention phases, followed by batch normalization and nonlinear activation. After that, we multiply the convolution layers by the transpose matrix and then use the softmax function to get the horizontal and vertical attention matrices. The transpose matrix refers to a matrix operation crucial for aligning features and computing attention scores in neural networks. By flipping matrices over their diagonals, this operation enhances the model’s ability to understand spatial relationships and accurately depict details like thin blood vessels in retinal images.

Channel Attention Module

A parallel channel attention module is considered in the second phase to stimulate and improve the recovery of retinal vessels. The module receives the input feature map and processes it using convolution kernels to produce feature maps. The exact process is carried out simultaneously with attention weight coefficients for each channel produced, and global information about each feature map channel is acquired using the sigmoid activation function and global average pooling. The input is subjected to the attention weight coefficient to create the channel-weighted attention map.

Experimental Results

We compare the new model’s performance with the current model to assess the improvement’s impact. The model was verified using the STARE [23], Drive [24], INSPIREAVR [25], and HRF [26] datasets with identical parameter settings. Three phases proposed in this paper are combined to verify the network’s efficacy. The model was trained with a batch size of 256, an iteration interval of 100, and an initial learning rate of 0.001.

Image Collection

Gathered fundus images from datasets STARE [23], Drive [24], INSPIREAVR [25], and HRF [26]. The datasets used in this study include STARE with 400 images, DRIVE with 40 images, HRF with 45 images, and INSPIRE-AVR containing 40 images. Each dataset was then divided into training and testing sets using an 80:20 split while considering the individual datasets. This resulted in 320 images for training and 80 images for testing from the STARE dataset, 32 images for training and 8 images for testing from the DRIVE dataset, 36 images for training and 9 images for testing from the HRF dataset, and 32 images for training and 8 images for testing from the INSPIRE-AVR dataset. This approach ensured a balanced and comprehensive training and testing process, enhancing the model’s robustness and accuracy across different retinal imaging datasets.

Performance Measures

The positives and negatives in vessel segmentation relate to background and vessel pixels, respectively. The positives in the A/V classification represent the vein pixels, whereas the negatives represent the artery pixels true positives (TP) are defined as those that were accurately identified as vascular pixels. True negatives (TN) are pixels that are accurately identified as non-vascular. False positives (FP) are those pixels that were incorrectly identified as non-vascular. False negatives (FN) are pixels that are not vascular yet are mistakenly identified as vascular. We presented four metrics: accuracy, precision, and recall to assess the proposed performance and confirm it is feasible.

Precision evaluates the method’s capacity to correctly identify true positives (both arteries and veins), reflecting the accuracy of positive predictions made by the algorithm. Recall, on the other hand, examines the algorithm’s capacity to identify true positives, ensuring that most actual positives are correctly detected.

Accuracy=TP+TNTP+FP+TN+FN 9
Precision=TPTP+FP 10
Recall=TPTP+FN 11

Segmentation and A/V Evaluation

Figures 2 and 3 present segmentation results for the DRIVE and STARE datasets respectively, showcasing overall performance across different models. In contrast, Figs. 4 and 5 focus specifically on the segmentation of tiny vessels and provide a comparative analysis of vessel segmentation results across various models, highlighting the precision and effectiveness of each model in handling smaller and more intricate vessel structures.

Fig. 2.

Fig. 2

DRIVE dataset segmentation results (i) actual, (ii) ground truth, (iii) GDF-Net, (iv) CSG-Net, (v) MPS-Net, (vi) MFI-Net, (vii) DCU-Net, (viii) proposed

Fig. 3.

Fig. 3

STARE dataset segmentation results (i) actual, (ii) ground truth, (iii) GDF-Net, (iv) CSG-Net, (v) MPS-Net, (vi) MFI-Net, (vii) DCU-Net, (viii) proposed

Fig. 4.

Fig. 4

DRIVE dataset tiny vessel segmentation comparison results (i) actual, (ii) ground truth (iii) GDF-Net (iv) CSG-Net (v) MPS-Net (vi) MFI-Net (vii) DCU-Net (viii) proposed

Fig. 5.

Fig. 5

STARE dataset tiny vessel segmentation comparison results (i) actual, (ii) ground truth, (iii) GDF-Net, (iv) CSG-Net, (v) MPS-Net, (vi) MFI-Net, (vii) DCU-Net, (viii) proposed

Retinal vessel analysis has several limitations, one of which is the accurate identification of small vessels. Detecting small vessels enhances the overall performance by improving sensitivity and robustness. Two main issues arise when analyzing tiny vessels: they are often overlooked, and their omission impacts the algorithm’s sensitivity. Accurate detection of small vessels makes the method more robust. We conducted a performance comparison between our proposed technique and methods reported in GDF-Net [27], CSG-Net [28], MPS-Net [29], MFI-Net [31], and DCU-Net [15] for the detection of small vessels. As seen in Fig. 3, our proposed method produced more accurate results for small vessel detection compared to the other methods.

Figures 4 and 5 show the tiny vessel segmentation comparison results. Nevertheless, a poor segmentation effect can arise from accidentally segmenting the diseased area into retinal blood vessels. The proposed approach increases the receptive field’s range and decreases characteristic information loss. As a result, the segmentation result is more accurate since the features of small blood vessels and lesion areas are better restored.

The proposed methodology precisely identifies thin blood vessels and guarantees their connection by using tiny kernel convolution layers to extract and reorganize detailed vessel information, ensuring fine structures are accurately detected. The incorporation of attention operations enhances the model’s ability to focus on vessel-specific features, improving the detection of thin vessels. Multi-scale semantic feature extraction from various levels preserves detailed information, while upsampling in the decoder phase maintains vessel continuity by merging multi-scale data. Additionally, the attention matrix focuses on vessel pixel similarity, ensuring continuous vessel structures. Finally, the parallel channel attention module processes feature maps with global information, preserving even the smallest vessel connections throughout the segmentation process.

Locally magnified regions along with the ground truth and segmentation results generated by our proposed method and other techniques are shown in Fig. 6. The magnified image in the figure shows the differences with other models, and the yellow box in the figure represents the local magnification portion. Figure 6 illustrates how our system is able to more precisely identify thin blood vessels and guarantee blood vessel connection. These experimental results show that our approach can better preserve vascular anatomy and discriminate between vascular and non-vascular pixels. Our method is able to segment small and large retinal blood vessels accurately, as demonstrated by the results obtained with the proposed approaches.

Fig. 6.

Fig. 6

Magnified segmentation results (i) actual, (ii) ground truth, (iii) GDF-Net, (iv) CSG-Net, (v) MPS-Net, (vi) MFI-Net, (vii) DCU-Net, (viii) proposed

Figure 7 illustrates the comparison of A/V classification across various methods. The actual images (i) and ground truth (ii) provide the baseline for evaluation. GDF-Net (iii) shows moderate accuracy but struggles with fine vessel details. CSG-Net (iv) improves on GDF-Net by better distinguishing between arteries and veins, yet still misses some thin vessels. MPS-Net (v) offers enhanced vessel continuity but lacks precision in densely packed regions. MFI-Net (vi) maintains good vessel connectivity but occasionally misclassifies vessel types. DCU-Net (vii) performs well in both connectivity and classification but has slight inconsistencies in identifying smaller vessels. Our proposed method (viii) excels by accurately identifying thin vessels and ensuring their connection, closely matching the ground truth and outperforming the other models in both detail preservation and classification accuracy.

Fig. 7.

Fig. 7

A/V classification (i) actual, (ii) ground truth, (iii) GDF-Net, (iv) CSG-Net, (v) MPS-Net, (vi) MFI-Net, (vii) DCU-Net, (viii) proposed

Performance Comparison

Table 1 tabulates the performance comparison of various modes on the DRIVE dataset. Figure 8 shows the visualization for the Table 1 results. The proposed approach for vessel segmentation is compared with existing models GDF-Net [27], CSG-Net [28], MPS-Net [29], HHNet [30], MFI-Net [31], DCU-Net [15], MAGF-Net [32], and CSU-Net [33]. GDF-Net achieved an accuracy of 96.46%, with a precision of 92.63% and a recall of 93.67%. This model demonstrates a high accuracy, indicating it effectively classifies both arterial and venous instances. The precision and recall scores show balanced performance, with a slight emphasis on correctly identifying positive instances. CSG-Net followed closely with an accuracy of 96.57%, a precision of 92.57%, and a recall of 95.99%. This model exhibits consistent performance across all metrics, with particularly high recall, indicating its strength in correctly identifying true positive instances. MPS-Net exhibited an accuracy of 93.47%, with a precision of 95.35% and a recall of 96.56%. While it has a slightly lower accuracy, it excels in precision and recall, showing its effectiveness in correctly identifying both true positives and true negatives. HHNet achieved an accuracy of 94.55%, with a precision of 93.73% and a recall of 98.57%. This model shows balanced performance between precision and recall, with particularly high recall, indicating its strong performance in identifying true positive instances. CSU-Net demonstrated an accuracy of 92.46%, with a precision of 91.46% and a recall of 94.85%. Although it has the lowest accuracy among the models, it maintains a reasonable balance between precision and recall. MFI-Net achieved an accuracy of 93.63%, with a precision of 94.73% and a recall of 93.56%. This model shows strong performance in precision, indicating its effectiveness in correctly identifying positive instances, while maintaining a good balance with recall. DCU-Net had an accuracy of 93.38%, with a precision of 94.56% and a recall of 95.85%. This model demonstrates robust performance in both precision and recall, indicating its ability to correctly identify true positive instances. MAGF-Net achieved an accuracy of 94.38%, with a precision of 91.56% and a recall of 95.85%. It shows good performance in recall, indicating its strength in correctly identifying positive instances, although with a slightly lower precision compared to other models. The proposed model achieved an accuracy of 97.5%, with a high precision of 96.96% and a perfect recall of 100%. This model demonstrates outstanding performance across all metrics, with exceptional precision in correctly identifying both true positives and true negatives.

Table 1.

Performance comparison of segmentation models for DRIVE dataset

Model Accuracy Precision Recall
GDF-Net 96.46 92.63 93.67
CSG-Net 96.57 92.57 95.99
MPS-Net 93.47 95.35 96.56
HHNet 94.55 93.73 98.57
CSU-Net 92.46 91.46 94.85
MFI-Net 93.63 94.73 93.56
DCU-Net 93.38 94.56 95.85
MAGF-Net 94.38 91.56 95.85
Proposed 97.5 96.96 100

Fig. 8.

Fig. 8

Drive dataset performance comparison

Table 2 and Fig. 9 show the performance comparison for the STARE dataset. GDF-Net achieved an accuracy of 98.87%, with a precision of 98.63% and a recall of 98.67%. This model demonstrates excellent performance in both precision and recall, indicating its effectiveness in correctly identifying both true positives and true negatives. CSG-Net followed closely with an accuracy of 98.53%, a precision of 98.57%, and a recall of 98.99%. This model exhibits consistent performance across all metrics, with particularly high recall, indicating its strength in correctly identifying true positive instances. MPS-Net exhibited an accuracy of 97.47%, with a precision of 96.35% and a recall of 98.56%. While it has a slightly lower accuracy, it maintains high recall and precision, showing its effectiveness in correctly identifying both true positives and true negatives. HHNet achieved an accuracy of 96.55%, with a precision of 93.73% and a recall of 94.57%. This model shows balanced performance between precision and recall, with slightly lower scores compared to other models, indicating room for improvement in identifying true positive instances. CSU-Net demonstrated an accuracy of 97.46%, with a precision of 96.46% and a recall of 94.85%. Although it has a slightly lower recall, it maintains high precision, indicating its strength in correctly identifying positive instances. MFI-Net achieved an accuracy of 97.63%, with a precision of 94.73% and a recall of 93.56%. This model shows strong performance in precision, indicating its effectiveness in correctly identifying positive instances, while maintaining a good balance with recall. DCU-Net had an accuracy of 93.38%, with a precision of 94.56% and a recall of 95.78%. This model demonstrates robust performance in both precision and recall, indicating its ability to correctly identify true positive instances, although with slightly lower overall accuracy. MAGF-Net achieved an accuracy of 94.38%, with a precision of 91.56% and a recall of 97.85%. It shows good performance in recall, indicating its strength in correctly identifying positive instances, although with slightly lower precision and overall accuracy compared to other models. The proposed model achieved an outstanding accuracy of 99.25%, with a high precision of 99.68% and a recall of 99.39%. This model demonstrates exceptional performance across all metrics, with near-perfect precision and recall, indicating its superior ability to correctly identify both true positives and true negatives.

Table 2.

Performance comparison for STARE dataset

Model Accuracy Precision Recall
GDF-Net 98.87 98.63 98.67
CSG-Net 98.53 98.57 98.99
MPS-Net 97.47 96.35 98.56
HHNet 96.55 93.73 94.57
CSU-Net 97.46 96.46 94.85
MFI-Net 97.63 94.73 93.56
DCU-Net 93.38 94.56 95.78
MAGF-Net 94.38 91.56 97.85
Proposed 99.25 99.68 99.39

Fig. 9.

Fig. 9

STARE dataset segmentation performance comparison. a Accuracy. b Sensitivity. c Specificity

Figure 10 shows the classification results for veins, arteries, and background in the INSPIRE-AVR dataset. The confusion matrix reveals that veins are correctly classified 98% of the time, with 1% misclassified as arteries and 1% as background. Similarly, arteries are accurately identified 98% of the time, with only 1% misclassified as veins and 1% as background. The background class is correctly identified 99% of the time, with minimal misclassification. In terms of recall, the model achieves 98% for veins and arteries, and 99% for background. This indicates the model’s high sensitivity and its ability to correctly identify most true positive instances across all classes. The precision scores are also noteworthy, with 98% for veins, 98.99% for arteries, and 98.02% for background, highlighting the model’s accuracy in predicting each class with a low number of false positives. The overall accuracy scores for veins, arteries, and background are 98.67%, 99%, and 99% respectively, underscoring the model’s reliability and precision.

Fig. 10.

Fig. 10

Confusion matrix for INSPIRE-AVR dataset

Table 3 and Fig. 11 show the comparison of the INSPIRE-AVR dataset. The proposed approach for A/V classification is compared with existing models AC-Net [34], AV-Net [35], UA-Net [36], MSC-Net [37], U-Net [38], and VC-Net [39]. AC-Net achieved an accuracy of 94.87%, with a precision of 98.63% and a recall of 98.67%. This model demonstrates strong performance in both precision and recall, indicating its effectiveness in correctly identifying both true positive and true negative instances. However, its overall accuracy is slightly lower compared to other models. AV-Net followed with an accuracy of 97.53%, a precision of 98.57%, and a recall of 96.99%. This model exhibits high precision, indicating its strength in correctly identifying positive instances, while maintaining a good balance with recall. The slightly lower recall suggests some room for improvement in identifying all true positive instances.UA-Net exhibited an accuracy of 96.47%, with a precision of 96.35% and a recall of 98.56%. While it has a slightly lower precision, it maintains high recall, showing its effectiveness in correctly identifying true positives. The overall performance is solid, although with slightly lower scores compared to other models. MSC-Net achieved an accuracy of 97.55%, with a precision of 98.73% and a recall of 98.57%. This model shows exceptional performance in both precision and recall, indicating its superior ability to correctly identify both true positive and true negative instances. The proposed model achieved the highest accuracy of 98.67%, with perfect precision of 98.98% and a recall of 99%. This model demonstrates outstanding performance, with exceptional precision indicating its effectiveness in correctly identifying all positive instances. The high accuracy and balanced recall underscore its superior performance in classification tasks.

Table 3.

Performance comparison of A/V classification models for INSPIREAVR dataset

Model Accuracy Precision Recall
AC-Net 94.87 98.63 98.67
AV-Net 97.53 98.57 96.99
UA-Net 96.47 96.35 98.56
MSC-Net 97.55 98.73 98.57
Proposed 98.33 98.98 99

Fig. 11.

Fig. 11

INSPIRE-AVR A/V classification performance comparison

Figure 12 shows the classification results for veins, arteries, and background in the HRF dataset. The confusion matrix indicates that veins are correctly classified 98% of the time, with only 1% misclassified as arteries or background, and a mere 1% misclassified in total. Arteries are accurately identified 99% of the time, with only 1% misclassified as background and no instances of veins being misclassified as arteries. The background class is also identified with a high accuracy rate of 99%, with only 1% of veins being misclassified as background. In terms of recall, the model achieves 98% for veins, 99% for arteries, and 99% for background, demonstrating its ability to correctly identify the most true positive instances across all classes. The precision scores are equally impressive, with 98.99% for veins, 99% for arteries, and 98.02% for background, indicating the model’s accuracy in predicting each class without a significant number of false positives. The overall accuracy scores for veins, arteries, and background are 99%, 99.33%, and 99% respectively, confirming the model’s high reliability and precision.

Fig. 12.

Fig. 12

Confusion matrix for HRF dataset

Table 4 and Fig. 13 show the comparison of the HRF dataset. AC-Net achieved an accuracy of 94.87%, with a precision of 98.63% and a recall of 98.67%. This model demonstrates strong performance in both precision and recall, indicating its effectiveness in correctly identifying both true positive and true negative instances. However, its overall accuracy is slightly lower compared to other models, suggesting the need to improve. AV-Net followed with an accuracy of 97.53%, a precision of 98.57%, and a recall of 96.99%. This model exhibits high precision, indicating its strength in correctly identifying positive instances, while maintaining a good balance with recall. The slightly lower recall suggests that it may miss some true positive instances, but overall, it performs well. UA-Net exhibited an accuracy of 96.47%, with a precision of 96.35% and a recall of 98.56%. While it has a slightly lower precision, it maintains high recall, showing its effectiveness in correctly identifying true positives. The overall performance is solid, though with slightly lower scores compared to some other models. MSC-Net achieved an accuracy of 97.55%, with a precision of 98.93% and a recall of 98.87%. This model shows excellent performance in both precision and recall, indicating its superior ability to correctly identify both true positive and true negative instances. The high scores across all metrics highlight its robustness and reliability. The proposed model achieved the highest accuracy of 98.67%, with a precision of 99% and a perfect recall of 99%. This model demonstrates outstanding performance, with exceptional recall indicating its effectiveness in correctly identifying all positive instances. The high accuracy and balanced precision underscore its superior performance in classification tasks.

Table 4.

Performance comparison of A/V classification models for HRF dataset

Model Accuracy Precision Recall
AC-Net 94.87 98.63 98.67
AV-Net 97.53 98.57 96.99
UA-Net 96.47 96.35 98.56
MSC-Net 97.55 98.93 98.87
Proposed 98.67 99 99

Fig. 13.

Fig. 13

HRF A/V classification performance comparison: a Accuracy. b Precision. c Recall

Conclusion

The proposed methodology for segmenting and classifying retinal vessels demonstrates remarkable effectiveness and accuracy, as evidenced by the results obtained from multiple datasets. By applying an ensemble filter approach and advanced pre-processing techniques, the method significantly enhances image contrast while preserving critical details. The comprehensive convolution network, combined with attention operations, successfully extracts and refines vessel features, resulting in highly precise vessel maps. The proposed model demonstrated exceptional performance across all datasets, achieving an accuracy of 97.5%, precision of 96.96%, and perfect recall of 100% on the DRIVE dataset; an outstanding accuracy of 99.25%, precision of 99.68%, and recall of 99.39% on the STARE dataset; the highest accuracy of 98.33%, precision of 98.98%, and recall of 99% on the INSPIREAVR dataset; and the highest accuracy of 98.67%, precision of 99%, and recall of 99% on the HRF dataset. These high metrics underscore the robustness and reliability of our approach in real-world applications. In conclusion, our method’s superior accuracy, precision, and recall make it a powerful tool for the segmentation and classification of retinal vessels. This can substantially aid in the early detection and diagnosis of retinal diseases, contributing to better patient outcomes and advancing the field of medical imaging. Future work will focus on enhancing the model’s ability to handle diverse retinal images from various populations and pathological conditions. We plan to integrate more advanced deep learning techniques to further improve segmentation accuracy and efficiency. Additionally, expanding the dataset to include more annotated samples will help refine the model’s performance and generalizability.

Declarations

Conflict of Interest

The authors declare no competing interests.

Footnotes

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

References

  • 1.Bellapianta A, Cetkovic A, Bolz M, Salti A. Retinal organoids and retinal prostheses: an overview. International Journal of Molecular Sciences. 2022 Mar 8;23(6):2922. [DOI] [PMC free article] [PubMed]
  • 2.Chew S, Colville D, Hutchinson A, Canty P, Hodgson L, Savige J. Obstructive sleep apnea, chronic obstructive pulmonary disease and hypertensive microvascular disease: a cross-sectional observational cohort study. Scientific Reports. 2022 Aug 3;12(1):13350. [DOI] [PMC free article] [PubMed]
  • 3.Kulikauskas MR, Bautch VL. The versatility and paradox of BMP signaling in endothelial cell behaviors and blood vessel function. Cellular and Molecular Life Sciences. 2022 Feb;79(2):77. [DOI] [PMC free article] [PubMed]
  • 4.Li W, Zhang H, Li F, Wang L. RPS-Net: An effective retinal image projection segmentation network for retinal vessels and foveal avascular zone based on OCTA data. Medical Physics. 2022 Jun;49(6):3830-44. [DOI] [PubMed]
  • 5.Saini DJ, Sivakami R, Venkatesh R, Raghava CS, Dwarkanath PS, Anwer TM, Smirani LK, Ahammad SH, Pamula U, Hossain MA, Rashed AN. Convolution neural network model for predicting various lesion-based diseases in diabetic macula edema in optical coherence tomography images. Biomedical Signal Processing and Control. 2023 Sep 1;86:105180.
  • 6.Sarki R, Ahmed K, Wang H, Zhang Y, Ma J, Wang K. Image preprocessing in classification and identification of diabetic eye diseases. Data Science and Engineering. 2021 Dec;6(4):455-71. [DOI] [PMC free article] [PubMed]
  • 7.Neto LC, Ramalho GL, Neto JF, Veras RM, Medeiros FN. An unsupervised coarse-to-fine algorithm for blood vessel segmentation in fundus images. Expert Systems with Applications. 2017 Jul 15;78:182-92.
  • 8.Wang X, Jiang X, Ren J. Blood vessel segmentation from fundus image by a cascade classification framework. Pattern Recognition. 2019 Apr 1;88:331-41.
  • 9.Chen Q, Leng T, Zheng L, Kutzscher L, Ma J, de Sisternes L, Rubin DL. Automated drusen segmentation and quantification in SD-OCT images. Medical image analysis. 2013 Dec 1;17(8):1058-72. [DOI] [PMC free article] [PubMed]
  • 10.Alex V, Motevasseli T, Freeman WR, Jayamon JA, Bartsch DU, Borooah S. Assessing the validity of a cross-platform retinal image segmentation tool in normal and diseased retina. Scientific reports. 2021 Nov 8;11(1):21784. [DOI] [PMC free article] [PubMed]
  • 11.Schmidt-Erfurth U, Sadeghipour A, Gerendas BS, Waldstein SM, Bogunović H. Artificial intelligence in retina. Progress in retinal and eye research. 2018 Nov 1;67:1-29. [DOI] [PubMed]
  • 12.Moradi M, Chen Y, Du X, Seddon JM. Deep ensemble learning for automated non-advanced AMD classification using optimized retinal layer segmentation and SD-OCT scans. Computers in Biology and Medicine. 2023 Mar 1;154:106512. [DOI] [PMC free article] [PubMed]
  • 13.Abdulla AA. Efficient computer-aided diagnosis technique for leukaemia cancer detection. IET Image Processing. 2020 Dec;14(17):4435-40.
  • 14.Atli I, Gedik OS. Sine-Net: A fully convolutional deep learning architecture for retinal blood vessel segmentation. Engineering Science and Technology, an International Journal. 2021Apr 1;24(2):271–83.
  • 15.Yang X, Li Z, Guo Y, Zhou D. DCU-net: A deformable convolutional neural network based on cascade U-net for retinal vessel segmentation. Multimedia Tools and Applications. 2022 May;81(11):15593-607.
  • 16.Guo X, Chen C, Lu Y, Meng K, Chen H, Zhou K, Wang Z, Xiao R. Retinal vessel segmentation combined with generative adversarial networks and dense U-Net. IEEE Access. 2020 Oct 23;8:194551-60.
  • 17.Kondermann C, Kondermann D, Yan M. Blood vessel classification into arteries and veins in retinal images. InMedical Imaging 2007: Image Processing 2007 Mar 2 (Vol. 6512, pp. 1401–1409). SPIE.
  • 18.Sambhav K, Grover S, Chalam KV. The application of optical coherence tomography angiography in retinal diseases. Survey of ophthalmology. 2017 Nov 1;62(6):838-66. [DOI] [PubMed]
  • 19.Cao H, Wang Y, Chen J, Jiang D, Zhang X, Tian Q, Wang M. Swin-unet: Unet-like pure transformer for medical image segmentation. InEuropean conference on computer vision 2022 Oct 23 (pp. 205–218). Cham: Springer Nature Switzerland.
  • 20.Li Y, Zhang Q, Zhou H, Li J, Li X, Li A. Cerebrovascular segmentation from mesoscopic optical images using Swin Transformer. Journal of Innovative Optical Health Sciences. 2023 May 11:2350009.
  • 21.Wen H, Zhao J, Xiang S, Lin L, Liu C, Wang T, An L, Liang L, Huang B. Towards more efficient ophthalmic disease classification and lesion location via convolution transformer. Computer Methods and Programs in Biomedicine. 2022 Jun 1;220:106832. [DOI] [PubMed]
  • 22.Hong Z, Chen M, Hu W, Yan S, Qu A, Chen L, Chen J. Dual encoder network with transformer-CNN for multi-organ segmentation. Medical & Biological Engineering & Computing. 2023 Mar;61(3):661-71. [DOI] [PubMed]
  • 23.STARE dataset, https://www.idiap.ch/software/bob/docs/bob/bob.db.stare/v1.0.0/, online accessed: 10-Aug -23
  • 24.DRIVE dataset:https://www.kaggle.com/datasets/andrewmvd/drive-digital-retinal-images-for-vessel-extraction;online accessed: 10-Aug -23
  • 25.Niemeijer M, Xu X, Dumitrescu AV, Gupta P, Van Ginneken B, Folk JC, Abramoff MD. Automated measurement of the arteriolar-to-venular width ratio in digital color fundus photographs. IEEE Transactions on medical imaging. 2011 Jun 16;30(11):1941-50. [DOI] [PubMed]
  • 26.Odstrcilik J, Kolar R, Budai A, Hornegger J, Jan J, Gazarek J, Kubena T, Cernosek P, Svoboda O, Angelopoulou E. Retinal vessel segmentation by improved matched filtering: evaluation on a new high-resolution fundus image database. IET Image Processing. 2013 Jun;7(4):373-83.
  • 27.Li J, Gao G, Yang L, Liu Y. GDF-Net: A multi-task symmetrical network for retinal vessel segmentation. Biomedical Signal Processing and Control. 2023 Mar 1;81:104426.
  • 28.Guo S. CSGNet: Cascade semantic guided net for retinal vessel segmentation. Biomedical Signal Processing and Control. 2022 Sep 1;78:103930.
  • 29.Lin Z, Huang J, Chen Y, Zhang X, Zhao W, Li Y, Lu L, Zhan M, Jiang X, Liang X. A high resolution representation network with multi-path scale for retinal vessel segmentation. Computer Methods and Programs in Biomedicine. 2021 Sep 1;208:106206. [DOI] [PubMed]
  • 30.Guo S. Fundus image segmentation via hierarchical feature learning. Computers in Biology and Medicine. 2021 Nov 1;138:104928. [DOI] [PubMed]
  • 31.Ye Y, Pan C, Wu Y, Wang S, Xia Y. MFI-Net: Multiscale feature interaction network for retinal vessel segmentation. IEEE Journal of Biomedical and Health Informatics. 2022 Jun 13;26(9):4551-62. [DOI] [PubMed]
  • 32.Li J, Gao G, Liu Y, Yang L. MAGF-Net: A multiscale attention-guided fusion network for retinal vessel segmentation. Measurement. 2023 Jan 1;206:112316.
  • 33.Wang B, Wang S, Qiu S, Wei W, Wang H, He H. CSU-Net: A context spatial U-Net for accurate blood vessel segmentation in fundus images. IEEE Journal of Biomedical and Health Informatics. 2020 Jul 22;25(4):1128-38. [DOI] [PubMed]
  • 34.Ma, W., Yu, S., Ma, K., Wang, J., Ding, X. and Zheng, Y., 2019. Multi-task neural networks with spatial activation for retinal vessel segmentation and artery/vein classification. In Medical Image Computing and Computer Assisted Intervention–MICCAI 2019: 22nd International Conference, Shenzhen, China, October 13–17, 2019, Proceedings, Part I 22 (pp. 769–778). Springer International Publishing.
  • 35.Kang H, Gao Y, Guo S, Xu X, Li T, Wang K. AVNet: A retinal artery/vein classification network with category-attention weighted fusion. Computer Methods and Programs in Biomedicine. 2020 Oct 1;195:105629. [DOI] [PubMed]
  • 36.Galdran A, Meyer M, Costa P, Campilho A. Uncertainty-aware artery/vein classification on retinal images. In2019 IEEE 16th International Symposium on Biomedical Imaging (ISBI 2019) 2019 Apr 8 (pp. 556–560). IEEE.
  • 37.Yi J, Chen C. Multi-Task Segmentation and Classification Network for Artery/Vein Classification in Retina Fundus. Entropy. 2023 Jul 31;25(8):1148. [DOI] [PMC free article] [PubMed]
  • 38.Ronneberger O, Fischer P, Brox T. U-net: Convolutional networks for biomedical image segmentation. InMedical Image Computing and Computer-Assisted Intervention–MICCAI 2015: 18th International Conference, Munich, Germany, October 5–9, 2015, Proceedings, Part III 18 2015 (pp. 234–241). Springer International Publishing.
  • 39.J Hu H Wang Z Cao G Wu JB Jonas YX Wang J Zhang 2021 Jun Automatic artery/vein classification using a vessel-constraint network for multicenter fundus images Frontiers in cell and developmental biology. 11 9 659941 [DOI] [PMC free article] [PubMed]

Articles from Journal of Imaging Informatics in Medicine are provided here courtesy of Springer

RESOURCES