Abstract
Background/objectives
We aim to develop an objective fully automated Artificial intelligence (AI) algorithm for MNV lesion size and leakage area segmentation on fluorescein angiography (FA) in patients with neovascular age-related macular degeneration (nAMD).
Subjects/methods
Two FA image datasets collected form large prospective multicentre trials consisting of 4710 images from 513 patients and 4558 images from 514 patients were used to develop and evaluate a deep learning-based algorithm to detect CNV lesion size and leakage area automatically. Manual segmentation of was performed by certified FA graders of the Vienna Reading Center. Precision, Recall and F1 score between AI predictions and manual annotations were computed. In addition, two masked retina experts conducted a clinical-applicability evaluation, comparing the quality of AI based and manual segmentations.
Results
For CNV lesion size and leakage area segmentation, we obtained F1 scores of 0.73 and 0.65, respectively. Expert review resulted in a slight preference for the automated segmentations in both datasets. The quality of automated segmentations was slightly more often judged as good compared to manual annotations.
Conclusions
CNV lesion size and leakage area can be segmented by our automated model at human-level performance, its output being well-accepted during clinical applicability testing. The results provide proof-of-concept that an automated deep learning approach can improve efficacy of objective biomarker analysis in FA images and will be well-suited for clinical application.
Subject terms: Retinal diseases, Macular degeneration, Medical imaging
Introduction
Age-related macular degeneration (AMD) is the leading cause for vision loss affecting the population over 50 in the developed world [1].
Retina specialists rely heavily on retinal imaging in the diagnosis, monitoring and therapy of patients with neovascular AMD (nAMD). Although optical coherence tomography (OCT) and in recent years OCT Angiography (OCTA) became the most common diagnostic modalities for diagnosis and monitoring, fluorescein angiography (FA) is still considered as the gold standard of solidifying the diagnosis of nAMD [2]. The most important biomarker of macular neovascularization (MNV) activity seen in FA is leakage area, which is characterized by increasing hyperfluorescence originating from an active MNV lesion seen in late phase images [3].
FA biomarkers such as presence, size and location of CNV lesions as well as its lesion type and activity are very useful in the clinical diagnostics and research [4, 5].
FA evaluation is especially critical in the context of clinical studies for determining study eligibility of patients, tracking disease progression or analyzing the efficacy of treatment. This requires manual segmentation, which is time-consuming, subjective as it suffers from high inter- and intra-grader variability and has to be performed by retina experts [6]. Particularly in large multicentre studies, these drawbacks are very significant, as the process of manual annotation does not scale well with big data, and needs to be harmonized among multiple retina experts within as well as across different reading centres involved in each study. Therefore, developing an automated and objective way to detect and segment CNV lesion size and leakage area is desirable. This would not only allow analyzing FA images in real-time, but also offers the possibility of conducting even large-scale analyses fast and objective to reduce variability of manual readings. Deep-learning based convolutional neural networks (CNN) have already been successfully applied to a multitude of medical imaging tasks outside of ophthalmology such as vessel [7] or brain tissue segmentation [8], and this technique has also shown impressive performance in multiple retinal imaging tasks [9].
In this study, we developed deep learning-based models for fully automated segmentation of CNV lesion size and leakage area. The proposed method additionally provides an uncertainty estimation of the prediction, being important for adopting into clinical routines and eventual subsequent revision by humans [10].
Materials/subjects and methods
Dataset
This study was conducted in adherence to the tenets of the Declaration of Helsinki, and ethics approval was obtained by the Ethics Committee of the Medical University of Vienna Submission Nr 1246/2016. The study is a retrospective data analysis and the ethics committee decided that informed patient consent is not necessary.
In this study, two datasets from large prospective multicentre trials for evaluation of anti-vascular endothelial growth factor drugs in patients with nAMD were used, consisting of FA images from the Vienna Reading Center (VRC) image database. The first dataset ‘FA-CNV’ is comprised of 4710 FA images containing CNV lesions from 513 eyes of 513 patients. The second dataset ‘FA-Leakage’ consists of 4558 FA images from 514 eyes of 514 patients.
Manual ground truth image segmentation was performed by certified and trained FA graders of the VRC in a standardized manner to reach best possible human objectiveness. CNV lesion size was annotated in ‘FA-CNV’ and leakage area was annotated in ‘FA-Leakage’ dataset for each individual image. The images to be segmented were selected manually from the best images of the FA image series between 45–90 s for CNV lesions and 5–10 min for leakage area.
All FA images in the datasets have been acquired using different conventional FA cameras certified by the VRC following a standardized image acquisition protocol and fulfilling VRC’s minimal resolution criteria. In particular, the FA images were acquired using FF540+, Spectralis HRA + OCT, TRC50 & TRC50IX, Topcon 2000 and Megaplus 1,4 l from manufacturers Zeiss, Heidelberg, Topcon, Megaplus and Optovue.
No images were excluded due to bad quality. Instead, all images that had been manually annotated were also used for training the automated segmentation model. The used datasets are thus characterized by a large variability in terms of image quality, pixel size, field-of-view, contrast, viewing angle, imaging artefacts or motion blur.
Technical setup: Network architecture
We propose an AI-based approach using a convolutional neural network (CNN) architecture requiring labelled data [11].
In general, the model follows a U-Net [12] based architecture, with some adaptations. It consists of a contracting (encoder) and expanding (decoder) part, as shown in Fig. 1. Skip connections are used between and within each level, as they have shown to be beneficial for the overall performance of the network [13].
In order to facilitate annotations during clinical application, the proposed model also provides an uncertainty map. For each pixel the model both outputs a prediction and an uncertainty estimation of this prediction [14] (Fig. 1). In this context, uncertainty estimates quantify the confidence of the model regarding its prediction, indicating when and where we can trust the prediction of the model.
Experimental setup and statistical analysis
Both datasets were randomly split on a patient-distinct basis into 70% training, 10% validation, and 20% test set. While the training set was used for learning the optimal parameters of the model, the validation set was utilized to tune the hyperparameter settings such as the learning rate, number of model parameters or the dropout rate. The final generalization performance of the algorithm was evaluated in the test set. The split resulted in 3343/507/860 training/validation/test FA images in the ‘FA-CNV’ dataset and 3210/493/860 in the ‘FA-Leakage’ dataset.
We assessed the networks segmentation performance by comparing the model predictions with the manual annotations. F1 score, Precision and Recall were used for quantitative evaluation, which are well-known and widely used metrics for comparing segmentation results.
Precision indicates the amount of true positive pixels among all predicted target pixels. Recall is the amount of true positive pixels among all annotated target pixels of the ground truth. In other words, recall is a measure for how good the model is in detecting pixels of the target class, while precision indicates how relevant the detected target pixels are. The F1 score is the harmonic mean of precision and recall used to balance both [15].
We also evaluated the impact of image quality on the segmentation performance of the model. For this we used an internal automated image quality assessment algorithm for FA, providing a score in the range [0,1] for multiple image quality categories (‘overall-quality’, ‘noise’, ‘focus’, ‘contrast’), with 0 indicating bad and 1 good quality. We then computed the Spearman correlation coefficient between these image quality scores and the F1 score of the AI segmentation models on both datasets, across all images in the test set. A description of the automated quality assessment algorithm is provided in the Supplemental information.
Evaluation of clinical applicability
To evaluate the clinical applicability of the deep-learning model, an extensive evaluation was performed by two trained retina experts of the VRC (GM, BH). For each of the FA images of the test set, both masked experts judged the quality of the manual as well as the automated segmentations as ‘good segmentation’ vs. ‘bad segmentation’. In addition, both experts had to decide which segmentation they preferred for each image. Both experts were masked for this evaluation, meaning that they did not know which image showed the manual annotation and which showed the automated segmentation result coming from the deep learning model.
Results
Quantitative and qualitative results
Table 1 shows an overview of the results of the quantitative evaluations including the average F1 score, Precision and Recall for both datasets. The F1 score for the FA-Leakage dataset (0.73) is slightly higher than the F1 score for the FA-CNV dataset (0.65), as are precision and recall. A quantitative comparison of the presented AI algorithm with a baseline is provided in the Supplemental information.
Table 1.
Dataset | Average F1 score ± standard deviation | Average precision ± standard deviation | Average recall ± standard deviation |
---|---|---|---|
FA-CNV | 0.65 ± 0.26 | 0.75 ± 0.29 | 0.72 ± 0.28 |
FA-Leakage | 0.73 ± 0.21 | 0.80 ± 0.23 | 0.78 ± 0.25 |
On average, the automatic segmentation together with its uncertainty estimation by the AI model required 90 milliseconds of computation time for a single FA image, using a ‘GeForce GTX 1080 Ti’ GPU. The whole segmentation pipeline including loading, pre-processing, segmentation, uncertainty estimation, post-processing and saving of the image took around 3 s, while this pipeline was not optimized for speed.
Qualitative segmentation results with uncertainty estimates are presented both for the FA-CNV and FA-Leakage dataset in Fig. 2. Typically, the highest uncertainty is present at the edges of segmentations, and in areas of heterogeneous appearance in the centre of the leakage area. Figure 2 depicts a representative set of cases to allow insights into the characteristics of the dataset and results, including the heterogeneity of images (e.g., low contrast in third row of Fig. 2b, differences in field-of-view between third and fourth row in Fig. 2a), improved segmentation of the AI model in comparison with manual annotations (first row of Fig. 2b) or false positive predictions of the algorithm (third row of Fig. 2b).
We observed that the performance of the proposed deep learning models is stable even when facing a wide variety of different input image appearances. This is supported by the Spearman correlation coefficients between the image quality scores and the F1 scores being close to zero: 0.005/0.04/0.02/0.01 in ‘FA-CNV’ and 0.04/0.06/0.07/0.02 in ‘FA-Leakage’ for overall-quality/noise/focus/contrast, respectively.
Evaluation of clinical applicability
The deep learning based automated segmentations of the proposed approach were preferred more often than their human counterparts for both datasets (Fig. 3a).
Both automated predictions and manual human annotations were judged as being of good quality in the majority of cases (Fig. 3b). Compared to manual human annotations, automated predictions were judged more often as good quality, indicating that the proposed deep learning models achieve at least human-level performance in both the ‘FA-CNV’ and the ‘FA-Leakage’ datasets.
Discussion
In order to streamline the segmentation of CNV lesions and their FA leakage areas, a new deep learning-based approach using uncertainty estimation was proposed. The models produce segmentations in a fast, precise, and upscale-able way, achieving human-level performance in our clinical applicability evaluation.
Manual annotations suffer from a wide variety of problems, such as the time it takes to produce high-quality annotations or the subjectivity leading to high inter-, or intra-reader variability. In contrast, automated models are objective and able to segment in real time, offering the potential for big data analysis and for producing high-quality segmentations even in small or understaffed clinics.
In recent years OCT and OCT angiography (OCTA) more or less replaced FA in routine clinical applications, as these examinations are non-invasive, and faster than FA. Although for the clinical diagnosis of a treatment-naïve CNV, OCT and OCTA are sufficient in most cases, in a fair number of cases FA is still necessary, as only this examination depicts the dynamics of dye leakage over time. Furthermore, although wide-field OCTA systems are being developed, ultra-wide-field FA systems are readily available, and thus are far superior in detecting changes outside of the macular area [16].
Together with signs of exudation in OCT, FA parameters such as CNV lesion size and area of FA leakage are still highly important biomarkers in monitoring disease activity and recurrence during anti-VEGF therapy, and as such are key part of any clinical trial investigating the efficacy of new drugs. Currently such analyses are performed manually by trained graders. Although the graders in our study are certified from a highly standardized renown centre, the VRC, perform numerous alignment trainings, and their gradings follow standardized protocols, inter- and intra-grader reproducibility of such readings is still only moderate to fair [17, 18]. Artificial intelligence (AI) can further improve the reproducibility of the grading of these biomarkers. In future research, combining multiple imaging modalities resulting in high-quality information about the CNV lesion could be achieved. For example, by localizing potential lesions using FA in 2D first, and secondly getting more detailed information by using 3D OCTA. In this context, the proposed model could be also used in future studies to identify potential still unknown „patterns of leakage“ in OCTA by automatically segmenting leakage areas in FA and mapping the corresponding area to OCTA, in datasets where both FA and OCTA are available. Mapping the automatic segmentation in FA to OCTA would allow training an automatic leakage segmentation model for OCTA, learning “leakage patterns” in OCTA without involving manual labels. This may help to replace invasive FA with non-invasive OCTA, that allows faster acquisition, and carries no risk of the potentially adverse reactions [16].
For CNV segmentation, deep learning-based segmentation has been proposed in prior publications for OCTA scans only [19]. For FA images, parametric modelling was used to segment CNV lesions, the algorithm however requiring FA images of each phase, leading to failure in case of non-providence [5]. In contrast, in our work we propose a fully automated segmentation model that operates on individual FA images, providing an output segmentation map using a single FA image as input. On the one hand, this overcomes the drawback of needing multiple images of several phases as input for the segmentation model. On the other hand, this also means that capturing the leakage dynamic can only be done indirectly by segmenting multiple subsequent images.
In the clinical applicability evaluation our models reached human-level performance, meaning that more automated segmentations than manual human segmentations were preferred by the retina experts. With an overall selection rate for ‘FA-Leakage’ of 52.2% and ‘FA-CNV’ of 51.9%, the predictions of the proposed deep learning-based models can be considered as high-quality segmentations on-par with human annotators. The results show that the automated predictions even slightly outperform the manual segmentations in our clinical applicability evaluation. This is a remarkable outcome, especially as the used datasets were particularly challenging due to the heterogeneous composition featuring images of different sharpness, contrast, resolution, field-of-view, etc.
The proposed models allow producing uncertainty estimations simultaneously with its predictions, providing valuable visualizations for subsequent processing or usage. The visualized uncertainty maps in Fig. 2 align with our expectations, being high in border regions and areas of low contrast or heterogeneous appearance. In general, uncertainty maps support the procedure of human revision and offer insights when and where the prediction should be re-evaluated. We believe that this may be a cornerstone of clinical application, helping to open the black box of a convolutional network a bit.
The results did not show a correlation between image quality and segmentation accuracy, meaning that the performance of the AI models was robust with respect to image quality. This demonstrates the potential of the algorithms, as this correlation has been shown to be present with humans [20]. In the application scenario where a subsequent check of the automated segmentation result by a human is intended, an image quality detection algorithm could be utilized as a pre-processing step to ensure a certain quality of the overall pipeline [21, 22].
While both segmentation models show stable and accurate results in general, one limitation is that the models tend to over-segment target areas in complex cases such as multiple pathologies appearing within the same image. However, we also observed that these cases show high uncertainty at the same time and could therefore be identified in clinical practice. Examples are provided in the Supplemental information. Compared to undersegmentation, oversegmentation is less severe, as no lesion would be missed (e.g., in a screening setting). Nevertheless, this limitation with respect to oversegmentation could be addressed in future work by using more training data featuring multiple pathologies, or by using a specific loss function during training that penalizes oversegmentation of the model in a stronger way. One key aspect of FA evaluation in patients with CNV is the determination of dynamic leakage compared with static staining. Our AI algorithm cannot determine this directly, and can only give indirect indication of a dynamic process by providing an accurately measured growth in the area covered by hyperfluorescence when being applied to multiple subsequent images (indicating leakage). Further development is needed to be able to differentiate between these two angiographic features automatically.
In conclusion, we have presented a deep learning-based approach for segmenting CNV lesion sizes and leakage area in individual FA images. The additional capability of the method to provide uncertainty estimations of the automatically created predictions constitutes a key concept for clinical applicability besides real-time performance and scalability. Moreover, the results showing human-level performance of the proposed models in both tasks demonstrate the potential of the method for clinical practice.
Summary
What was known before
The reproducibility of manual grading in FA is moderate to low.
Due to low intra- and inter-grader reproducibility, the usefulness of biomarkers seen in FA cannot be fully deployed, particularly in the comparison of novel therapies.
Manual grading is also time-consuming and costly, hindering analyses on a large-scale basis.
What this study adds
A new AI segmentation model that provides a fully automated, fast and precise segmentation of CNV lesions and leakage area in FA.
The model also provides an uncertainty estimation to quantify the confidence of the model regarding its prediction, allowing better interpretability.
The deep learning algorithm could have largest positive impact on the reproducibility of FA grading, particularly in the setting of randomized multicentre trials.
Supplementary information
Author contributions
Concept and design: DH, PS, BG, GD. Acquisition, analysis, or interpretation of data: All authors. Methodological development and statistical analysis: DH, PS. Drafting of the paper: DH, PS, BG, GD. Critical revision of the paper for important intellectual content: All authors.
Data availability
The datasets generated during and/or analyzed during the current study are not publicly available due to privacy restrictions but are available from the corresponding author on reasonable request.
Competing interests
DH: none; PS: none; BG: Consultant for Roche, Bayer, Novartis, Financial Research Support from Digital Diagnostics; GM: none; BHN: RetInSight; SE: Scientific consultancy for Genentech, Novartis, Roche, Heidelberg Engineering, Kodiak, RetInSight; GGD: none.
Footnotes
Publisher’s note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
These authors contributed equally: David Holomcik, Philipp Seeböck.
Supplementary information
The online version contains supplementary material available at 10.1038/s41433-022-02156-6.
References
- 1.Wong WL, Su X, Cheung CMG, Klein R, Cheng CY, Wong TY. Global prevalence of age-related macular degeneration and disease burden projection for 2020 and 2040: a systematic review and meta-analysis. Lancet Glob Health. 2014;2:e106-16. [DOI] [PubMed]
- 2.Schmidt-Erfurth U, Chong V, Loewenstein A, Larsen M, Souied E, Schlingemann R, et al. Guidelines for the management of neovascular age-related macular degeneration by the European Society of Retina Specialists (EURETINA) Br J Ophthalmol. 2014;98:1144–67. doi: 10.1136/bjophthalmol-2014-305702. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 3.Klein ML, Ferris FL, Armstrong J, Hwang TS, Chew EY, Bressler SB, et al. Retinal Precursors and the Development of Geographic Atrophy in Age-Related Macular Degeneration. Ophthalmology. 2008;115:1151026–31. [DOI] [PubMed]
- 4.Domalpally A, Danis RP. Fluorescein Angiography in Neovascular AMD. Rev. Opthalmol. 2008. https://www.reviewofophthalmology.com/article/fluorescein-angiography-in-neovascular-amd.
- 5.Walid A, Syed MS, Ahmed SF. Segmentation of Choroidal Neovascularization in Fundus Fluorescein Angiograms. IEEE Trans Biomed Eng. 2013;60:06. doi: 10.1109/TBME.2013.2237906. [DOI] [PubMed] [Google Scholar]
- 6.Shah SM, Tatlipinar S, Quinlan E, Sund JU, Tabandeh H, Nguyen QD, et al. Dynamic and Quantitative Analysis of Choroidal Neovascularization by Fluorescein Angiography. Investig Ophthalmol Vis Sci. 2006;47:5460–8. doi: 10.1167/iovs.06-0012. [DOI] [PubMed] [Google Scholar]
- 7.Ding L, Bawany MH, Kuriyan AE, Ramchandran RS, Wykoff CC, Sharma G. A Novel Deep Learning Pipeline for Retinal Vessel. IEEE Transact Image Proc. 2020;29:6561–73. [DOI] [PMC free article] [PubMed]
- 8.Zhang F, Breger A, Cho KIK, Ning L, Westin CF, O’Donnell L, et al. Deep learning based segmentation of brain tissue from diffusion MRI. NeuroImage. 2021;223:117934. [DOI] [PMC free article] [PubMed]
- 9.Schmidt-Erfurth U, Sadeghipour A, Gerendas BS, Waldstein SM, Bogunović H. Artificial intelligence in retina. Prog Retin Eye Res. 2018;67:1–29. [DOI] [PubMed]
- 10.Nair T, Precup D, Arnold DL, Arbel T. Exploring uncertainty measures in deep networks for Multiple sclerosis lesion detection and segmentation. Med Image Analy. 2020;59:101557. [DOI] [PubMed]
- 11.Yamashita R, Nishio M, Kinh Gian Do R, Togashi K. Convolutional neural networks: an overview and application in radiology. Insights Imaging. 2018;9:611–29. doi: 10.1007/s13244-018-0639-9. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 12.Ronneberger O, Fischer P, Brox T. U-Net: Convolutional Networks for Biomedical Image Segmentation. Med Image Comput Computer-Assist Intervention Springe Lncs. 2015;9351:234–41. [Google Scholar]
- 13.Drozdzak N, Vorontsov E, Chartrand G, Kadoury S, Pal C. The Importance of Skip Connections in Biomedical Image Segmentation. Deep Learning Data Labeling Med Appl. 2016:179–87.
- 14.Kendall A, Gal Y. What Uncertainties Do We Need in Bayesian Deep Learning for Computer. In 31st International Conference on Neural Information Processing Systems; 2017;30.
- 15.Sokolova M, Japkowicz N, Szpakowicz S. Beyond Accuracy, F-Score and ROC: A Family of Discriminant Measures for Performance Evaluation. Adv Artif Intell. 2006;4304:1015–21. [Google Scholar]
- 16.de Carlo TE, Salz DA, Waheed NK, Baumal CR, Duker JS, Wiktin JA. Visualization of the retinal vasculature using side-field montage optical coherence tomography angiography. Ophthalmic Surg Lasers Imaging Retina. 2015;46:611. [DOI] [PubMed]
- 17.Kaiser RS, Berger JW, Williams GA, Tolentino MJ, Maguire AM, Alexander J, et al. Variability in fluorescein angiography interpretation for photodynamic therapy in age-related macular degeneration. Retina. 2002;22:683–90. doi: 10.1097/00006982-200212000-00001. [DOI] [PubMed] [Google Scholar]
- 18.Goldstein M, Zait S, Yassur Y, Loewenstein A. Reproducibility in Grading Size, Leakage and Classical Component of Subfoveal Choroidal Neovascularization by Fluorescein Angiography. Investig Ophthalmol Vis Sci. 2005;46:1395–1395. [Google Scholar]
- 19.Wang J, Hormel TT, Gao L, Zang P, Guo Y, Wang X, et al. Automated diagnosis and segmentation of choroidal neovascularization in OCT angiography using deep learning. Biomed Opt Express. 2020;11:927–44. [DOI] [PMC free article] [PubMed]
- 20.Reiner BI. Hidden costs of poor image quality: a radiologist’s perspective. J Am Coll Radiol. 2014;11:974–8. doi: 10.1016/j.jacr.2014.04.008. [DOI] [PubMed] [Google Scholar]
- 21.Li HH, Abraham JR, Sevgi DD, Srivastava SK, Hach JM, Whitney J, et al. Automated quality assessment and image selection of ultra-widefield fluorescein angiography images through deep learning. Transl Vis Sci Technol. 2020;9:52–52. doi: 10.1167/tvst.9.2.52. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 22.Sadeghipour A, Arikan M, Ismail O, König M, Baltali B, Gerendas BS, et al. Imageability and Registration of Multimodal Imaging using Machine Learning. Investig Ophthalmol Vis Sci. 2019;60:2197–97. [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Supplementary Materials
Data Availability Statement
The datasets generated during and/or analyzed during the current study are not publicly available due to privacy restrictions but are available from the corresponding author on reasonable request.