Glaucoma is an optic neuropathy that results in visual field defects, and its diagnosis requires assessment of multiple domains, including optic nerve assessment on fundus exam or optical coherence tomography imaging, visual field assessment, and intraocular pressure [1]. Despite the multidimensionality and complexity of glaucoma diagnosis, many researchers tried to develop machine learning models that aid in the diagnosis or assessment of glaucoma. It is expected that machine learning models will have important impact on glaucoma patient assessment in the near future [2]. Such models mostly depended on fundus photography of optic disc, as the most important domain in glaucoma diagnosis [3]. Fundus photography used to train machine learning models are either derived from local hospital settings, where expert ophthalmologists are involved in the project to diagnose glaucoma patients and provide the ground truth for the data included, or more commonly obtained from openly available datasets.
Currently, there are multiple openly accessible fundus photography datasets commonly used by machine learning researchers. Table 1 provides the details of these commonly used datasets. The following concerns are raised about these datasets and need to be considered by prospective researchers:
The criteria for glaucoma diagnosis in most datasets are vaguely described and are usually dependent on clinical decision by treating physician. The glaucoma diagnosis in the studies using the dataset is usually taken for granted.
Most datasets did not specify the severity of glaucoma, but usually included moderate to severe cases. Training models on moderate to advanced glaucoma will not yield a significant added value, as these cases do not impose a diagnostic dilemma for ophthalmologists and are usually already on treatment [4].
Another issue is also raised by “glaucoma mimickers” on fundus photography, where the optic disc may appear severely cupped without the presence of glaucoma. Such mimickers include large disc area, myopic features, or even physiological cupping [5]. Models will yield false positive results if presented with these fundus photographs if the model was not properly trained on. Openly accessible datasets rarely include fundus photographs for such cases.
Table 1.
Dataset | Fundus photographs | Confirmation by visual field | Confirmation by clinical exam | Presence of myopes | Presence of physiological cupping | Presence of normal tension glaucoma | Total sample size | Glaucoma sample size | Severity assessment | |
---|---|---|---|---|---|---|---|---|---|---|
Disc | Maula | |||||||||
ORIGA | Yes | No | No | No | No | No | No | 650 | 168 | No |
RIGA | Yes | Yes | No | No | No | No | No | 750 | 0 | No |
REFUGE | Yes | No | Yes | Yes | Yes | No | Yes | 1200 | 120 | No |
ACRIMA | Yes | No | No | Yes | No | No | No | 705 | 396 | No |
RIM-ONEv1 | Yes | No | No | Yes | No | No | Yes | 169 | 51 | Yes |
RIM-ONEv2 | Yes | No | No | Yes | No | No | No | 452 | 197 | No |
RIM-ONEv3 | Yes | No | No | Yes | No | No | No | 127 | 35 | No |
DRISHTI-GS1 | Yes | No | No | Yes | No | No | No | 101 | 80 | No |
High-Resolution Fundus (HRF) | Yes | Yes | No | Yes | No | No | No | 45 | 15 | No |
LES-AV | Yes | No | No | Yes | No | No | Yes | 22 | 11 | No |
As a low prevalence disease, a cost-effective model to screen glaucoma needs a sensitivity and specificity of more than 95% [6]. Training such a model will need a large dataset of mild to moderate glaucoma patients, along with data on glaucoma mimickers on fundus photography, a dataset that is not yet available for public use.
Compliance with ethical standards
Conflict of interest
The author declares no competing interests.
Footnotes
Publisher’s note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
References
- 1.Anon. Recommendations | Glaucoma: diagnosis and management | Guidance | NICE. https://www.nice.org.uk/guidance/ng81/chapter/Recommendations. Accessed 29 Jan 2021.
- 2.Mayro EL, Wang M, Elze T, Pasquale LR. The impact of artificial intelligence in the diagnosis and management of glaucoma. Eye. 2020;34:1–1. doi: 10.1038/s41433-019-0577-x. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 3.Thompson AC, Jammal AA, Medeiros FA. A review of deep learning for screening, diagnosis, and detection of glaucoma progression. Trans Vis Sci Tech. 2020;9:42–42. doi: 10.1167/tvst.9.2.42. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 4.Medeiros FA. Deep learning in glaucoma: progress, but still lots to do. Lancet Digital Health. 2019;1:e151–e152.. doi: 10.1016/S2589-7500(19)30087-1. [DOI] [PubMed] [Google Scholar]
- 5.Ran AR, Cheung CY, Wang X, Chen H, Luo L, Chan PP, et al. Detection of glaucomatous optic neuropathy with spectral-domain optical coherence tomography: a retrospective training and validation deep-learning analysis. Lancet Digital Health. 2019;1:e172–e182.. doi: 10.1016/S2589-7500(19)30085-8. [DOI] [PubMed] [Google Scholar]
- 6.Vaahtoranta-Lehtonen H, Tuulonen A, Aronen P, Sintonen H, Suoranta L, Kovanen N, et al. Cost effectiveness and cost utility of an organized screening programme for glaucoma. Acta Ophthalmol Scand. 2007;85:508–18. doi: 10.1111/j.1755-3768.2007.00947.x. [DOI] [PubMed] [Google Scholar]