Abstract
Results
The CNN trained using OCT alone showed a diagnostic accuracy of 94%, whilst the OCT-A trained CNN resulted in an accuracy of 91%. When multiple modalities were combined, the CNN accuracy increased to 96% in the AMD cohort.
Conclusions
Here we demonstrate that superior diagnostic accuracy can be achieved when deep learning is combined with multimodal image analysis.
1. Introduction
In the developed world, age-related macular degeneration (AMD) is the leading cause of irreversible vision loss in the population over 60 years old [1, 2]. Broadly AMD can be categorised in neovascular and non-neovascular AMD, with the later being far more prevalent. Characteristic features of non-neovascular AMD include macular drusen, RPE abnormalities, and, in the late stage, geographic atrophy [3]. Pigment abnormalities detected by colour fundus photography (CFP) are now well recognised to be one of the major risk factors for the development of late stage AMD [3–6].
Over time, optical coherence tomography (OCT) and fundus autofluorescence (FAF) have also joined the battery of imaging techniques that are now considered essential for the monitoring of non-neovascular AMD. This list has more recently been joined by optical coherence tomography angiography (OCT-A) and multicolour confocal scanning laser ophthalmoscopy (SLO) [7]. With the introduction of OCT and OCT-A guided AMD treatment regimens, there has been an exponential increase in the number of images obtained and stored in large electronic databases [8]. Furthermore, the multiple imaging techniques that are now routinely used in clinical practice mean that the cumulative time spent by retinal specialists interpreting these images is also increasing [8]. Utilising software or computation power to automate image analysis may have the potential to improve the efficiency and accuracy of this process in the clinic [9]. Computer-assisted image assessment is also less prone to human factors such as bias, fatigue, and mindset [10]. Computer-assisted diagnosis is not a new concept: radiology adopted this approach when the increasing demand for imaging studies outstripped the ability of radiologists to interpret and report on the studies [11].
Within the field of ophthalmology, automated image analysis has been applied to the detection of diabetic retinopathy, mapping visual field defects in glaucoma, and grading cataract [12–14]. Concentrating on macular disease, both semiautomated and automated techniques have already been used and validated in the detection of drusen, reticular drusen, and geographic atrophy [15–19]. The majority of these studies have developed networks using only a single imaging modality, namely, OCT [20–24], and very few have combined image data from more than one modality, e.g., CFP and OCT images [25], or infrared, green FAF, and SLO [26].
The purpose of this study is to determine whether a multimodal deep learning approach; training the CNN on OCT, OCT-A, and CFP, will diagnose intermediate dry AMD more accurately, when compared to conventional CNN trained on the single modalities of CFP, OCT, and OCT-A.
2. Methods
Seventy-five participants were recruited through Auckland Optometry Clinic and Milford Eye Clinic, Auckland, New Zealand. All participants provided written informed consent prior to imaging. Ethics approval (#018241) from the University of Auckland Human Participants Ethics Committee was obtained for this study. This research adhered to the tenets of the Declaration of Helsinki. Participants were divided into three groups: young healthy (YH), old healthy, (OH) and high-risk intermediate AMD (AMD). A total of 20 participants were recruited into the YH group, 21 into the OH group, and 34 into the AMD group. A comprehensive eye exam was conducted on each participant prior to the OCT and OCT-A scans including dilated fundal examination and high contrast best-corrected visual acuity (BCVA) to determine the ocular health status of the fundus. Patients with any posterior eye disease that could potentially affect the choroidal or retinal vasculature including but not limited to glaucoma, polypoidal choroidal vasculopathy, DR, hypertensive retinopathy, and high myopia (≥6 D) were excluded from the study. Patients with any history of neurological disorders were also excluded. None of our recruited participants fit the exclusion criteria. The YH group consisted entirely of individuals between the ages of 20 and 26 and a best corrected visual acuity of ≥6/9 in the eye under test. The OH group consisted of individuals over the age of 55 years who had a best corrected visual acuity of ≥6/9 in the eye under test and a normal ocular examination. The “AMD cohort” consisted entirely of patients with high-risk intermediate dry AMD. This was diagnosed if the individual had at least two of the following risk factors: reticular pseudodrusen, established neovascular AMD in the fellow eye, and confluent soft drusen with accompanying changes within the retinal pigment epithelium. In order to ensure that all patients in the “AMD cohort” could maintain fixation during OCT-A imaging, only those patients with a BCVA of 6/15 or better were enrolled. The mean age of the participants in the YH, OH, and AMD groups were 23 ± 3, 65 ± 10, and 75 ± 8 years, respectively. Only one eye of each patient was used for the analysis, and if the patient had both eyes scanned, the OCT-A scan that had the better quality (assessed subjectively by the clinical grader) of the two was used. Mean BCVA for the YH, OH, and AMD groups were 6/6, 6/9, and 6/12, respectively. The ocular health of all participants was assessed at Auckland Optometry Clinic, by a registered optometrist, prior to enrolment in the study. The macular status of patients enrolled into the AMD group was assessed separately by an experienced retinal specialist (DS).
2.1. SS-OCT-A Device and Scanning Protocol
Participants were dilated with 1.0% tropicamide if the pupils were deemed too small for adequate OCT scans. The swept source (SS) OCT-A device (Topcon DRI OCT Triton, Topcon Corporation, Tokyo, Japan) was used to obtain the following images: 3 × 3 mm2 macular en-face OCT and 3 × 3 mm2 macular en-face OCT-A.
Raw OCT, OCT-A, and CFP image data were exported using Topcon IMAGEnet 6.0 software. No image processing was performed prior to image analysis. The retinal layers were identified using the IMAGENET 6.0 automated layer detection tool (Figure 1). En-face OCT and OCT-A images from layers 6 to 9 of the scan were selected (ONL-IS, OS, RPE, and choroid), plus a single fundus photo. The dataset was divided into 0.6/0.2/0.2 for training, validation, and test sets. This division was based on participants (not images) in order to avoid data leakage between the training, validation, and test sets. As there were multiple OCT and OCT-A images per patient, appropriate measures were taken to ensure that there was no patient overlap between the training, validation, or test sets (Figure 2).
Figure 1.

The retinal layers automatically identified by IMAGEnet 6.0 software. NFL: nerve fiber layer; GCL: ganglion cell layer; IPL: inner plexiform layer; INL: inner nuclear layer; OPL: outer plexiform layer; ONL: outer nuclear layer; IS: inner segment; OS: outer segment; RPE: retinal pigment epithelium.
Figure 2.

Examples of en-face OCT (top row), en-face OCT-A (middle row), and colour fundus photographs (bottom row) from each of the three cohorts (AMD, OH, and YH). The images within each column are from the same patient, from the labeled cohort at the top.
2.2. Convolutional Neural Network Design
The original design of the convolutional neural network (CNN) used here was based on the INCEPTION-RESNET-V2 design, since it appeared to have the advantage of faster convergence speed. This design was further modified to enable the network to be trained on multiple image modalities at the same time (Figure 3). Each imaging stream was then set up to initiate with a resizing layer, which was then followed by three repetitions of a 2D convolutional layer, batch normalization, and RELU activation layer. Separate modalities were then concatenated using a global pooling layer. The main body of the CNN followed the classic INCEPTION-RESNET-V2 design, where blocks of inception cells (A, B, and C) were used in series, where each block included cells with varying kernel sizes and channels, as well as feed forwards bypasses. The Python code for creating the CNN is released and freely available here: https://medium.com/@mannasiladittya/building-inception-resnet-v2-in-keras-from-scratch-a3546c4d93f0.
Figure 3.

Modification to the top layers of INCEPTION-RESNET-V2 architecture.
Experiments were run on an Intel Xeon Gold 6128 CPU @ 3.40 GHz with 16 GB of RAM memory and a NVIDIA GeForce TiTan V VOLTA 12 GB, for 100 epochs. The training-stop criterion was as CNN validation loss (measured as negative log-likelihood and residual sum of squares) had reached a stable minimum over the last 3 EPOCHs of training. Hence, any further training would have led to model “overfitting,” in which the neural network “memorizes” the training examples. It is worth noting that this approach is scalable to include more modalities (>3).
Several network training episodes were undertaken. Firstly, only a single image modality was used: OCT, OCT-A, or CFP. Further training episodes combined both OCT and OCT-A image data. The last training episode utilised multimodal image analysis from OCT, OCT-A, and CFP.
To better understand the nature of “learning” from our CNN in each modality, attention maps were produced from each modality. This method is explained in detail elsewhere [27–29]. Briefly, the output of the last convolutional layer prior to the global concatenation was extracted, resized, and smoothed for visualization.
3. Results
Table 1 shows that the sensitivity and specificity of the CNNs trained using a single modality are high. However, when more than one image modality is used during the training of the CNN, the sensitivity and specificity increase with each additional image modality added. The CNN with the highest accuracy, sensitivity, and specificity was the multimodal image-trained network.
Table 1.
The accuracy, sensitivity, and specificity of each CNN design (single modality vs. dual modality vs. multimodality) in classification of three cohorts of YH, OH, and AMD.
| Patient cohort | |||
|---|---|---|---|
| YH | OH | AMD | |
| Single modality CNN | Overall accuracy 94.4% | ||
| OCT sensitivity (%) | 99.6 | 98.9 | 77.8 |
| OCT specificity (%) | 98.8 | 86.7 | 100 |
| Single modality CNN | Overall accuracy 91.9% | ||
| OCT-A sensitivity (%) | 95.5 | 83.2 | 97.6 |
| OCT-A specificity (%) | 99.6 | 96.2 | 76.4 |
| Single modality CNN | Overall accuracy 93.8% | ||
| CFP sensitivity (%) | 81.0 | 84.6 | 100 |
| CFP specificity (%) | 94.4 | 78.6 | 86.7 |
| Dual modality CNN (OCT + OCT-A) | Overall accuracy 96.7% | ||
| Sensitivity (%) | 100 | 95.7 | 92.1 |
| Specificity (%) | 97.6 | 94.6 | 98.7 |
| Dual modality CNN (OCT + CFP) | Overall accuracy 92.9% | ||
| Sensitivity (%) | 98.1 | 96.3 | 100 |
| Specificity (%) | 100 | 97.0 | 96.0 |
| Multimodality CNN (OCT + OCT-A + CFP) | Overall accuracy 99.8% | ||
| Sensitivity (%) | 100 | 99.3 | 100 |
| Specificity (%) | 100 | 100 | 99.2 |
If the CNN results are considered for each imaging modality, it would suggest that each modality is better suited for classifying certain categories. The CFP single modality CNN was the most sensitive to AMD closely followed by OCT-A. In contrast, the single modality OCT was more sensitive to ageing, identifying the young and old cohorts more accurately. Combining the imaging modalities into a single “multimodal” CNN resulted in an accuracy of 99.8%, being able to identify both ageing and disease with high sensitivity and specificity (Table 1).
To investigate the apparent different sensitivities for each modality, “attention maps” of each modality were generated (Figure 4) from the same images in Figure 2. These maps demonstrate the image features that were “noticed” by multiple CNNs for each modality or participant group.
Figure 4.

Same images as Figure 2 were used to generate these attention maps. Normalized images [0-1] are pseudocoloured based on the level of “attention” the CNN gave to the region in the image modality (low represented by blue, to high represented by red).
The attention maps of the OCT images in each cohort pay highest attention to the background homogeneity. In the OCT-A images, the areas of highest attention are what appear to be projection artefacts of the retinal vessels seen within the choricapillaris OCTA slab, and the surrounding background OCTA signal of the adjacent tissues. In CFP, the area of highest attention is at the optic disc and the peripapillary region in young and AMD and additionally at the macula in the old cohort.
4. Discussion
In the current study, we wanted to compare the accuracy of different CNN designs, trained on the same dataset, to investigate whether combing imaging modalities improved the ability of the CNN to accurately identify the 3 distinct clinical groups under test. A secondary aim was to evaluate the attention maps of each design to investigate which components of the image(s), the CNN “paid attention” to.
To date, a number of single image modality CNNs designed to identify AMD have been published, with differing rates of success [30–36]. Furthermore, the majority of these studies have been based on single transverse OCT image (cross section) taken through the fovea [23, 37–40]. Although such an approach, training a CNN on a single OCT image, can achieve impressive results, this approach is flawed as a single transverse image will only sample a very limited part of the macula and relies on the pathology being present within the scan analysed.
The use of en-face OCT images may overcome the segmentation error and sampling bias associated with the use of transverse OCT scans. This technique has previously been utilised by Zhao et al. [33], to automatically detect the presence of drusen, including pseudodrusen. CFPs have also been used as a single imaging modality in the identification of AMD using an appropriately trained CNN [30, 32, 34–36, 41]. Good levels of accuracy have been obtained [30, 34], but again often only after image resizing and significant image preprocessing [36].
To the best of our knowledge, this is the first study to develop a CNN that utilises both en-face OCT and OCT-A and then combines them with CFP to develop a truly “multimodal” CNN, one that truly samples the entirety of the macular under test. To avoid sampling bias, the en-face OCT and OCT-A data slabs of the entire macula were used. We found that the multimodal CNN was superior to the single modality CNNs; moreover, incorporating additional modalities led to an incremental improved accuracy of the CNN.
To the best of our knowledge, only two other groups have utilised a similar multimodal approach [25, 42]. Yoo et al. [25] used paired fundus and OCT images, employing a VGG19 model pretrained on ImageNet to extract visual features from both images. More impressive is the approach taken by Wang et al. [42], who used a training method they termed “loose pairing” whereby pairings were constructed based on labels instead of eyes. In this method, a fundus image is allowed to be paired with an OCT image if their labels are identical, an approach which yielded an overall accuracy of 97%. The progressive improvement in the performance of the incrementally complex CNNs that Wang et al. [42] and ourselves describe strongly suggest that, like clinicians, the accuracy of a CNN to detect pathology and “normal” ageing is enhanced if complimentary imaging modalities are employed.
The outcome of our single modality CNNs (Table 1) suggest that OCT and OCT-A modalities are inherently sensitive to different aspects of the retinal health. It appears that OCT is more sensitive to ageing of the retina, while CFP and OCT-A are better suited to detect the pathological changes attributable to AMD. The black box nature of the neural networks makes any interpretation of the results problematic [43]. We therefore produced attention maps to aid our understanding of the CNN activity. The en-face OCT images in each cohort revealed that the highest attention was directed to the background homogeneity with a lesser emphasis on the fovea. OCT-A in contrast appears to be more sensitive to disease. Review of the attention maps reveals that within the OCT-A images, the regions of higher attention are the retinal vessels and the tissues immediately adjacent to them. Again the fovea appears to contribute very little to the OCT-A CNN classifier.
In conclusion, although trained on a small number of images, this study demonstrates that, compared with CNNs trained on a single image modality, superior diagnostic accuracy can be achieved when deep learning is combined with multimodal image analysis. We should also emphasise that this is only a “proof of concept” study, and larger studies to test its validity are needed. It should also be noted that the attention maps produced here are from our small cohort and may not prove to be generalisable after further studies. This approach is more similar to the multimodal image interpretation used by retinal specialists within the clinical environment, and our results suggest that this approach warrants further investigation. The limitations of this study include the small sample size of images, with the CNN trained using images from a single academic center. These clearly limit the generalisability of the CNNs we have trained, but arguably do not detract from the conclusion that a multimodal CNN is superior to a single modality CNN. Further research would include the utilisation of images from other OCT manufacturers, and validation of this CNN on a dataset from another academic or clinical institution.
Acknowledgments
This research was funded by the Ministry of Business, Innovation and Education (MBIE) Endeavour Fast Start Fund (UOAX1805).
Data Availability
The ophthalmic imaging data used to support the findings of this study were obtained under appropriate approval by the University of Auckland Ethics Committee and so cannot be made freely available. Requests for access to these data should be made to the corresponding author Dr Ehsan Vaghefi e.vaghefi@auckland.ac.nz, which will be then passed on to the University of Auckland Ethics Committee for further process.
Conflicts of Interest
The authors declare that they have no conflicts of interest.
References
- 1.Klein R., Klein B. E. K., Linton K. L. P. Prevalence of age-related maculopathy. Ophthalmology. 1992;99(6):933–943. doi: 10.1016/s0161-6420(92)31871-8. [DOI] [PubMed] [Google Scholar]
- 2.Klein R., Cruickshanks K. J., Nash S. D., et al. The prevalence of age-related macular degeneration and associated risk factors. Archives of Ophthalmology. 2010;128(6):750–758. doi: 10.1001/archophthalmol.2010.92. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 3.Ferris F. L., Davis M. D., Clemons T. E., et al. A simplified severity scale for age-related macular degeneration. Archives of Ophthalmology. 2005;123(11):1570–1574. doi: 10.1001/archopht.123.11.1570. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 4.Klein M. L., Ferris F. L., 3rd, Armstrong J., et al. Retinal precursors and the development of geographic atrophy in age-related macular degeneration. Ophthalmology. 2008;115(6):1026–1031. doi: 10.1016/j.ophtha.2007.08.030. [DOI] [PubMed] [Google Scholar]
- 5.Wang J. J., Rochtchina E., Lee A. J., et al. Ten-year incidence and progression of age-related maculopathy: the blue mountains eye study. Ophthalmology. 2007;114(1):92–98. doi: 10.1016/j.ophtha.2006.07.017. [DOI] [PubMed] [Google Scholar]
- 6.Klein R., Klein B. E. K., Knudtson M. D., Meuer S. M., Swift M., Gangnon R. E. Fifteen-year cumulative incidence of age-related macular degeneration: the Beaver dam eye study. Ophthalmology. 2007;114(2):253–262. doi: 10.1016/j.ophtha.2006.10.040. [DOI] [PubMed] [Google Scholar]
- 7.Garrity S. T., Sarraf D., Freund K. B., Sadda S. R. Multimodal imaging of nonneovascular age-related macular degeneration. Investigative Opthalmology & Visual Science. 2018;59(4):AMD48–AMD64. doi: 10.1167/iovs.18-24158. [DOI] [PubMed] [Google Scholar]
- 8.Lee C. S., Baughman D. M., Lee A. Y. Deep learning is effective for the classification of OCT images of normal versus age-related macular degeneration. Ophthalmology Retina. 2017;1(4):322–327. doi: 10.1016/j.oret.2016.12.009. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 9.Abramoff M. D., Garvin M. K., Sonka M. Retinal imaging and image analysis. IEEE Reviews in Biomedical Engineering. 2010;3:169–208. doi: 10.1109/rbme.2010.2084567. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 10.Schmitz-Valckenberg S., Göbel A. P., Saur S. C., et al. Automated retinal image analysis for evaluation of focal hyperpigmentary changes in intermediate age-related macular degeneration. Translational Vision Science & Technology. 2016;5(2):p. 3. doi: 10.1167/tvst.5.2.3. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 11.Smieliauskas F., MacMahon H., Salgia R., Shih Y.-C. T. Geographic variation in radiologist capacity and widespread implementation of lung cancer CT screening. Journal of Medical Screening. 2014;21(4):207–215. doi: 10.1177/0969141314548055. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 12.Abràmoff M. D., Lou Y., Erginay A., et al. Improved automated detection of diabetic retinopathy on a publicly available dataset through integration of deep learning. Investigative Opthalmology & Visual Science. 2016;57(13):5200–5206. doi: 10.1167/iovs.16-19964. [DOI] [PubMed] [Google Scholar]
- 13.Asaoka R., Murata H., Iwase A., Araie M. Detecting preperimetric glaucoma with standard automated perimetry using a deep learning classifier. Ophthalmology. 2016;123(9):1974–1980. doi: 10.1016/j.ophtha.2016.05.029. [DOI] [PubMed] [Google Scholar]
- 14.Gao X., Lin S., Wong T. Y. Automatic feature learning to grade nuclear cataracts based on deep learning. IEEE Transactions on Biomedical Engineering. 2015;62(11):2693–2701. doi: 10.1109/tbme.2015.2444389. [DOI] [PubMed] [Google Scholar]
- 15.van Grinsven M. J. J. P., Lechanteur Y. T. E., van de Ven J. P. H., et al. Automatic drusen quantification and risk assessment of age-related macular degeneration on color fundus images. Investigative Opthalmology & Visual Science. 2013;54(4):3019–3027. doi: 10.1167/iovs.12-11449. [DOI] [PubMed] [Google Scholar]
- 16.Chen Q., Leng T., Zheng L., et al. Automated drusen segmentation and quantification in SD-OCT images. Medical Image Analysis. 2013;17(8):1058–1072. doi: 10.1016/j.media.2013.06.003. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 17.Mora A. D., Vieira P. M., Manivannan A., Fonseca J. M. Automated drusen detection in retinal images using analytical modelling algorithms. BioMedical Engineering OnLine. 2011;10(1):p. 59. doi: 10.1186/1475-925x-10-59. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 18.Sivagnanavel V., Smith R. T., Lau G. B., Chan J., Donaldson C., Chong N. V. An interinstitutional comparative study and validation of computer aided drusen quantification. British Journal of Ophthalmology. 2005;89(5):554–557. doi: 10.1136/bjo.2004.046813. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 19.Schmitz-Valckenberg S., Brinkmann C. K., Alten F., et al. Semiautomated image processing method for identification and quantification of geographic atrophy in age-related macular degeneration. Investigative Opthalmology & Visual Science. 2011;52(10):7640–7646. doi: 10.1167/iovs.11-7457. [DOI] [PubMed] [Google Scholar]
- 20.Poplin R., Varadarajan A. V., Blumer K., et al. Prediction of cardiovascular risk factors from retinal fundus photographs via deep learning. Nature Biomedical Engineering. 2018;2(3):158–164. doi: 10.1038/s41551-018-0195-0. [DOI] [PubMed] [Google Scholar]
- 21.Krause J., Gulshan V., Rahimy E., et al. Grader variability and the importance of reference standards for evaluating machine learning models for diabetic retinopathy. Ophthalmology. 2018;125(8):1264–1272. doi: 10.1016/j.ophtha.2018.01.034. [DOI] [PubMed] [Google Scholar]
- 22.De Fauw J., Ledsam J. R., Romera-Paredes B., et al. Clinically applicable deep learning for diagnosis and referral in retinal disease. Nature Medicine. 2018;24(9):1342–1350. doi: 10.1038/s41591-018-0107-6. [DOI] [PubMed] [Google Scholar]
- 23.Kermany D. S., Goldbaum M., Cai W., et al. Identifying medical diagnoses and treatable diseases by image-based deep learning. Cell. 2018;172(5):1122–1131.e9. doi: 10.1016/j.cell.2018.02.010. [DOI] [PubMed] [Google Scholar]
- 24.Mansour R. F. Deep-learning-based automatic computer-aided diagnosis system for diabetic retinopathy. Biomedical Engineering Letters. 2018;8(1):41–57. doi: 10.1007/s13534-017-0047-y. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 25.Yoo T. K., Choi J. Y., Seo J. G., Ramasubramanian B., Selvaperumal S., Kim D. W. The possibility of the combination of OCT and fundus images for improving the diagnostic accuracy of deep learning for age-related macular degeneration: a preliminary experiment. Medical & Biological Engineering & Computing. 2019;57(3):677–687. doi: 10.1007/s11517-018-1915-z. [DOI] [PubMed] [Google Scholar]
- 26.Ly A., Nivison-Smith L., Assaad N., Kalloniatis M. Multispectral pattern recognition reveals a diversity of clinical signs in intermediate age-related macular degeneration. Investigative Opthalmology & Visual Science. 2018;59(5):1790–1799. doi: 10.1167/iovs.17-23076. [DOI] [PubMed] [Google Scholar]
- 27.Krizhevsky A., Sutskever I., Hinton G. E. Advances in Neural Information Processing Systems. New York, NY, USA: Curran Associates, Inc.; 2012. Imagenet classification with deep convolutional neural networks; pp. 1097–1105. [Google Scholar]
- 28.Cho K., Courville A., Bengio Y. Describing multimedia content using attention-based encoder-decoder networks. IEEE Transactions on Multimedia. 2015;17(11):1875–1886. doi: 10.1109/tmm.2015.2477044. [DOI] [Google Scholar]
- 29.Bahdanau D., Cho K., Bengio Y. Neural machine translation by jointly learning to align and translate. 2014. https://arxiv.org/abs/1409.0473.
- 30.Grassmann F., Mengelkamp J., Brandl C., et al. A deep learning algorithm for prediction of age-related eye disease study severity scale for age-related macular degeneration from color fundus photography. Ophthalmology. 2018;125(9):1410–1420. doi: 10.1016/j.ophtha.2018.02.037. [DOI] [PubMed] [Google Scholar]
- 31.Treder M., Lauermann J. L., Eter N. Automated detection of exudative age-related macular degeneration in spectral domain optical coherence tomography using deep learning. Graefe’s Archive for Clinical and Experimental Ophthalmology. 2018;256(2):259–265. doi: 10.1007/s00417-017-3850-3. [DOI] [PubMed] [Google Scholar]
- 32.Burlina P. M., Joshi N., Pacheco K. D., Freund D. E., Kong J., Bressler N. M. Use of deep learning for detailed severity characterization and estimation of 5-year risk among patients with age-related macular degeneration. JAMA Ophthalmology. 2018;136(12):1359–1366. doi: 10.1001/jamaophthalmol.2018.4118. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 33.Zhao R., Camino A., Wang J., et al. Automated drusen detection in dry age-related macular degeneration by multiple-depth, en face optical coherence tomography. Biomedical Optics Express. 2017;8(11):5049–5064. doi: 10.1364/boe.8.005049. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 34.Tan J. H., Bhandary S. V., Sivaprasad S., et al. Age-related macular degeneration detection using deep convolutional neural network. Future Generation Computer Systems. 2018;87:127–135. doi: 10.1016/j.future.2018.05.001. [DOI] [Google Scholar]
- 35.Acharya U. R., Hagiwara Y., Koh J. E. W., et al. Automated screening tool for dry and wet age-related macular degeneration (ARMD) using pyramid of histogram of oriented gradients (PHOG) and nonlinear features. Journal of Computational Science. 2017;20:41–51. doi: 10.1016/j.jocs.2017.03.005. [DOI] [Google Scholar]
- 36.Burlina P. M., Joshi N., Pekala M., Pacheco K. D., Freund D. E., Bressler N. M. Automated grading of age-related macular degeneration from color fundus images using deep convolutional neural networks. JAMA Ophthalmology. 2017;135(11):1170–1176. doi: 10.1001/jamaophthalmol.2017.3782. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 37.Russakoff D. B., Lamin A., Oakley J. D., Dubis A. M., Sivaprasad S. Deep learning for prediction of AMD progression: a pilot study. Investigative Opthalmology & Visual Science. 2019;60(2):712–722. doi: 10.1167/iovs.18-25325. [DOI] [PubMed] [Google Scholar]
- 38.Serrano-Aguilar P., Abreu R., Antón-Canalís L., et al. Development and validation of a computer-aided diagnostic tool to screen for age-related macular degeneration by optical coherence tomography. British Journal of Ophthalmology. 2012;96(4):503–507. doi: 10.1136/bjophthalmol-2011-300660. [DOI] [PubMed] [Google Scholar]
- 39.Sun W., Liu X., Yang Z. Automated detection of age-related macular degeneration in OCT images using multiple instance learning. Proceedings of the 9th International Conference on Digital Image Processing (ICDIP 2017); May 2017; Hong Kong, China. [DOI] [Google Scholar]
- 40.Wang Y., Zhang Y., Yao Z., Zhao R., Zhou F. Machine learning based detection of age-related macular degeneration (AMD) and diabetic macular edema (DME) from optical coherence tomography (OCT) images. Biomedical Optics Express. 2016;7(12):4928–4940. doi: 10.1364/boe.7.004928. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 41.Zheng Y., Vanderbeek B., Daniel E., et al. An automated drusen detection system for classifying age-related macular degeneration with color fundus photographs. Proceedings of the 2013 IEEE 10th International Symposium on Biomedical Imaging; April 2013; San Francisco, CA, USA. IEEE; pp. 1448–1451. [DOI] [Google Scholar]
- 42.Wang W., Xu Z., Yu W., et al. Two-stream CNN with loose pair training for multi-modal AMD categorization. 2019. https://arxiv.org/abs/1907.12023.
- 43.Castelvecchi D. Can we open the black box of AI? Nature. 2016;538(7623):20–23. doi: 10.1038/538020a. [DOI] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Data Availability Statement
The ophthalmic imaging data used to support the findings of this study were obtained under appropriate approval by the University of Auckland Ethics Committee and so cannot be made freely available. Requests for access to these data should be made to the corresponding author Dr Ehsan Vaghefi e.vaghefi@auckland.ac.nz, which will be then passed on to the University of Auckland Ethics Committee for further process.
