Abstract
Background
The ophthalmology field was among the first to adopt artificial intelligence (AI) in medicine. The availability of digitized ocular images and substantial data have made deep learning (DL) a popular topic.
Main text
At the moment, AI in ophthalmology is mostly used to improve disease diagnosis and assist decision-making aiming at ophthalmic diseases like diabetic retinopathy (DR), glaucoma, age-related macular degeneration (AMD), cataract and other anterior segment diseases. However, most of the AI systems developed to date are still in the experimental stages, with only a few having achieved clinical applications. There are a number of reasons for this phenomenon, including security, privacy, poor pervasiveness, trust and explainability concerns.
Conclusions
This review summarizes AI applications in ophthalmology, highlighting significant clinical considerations for adopting AI techniques and discussing the potential challenges and future directions.
Keywords: Artificial intelligence, Deep learning, Ophthalmology, Diabetic retinopathy, Glaucoma, Age-related macular degeneration
1. Introduction
As artificial intelligence (AI) technologies develop, ophthalmology has thrived partly due to its dependency on image-based clinical decision-making and investigations.1 Deep learning (DL) played a key role in this progress, which has led to breakthroughs in information technology via the utilization of tools for feature extraction.2 Recently, many AI researchers in ophthalmology used medical images to construct deep learning models to perform high-dimensional analyses. They have been used to automate screening and diagnosis of common vision-threatening diseases with expert-level accuracy, including diabetic retinopathy (DR),3 glaucoma,4 age-related macular degeneration (AMD),5 cataract and other anterior segment disease,6 with high accuracy. As deep learning technologies are maturing, researchers start trying to achieve disorder detection beyond the established scope of early-stage7 and prognosis prediction.8
Due to the broadness of AI research in ophthalmology, experts are increasingly agreeing that consolidating data foundation and adopting standardized reporting and regulatory guidelines are needed to enhance explainability, repeatability, security and ethical compliance.9, 10, 11 Even if the deployment of AI systems remains limited, this trend is expected to continue with the implementation of these frameworks to reduce inconsistencies among different studies. The development of high-performance DL systems alone is not sufficient to ensure eventual clinical translation of AI research, which poses a difficult conundrum to the clinical success of AI research. A wide range of factors must be aligned to achieve it, including algorithm development, clinical indication, security and public acceptance.12 As such, to successfully implement AI in ophthalmology, clinicians need to develop a deep understanding of the interactions among these factors to formulate a clear lab-to-clinic strategy.
In this study, we aim to review the current status and future prospect of AI in ophthalmology by analyzing a variety of issues such as the development of AI in ophthalmology, the challenges for its application and potential future directions.
2. Development of AI in ophthalmology
Table 1 sums the performance of AI algorithms, data size and diversity, etc. for each ophthalmology diseases.
Table 1.
Year | Reference | Topic | Dataset | AUC | SE | SP |
---|---|---|---|---|---|---|
2016 | Abramoff et al.14 | DR | CFPs, 1748 images | 0.980 | 96.8% | 87.0% |
2017 | Gargeya et al.15 | DR | CFPs, 75,137 images | 0.97 | 94% | 98% |
2017 | Ting et al.16 | DR, glaucoma, and AMD | CFPs, 494,661 images | 0.931–0.958 | 90.5%–100% | 87.2%–91.6% |
2020 | Pan et al.17 | DR lesions classification | FFAs, 4067 images | 0.870–0.965 | 79.7%–98.0% | 82.7%–99.5% |
2022 | Gao et al.18 | DR grading | FFAs, 11,214 images | 0.922–0.994 | / | / |
2020 | Varadarajan et al.20 | DME OCT-grading | CFPs, 7072 images | 0.89 | 85% | 80% |
2019 | Liu et al.23 | Glaucoma | CFPs, 274,413 images | 0.996 | 96.2% | 97.7% |
2021 | Sun et al.28 | Glaucoma | OCTs, 777 sets | 0.957 | 89.6% | 95.2% |
2020 | Thompson et al.29 | Glaucoma | OCTs, 20,806 images | 0.96 | 81% | 95% |
2019 | Phene et al.33 | Glaucoma | CFPs, 86,618 images | 0.945 | 80.0% | 90.2% |
2019 | Asaoka et al.34 | Glaucoma | OCTs, 4316 images | 0.937 | 82.5% | 93.9% |
2019 | Wen et al.37 | Forecast future Visual Fields | VFs, 32,443 VFs | / | / | / |
2022 | Huang et al.38 | Glaucoma visual field grading | VFs, 16,356 VFs | 0.93 | / | / |
2017 | Burlina et al.39 | AMD | CFPs, 213,997 iamges | 0.96 | 88.4% | 94.1% |
2019 | Peng et al.40 | AMD grading | CFPs, 59,302 images | 0.94 | 59.0% | 93.0% |
2020 | Yan et al.42 | AMD progression prediction | CFPs, 31,262 images | 0.85 | / | / |
2018 | Kermany et al.43 | AMD | OCTs, 207,130 images | 0.999 | 97.8% | 97.4% |
2020 | Erfurth et al.45 | Quantification of Fluid Volumes to AMD | OCTs, 24,362 images | / | / | / |
2021 | Yan et al.47 | AMD | OCTs, 56,091 images | 0.940–0.992 | 80.0%–96.5% | / |
2021 | Zhang et al.48 | Detect and quantify geographic atrophy | OCTs, 5049 images | / | / | / |
2021 | Xu et al.50 | AMD,PCV | CFPs and OCTs, 1099 CFPs, 821 OCTs | 0.939 | 88.8% | 95.6% |
2022 | Jin et al.51 | Identification of choroidal neovascularization activity in AMD | OCTs and OCTAs, 462 image pairs | 0.980 | 89.6% | 95.6% |
2011 | Chueng et al.52 | Cataract | Slit-lamp photographs, 5547 images | / | / | / |
2021 | Li et al.54 | Keratitis | Slit-lamp photographs, 6567 images | 0.998 | 97.7% | 98.2% |
2021 | Ye et al.55 | Myopic maculopathy | OCTs, 2342 images | 0.927–0.974 | 73.9%–92.8% | 84.8%–94.0% |
2019 | Yoo et al.56 | Identify candidate patients for corneal refractive surgery | Multi-instrument data, 13,201 subjects | 0.983 | / | / |
2020 | WANG et al.57 | Eyelid malignant melanoma | Pathological patches, 225,230 images | 0.989 | 94.7% | 95.3% |
2.1. Diabetic retinopathy
It's a globally recognized strategy for preventing blindness to screening for DR and prompt referral for treatment. The screening process can be conducted by a variety of medical professionals, including ophthalmologists, optometrists, general practitioners, screening technicians, and ophthalmologic technologists. The use of AI-assisted screening and staging of DR based on fundus images is currently among the most promising AI applications in clinical medicine. Recent studies have shown that such systems can reliably match experts' performance and, in some cases, perform better than them while being a more cost-effective13 and broader-reaching way to the traditional screening programs.
Abramoff et al. implemented a DL model to screening DR, which achieved an area under the curve (AUC) of 0.980, a sensitivity of 96.8% and a specificity of 87.0%.14 Gargeya et al. also used a DL system to detect DR15 using a publicly available dataset including 75,137 color fundus photographs (CFP) of diabetic patients to train and test the model to distinguish between healthy fundus images and DR images. The model achieved a sensitivity of 94% and a specificity of 98%, indicating that the AI-based algorithm could reliably perform screening of fundus photographs. Ting et al. published a clinically acceptable DRZ diagnostic system developed and tested based on 10 external datasets from the Singapore Integrated Diabetic Retinopathy Program, which ran over a period of 5 years in 6 different countries or regions, including Singapore, China, Hong Kong, Mexico, the United States and Australia. Their model demonstrated an AUC, sensitivity, and specificity of 0.936, 90.5% and 91.6%, respectively, achieving reliable diagnosis in multiple ethnic groups.16 Although most studies have developed robust DL models for DR screening and diagnosis based on CFP or optical coherence tomography (OCT) images, some studies focused on automatic DR lesion detection in fundus fluorescein angiography (FFA) images. Multi-label classification of non-perfusion areas, vascular leakages and microaneurysms were automatically classified based on DL models17 to construct an end-to-end DL system for staging DR severity.18 Further, DL technologies have also been used to determine the prevalence and related systemic cardiovascular risk factors for DR19 and predict diabetic macular edema (DME) severity based on OCT (AUC of 0.89, sensitivity of 85% and specificity of 80%) from two-dimensional fundus images.20 Furtherly, commercial products for DR screening have been developed, since American Food and Drug Administration (FDA) approved IDx-DR as the first automated AI diagnostic system21 and then the EyRIS SELENA22 received the permission for clinical use in the European Union.
2.2. Glaucoma
Glaucoma patients may lead to irreversible visual field (VF) loss unless the patients receive an early diagnosis and prompt intervention. This represents a well-defined clinical need that could benefit from AI use. Although AI research in glaucoma has faced several limitations, such as a lack of multimodal assessment and long-term natural progression, substantial progress has been made. Many researchers have successfully used AI for diagnosing glaucoma via structural changes, including retinal fundus photos16,23, 24, 25, 26, 27 and OCT.28, 29, 30 Zangwill et al. reported high accuracy in identifying glaucoma using an SVM classifier.31 Burgansky et al. used five classifiers such as SVM and machine learning analysis of OCT image data to assist in diagnosing glaucoma.32 Li et al. trained a machine-learning algorithm to detect glaucoma-like optic disc by defining the optic disc with a vertical cup-to-disc ratio of 0.7 on fundus photos, similar to glaucoma-like optic disc. The results showed that the algorithm could detect glaucoma optic neuropathy with considerable sensitivity (95.6%), specificity (92%) and AUC (0.986).4 In another study, Phene et al. built a DL model trained on over 80,000 CFPs to screen referable glaucoma with an AUC of 0.945. Besides, its performance was proved when applied to two other independent datasets and the AUC performance declined slightly to 0.855 and 0.881, respectively.33 Asaoka et al. reported a DL model trained for detecting early glaucoma with 4316 OCT images, which achieved an AUC of 93.7%, sensitivity of 82.5%, and specificity of 93.9%.34 Using over 4000 Anterior segment OCT (AS-OCT) scans, Xu et al. detected gonioscopic angle closure and primary angle closure disease (PACD) based on fully automated analysis with the AUC of 0.928 and 0.964.35
VF assessment plays a significant role in the clinical diagnosis and management of glaucoma. VF provides multiple validated parameters that could be useful to develop a DL system. Elze et al. developed an unsupervised algorithm to classify glaucomatous vision loss with 17 prototypes.36 They found that the unsupervised algorithm was useful in detecting VF loss in early glaucoma. DL models have been used to predict glaucomatous progression in the VF. Wen et al. trained a DL system that could generating a point-wise VF prediction for up to 5.5 years into the future, with a correlation of 0.92 between the MD of predicted and actual future HVF and an average difference of 0.41 dB37. The complexity of clinical situations requires fine-grained grading and comprehensive evaluation in order to provide an accurate diagnosis and management. Huang et al. proposed a DL system to grade VF of glaucoma with a high accuracy for data from two devices (Humphrey Field Analyzer and Octopus).38 This tool could be used by glaucomatous patients for self-assessment as well as promote telemedicine.
2.3. Age-related macular degeneration
Age-related macular degeneration is a main cause of irreversible vision loss in the elderly population. CFP is the most widely used screening method, which can identify drusen, geographic atrophy, retinal hemorrhage, and other lesions. Due to its fast, noninvasive and low-cost advantages, CFP plays a very important role in screening AMD populations. A DL algorithm based on CFP was able to fully automatically diagnose and grade AMD with an accuracy comparable to ophthalmologists. In 2018, Burlina et al. developed a DL algorithm in which feature extraction and classification were automatically completed on more than 130,000 CFP datasets. When compared with previous methods on the binary classification task, their DL algorithm demonstrated greater prospects for application in clinical practice.39 Meanwhile, Grassmann et al. defined a 13-category AMD fundus photography dataset, which included 12 AMD severity and one case that couldn't be graded because of poor image quality. They trained six advanced DL models and finally proposed an ensemble network on an untrained independent test set.5 Peng et al. used a DL algorithm called DeepSeeNet integrated by 3 seed networks to identify AMD severity-related characteristic lesions at the single eye level and could provide a final score combined with binocular severity at the patient level.40 In addition to diagnosing AMD based on CFP, some AI research has been focused on predicting the risk of disease progression. In 2019, Burlina et al. not only explored the diagnostic effect of the DL algorithm on the 4 and 9 classifications of AMD severity but also experimentally used a DL-based regression model to give patients a 5-year risk score for their estimated disease progression to advanced AMD, which further expands the application scope of the DL algorithm.41 In 2020, Yan et al. performed a study that combined CFP and patients' corresponding AMD-related genotypes and could predict the risk of advanced AMD progression based on DL algorithms.42
OCT can display the status and lesions of the macular region of the retina. In 2018, Kermany et al. used a transfer learning method that only used a small part of the data of traditional DL methods for training and applied it to an OCT dataset for choroidal neovascularization (CNV) and other three classifications. Their model achieved an accuracy of 96.6%, sensitivity of 97.8% and specificity of 97.4%, reaching the level of senior ophthalmologists.43 Recently, more and more studies are focusing on the quantitative analysis of OCT images using AI algorithms. Schlegl et al. developed a DL network which could automatically identify and quantify subretinal fluid (SRF) and intraretinal fluid (IRF) on OCT images and reported that their results were very close to expert annotations.44 Erfurth et al. also used a DL algorithm to identify and quantify retinal effusions, including SRF, IRF and pigment epithelial detachment, and explore the relationship between the amount of effusion and visual function after intravitreal injection in AMD patients.45 The quantitative study of OCT volume mode by Moraes et al. was not limited to retinal effusion but also used biomarkers such as hyperreflective foci and subretinal hyperreflective material on OCT images, which demonstrated results that were closely related to the treatment decision of AMD patients in follow-up reviews and demonstrated a high clinical application value.46 Yan et al. used an attention-based DL algorithm to interpret the activity of CNV on OCT images to assist doctor to diagnose AMD.47 In addition to wet AMD, Zhang et al. quantified geographic atrophy (GA) on OCT images using a DL model for determining retinal pigment epithelium loss, photoreceptor degeneration, and hyperprojection.48 Liefers et al. also quantified multiple key features on OCT images of early and late AMD patients as biomarkers associated with disease progression.49 The combined application of multiple modalities was shown to be closer to clinical practice implementation and is one of the hotspots in medical AI research. Xu et al. combined CFP and OCT images to diagnose AMD and polypoidal choroidal vasculopathy (PCV) and achieved an 87.4% accuracy, 88.8% sensitivity and 95.6% specificity.50 Jin et al. determined the effictiveness of a multimodal DL model using OCT and optical coherence tomography angiography (OCTA) images to assess CNV in neovascular AMD.51 The DL system reached an accuracy of 95.5% and an AUC of 0.9796 on multimodal data inputs, which was comparable to retinal specialists.
2.4. Cataract and other ophthalmologic diseases
AI research in ophthalmology used to pay more attention to the screening and diagnosing fundus diseases. The fact shows that AI has the potential for the diagnosis and management of various diseases, such as automated diagnosis and severity grading of cataracts based on slit-lamp or fundus photographs. AI agents have demonstrated good-to-excellent overall diagnostic performance in classifying different types of cataracts, with high AUC (0.86–1.0), accuracy (69.0%–99.5%), sensitivity (60.1%–99.5%), and specificity (63.2%–99.6%).52 Long et al. used a DL model to develop an artificial intelligence management application for congenital cataracts involving 3 functions: a congenital cataract screening network in the population, a risk stratification network for congenital cataract patients, and assisting a network of strategies for ophthalmologists to make treatment decisions.53 Li et al. improved the performance of a DL algorithm for diagnosing anterior segment diseases in slit-lamp images such as cataracts, keratitis and pterygium through segmenting the anatomical structures and annotating pathological lesions.6 Another DL system exhibited considerable performance automatically classifying keratitis, other cornea abnormalities, and normal cornea in slit-lamp images captured by the different devices and a smartphone with the super macro mode (all AUCs>0.96).54 The sensitivity and specificity of AI in keratitis detection were found to be comparable with experienced cornea specialists. Ye et al. developed a DL system for detecting and classifying myopic maculopathy in patients with high myopia. Their model achieved sensitivities equal to or even better than junior ophthalmologists.55 Yoo et al. combined preoperative data and compiled a machine learning model from 10,561 eye images and showed that their model could predict suitability for refractive surgery achieving an accuracy of 93.4% and an AUC of 0.97 in external validation.56 For ocular tumors, a large-scale statistical analysis of epidemiological and clinicopathological characteristics was performed in combination with public medical databases and multicenter clinical data to develop a set of convolutional neural networks for identifying malignant tumors. The DL diagnosis system for melanoma visualization could distinguish between benign and malignant tumors with an accuracy of 94.9% and a sensitivity of 94.7%.57
3. Challenges in the application of AI
3.1. Consolidate the data foundation
A major challenge of current DL models is that their training requires a large amount of data because insufficient data may decrease the performance of DL models.58 Thus, for some rare eye diseases, their low incidence makes it difficult for researchers to collect enough data for AI research.59 Some multimodal and longitudinal cohort studies have high requirements for patient data, and researchers need to collect data prospectively, which is very time- and energy-consuming. Hence, dealing with the limited amount or even no data in certain conditions still requires further attempts and efforts. Currently, most of the AI algorithms successfully applied in the field of ophthalmology use a fully supervised learning model which requires doctors to perform high-standard data annotation on the original data according to specific tasks. Medical data labeling is difficult to obtain because it's time-consuming and labor-intensive. Further, the gaps in labeling standards and levels among labelers should also be considered.
With the advancement of computer science, weakly supervised learning models that require only a small amount of annotation, semi-supervised learning models or even unsupervised learning models that do not require annotation have been proposed and applied in ophthalmology.60, 61, 62 It could reduce dependence on data labeling in a way, however the fact can't be ignored that various image-related factors have been shown to play more predominant parts than technical factors in determining the performance of AI models, suggesting DL training and application in real-world scenarios require robust training and testing datasets.63, 64
3.2. Reporting guidelines
A computational perspective on AI often overlooks clinically relevant details that clinicians believe are essential to determining effectiveness. For example, the criteria for participant recruitment, demographics, risk control and other abilities to administer the trail. Therefore, to better reflect the specific requirements for AI implementation, it is necessary to update some of the widely accepted international reporting protocols using extant evidence-based clinical trial workflows.65 In recent decades, many standardized reporting protocols have arisen, of which CONSORT-AI,66 STARD-AI,11 SPIRIT-AI67 and TRIPOD68 are among the most extensively applied. These protocols aremainly differentiated by the type of research and setting in which they are applicable and offer a list of items grouped according to different article sections. CONSORT69 was one of the first protocols and was introduced in 1996. It was updated in 2001 and 2010 to specify reporting guidelines for parallel-group randomized controlled trials. STARD70 was developed in 2000 to standardize studies comparing new or alternative diagnostic tests with defined reference standards. SPIRIT,71 which arose in 2007, covers general clinical intervention trials and incorporates some of the features of CONSORT by farther emphasizing the significance of pre-registration trials to reduce reporting selectivity and encourage transparency.72 In spite of the propagation of reported protocol expansions addressing AI in medicine, it should be acknowledged that AI is a fast developing technique, and the range of potential application would multiply in the not-too-distant future.73 Therefore, it is expectable that these protocol extensions would be updated while AI-specific protocols for clinical decision-making may be formalized in the future.
3.3. Security
Medical AI research should be performed on multicenter datasets; otherwise, the generalization of the proposed AI models might not be difficult. Data from a single-center database could be biased by factors such as patients' race and image equipment. Data transfers between research collaborators, especially for international collaborations, are often limited because of patient privacy and data security.74 Thus, the management of data from multicenter research centers in different countries is also a major challenge. One of the key shortcomings of the DL algorithm is its inexplicability.75 From the perspective of medical personnel, it is impossible to understand how DL models make decisions and predictions for specific tasks because existing methods such as heatmaps are inadequate to fully explain them. Therefore, the AI models should provide a certain degree of transparency to justify their medical diagnosis, treatment decisions and risk predictions so that medical personnel can fully trust them to further generate new clinical insights. Xu et al. proposed a possible approach to enhance the interpretability by comprehensively simulating the diagnostic thinking of human experts that they establish a hierarchical deep learning system that includes pre-diagnosis module, image segmentation module, and final diagnosis module and makes the AI diagnosis of glaucoma a visualized interactive process.27 Although DL models have achieved performance close to or even surpassed ophthalmologists in specific medical tasks, the real practical application of AI systems in complex ophthalmology clinics is still far from being achieved.12,76 Its problems include a relatively narrow range of disease spectrum applications, doctors’ interpretation of DL model results, integration of software systems and hardware devices, resistance to malicious attacks, and approval by relevant policies.
3.4. Ethics
As AI is gradually being integrated in clinical practice and medical professionals are getting used to it, there is also a growing trend for an ethical framework to guide the real application of DL systems. The concern roots in the possibility that AI could lead to discrimination against patients or worsen health inequalities. For example, vulnerable groups with complicated health problems may lose their priority if an AI health system is purely focused on maximizing efficiency.77 Psychosocial factors may also influence the acceptance and trust in AI-involved healthcare.78 Once AI systems turn into the gateway to healthcare, this could be disadvantageous for patients who don't trust this technique or can't operate it. Beyond access, we must also recognize the fundamental biases in AI systems.79 For instance, if an AI algorithm is trained or tested on a dataset in which only certain groups in the population are represented, the algorithm may perform poorly on other groups. Health inequalities can be exacerbated if AI system performance metrics are too strictly defined to ignore its impacts on ethnic minorities. As such risks seem difficult to mitigate, potential ethical frameworks for guiding consent and incorporation into DL datasets must be carefully considered. Models for informed consent for AI research in medicine have been proposed,80 and when balanced with the actual problem of managing large AI data, dropout models seem to be preferred.
4. Future directions
Recent achievements using ophthalmology-based AI will facilitate more research and clinical translation efforts. To increase the chance of success for DL systems in the future, several elements should be simultaneously considered. In terms of development, the systems should be guided to consider unmet clinical demands. Issues like ill-considered indications, doubtful methods, and unsuitable AI platforms should be addressed before full development. Additional focus directions include multiclass and multimodal AI networks for disease diagnosis, progression prediction, or treatment decision-making. The sensitivity and specificity of DL are both very important in ophthalmology and medical diagnosis. In disease screening and early diagnosis, the sensitivity of AI systems would be deemed more important sometimes. There is a definite unmet clinical need to improve diagnostic capabilities of these procedures, including improving test sensitivity and specificity. These may involve evaluating the optic nerve head using optic discs or fundus photographs to screening a variety of optic neuropathy in mydriatic or non-mydriatic conditions,81 prediction of myopia progression in kids, and the use of combined automatical detection and management of glaucoma VF. Advanced segmentation networks would also improve the localization of anatomical structures and identification of pathological features, enhancing the robustness of the AI systems.
While many diagnostic AI algorithms proposed before shows promising performances, it's worth noting that most of them were trained on small samples, and few algorithms have been tested in real clinical settings. Patient heterogeneity can lead to less accurate AI algorithms when tested in real-world settings. Additionally, considering the effects of reporting bias, algorithms with unsatisfactory performance are more likely to be rejected to be reported. These issues should be adequately handled before AI technologies can be translated into real-world applications. As eye diseases become more prevalent, such as cataracts and glaucoma, there is a wealth of patient data available for training, testing, and validating AI models, which were ideally derived from multi-ethnic patient cohorts to enhance algorithms. In addition to diagnostic accuracy, the effectiveness of AI tools should also be assessed whether it can improve patient outcomes by prospective clinical trials. The challenge is to democratize this technology for stakeholders when clinically proven and implement robust AI algorithms. As AI technologies may pave the way for a “smart” healthcare model before long, it has been recently highlighted that agreed-on international reporting guidelines for AI-related interventions are needed.63 Thanks to these guidelines, Reporting will become more consistent and transparent, while regulators and relevant stakeholders will also be able to assess the cost and effectiveness of the AI-related interventions more confidently. Other potential challenges, such as the “AI black box”, the explainability of AI by medical professionals, and forensic and regulatory challenges, should be carefully addressed prior to applying AI techniques to clinical settings.
5. Conclusions
Although medical AI research has attained many milestones and achieved breakthroughs in the field of ophthalmology, it also faces many difficulties and challenges. The emergence of the period of big data, the development of medical electronics, and people's demand for high-quality medical services are all driving AI systems to maximize their potential in improving clinical medical workflows, patient care and assessing patient prognosis. Ophthalmic medical workers and AI researchers should work closely with computer scientists to apply cutting-edge AI concepts and technologies to solve ophthalmic clinical problems, and should also pay more attention to the translation of the research results. Developments in digital ophthalmology could solve issues such as unbalanced medical resource allocation in countries like China, heavy workload of professional ophthalmologists and challenges in popularizing high-quality medical services. The future of AI interventions in ophthalmology seems promising, but there is still a long way to go prior to their wide clinical application.
Method of literature search
The literature search for this review was conducted in PubMed and Google Scholar from January 2016 to June 2022. The search terms combined a set of keywords from the medical field (ophthalmology, retina, diabetic retinopathy, glaucoma, age-related macular degeneration, cataract, anterior segment disease) and the machine learning field (artificial intelligence, deep learning, machine learning, and convolutional neural network). Term from each set was independently combined with term from the other set. A total of 800 studies assessed as eligible by inspecting the titles and abstracts were reviewed. The main inclusion criterion is that the researches must focus on artificial intelligence in ophthalmology with perceived quality. Several selected typical articles before 2016 and some researches with closely related topics were included as well. Most of the papers were in English. For non-English articles, only their abstract was considered.
Study approval
This study was approved by the Medical Ethics Committee of the Second Affiliated Hospital, Zhejiang University, and complied with the Declaration of Helsinki.
Author contributions
All the authors were involved in conceptualizing, researching and writing of the manuscript, and edited all versions.
Acknowledgements
Thanks to all the peer reviewers and editors for their opinions and suggestions.
Funding
This work was financially supported by the National Key Research and Development Program of China (grant number 2019YFC0118401), Zhejiang Provincial Key Research and Development Plan (grant number 2019C03020), Natural Science Foundation of Zhejiang Province (grant number LQ21H120002), National Natural Science Foundation Regional Innovation and Development Joint Fund (U20A20386).
Declaration of completing interest
The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.
Contributor Information
Kai Jin, Email: jinkai@zju.edu.cn.
Juan Ye, Email: yejuan@zju.edu.cn.
Abbreviations
- AI
artificial intelligence
- DL
deep learning
- DR
diabetic retinopathy
- AMD
age-related macular degeneration
- AUC
area under the curve
- CFP
color fundus images
- OCT
optical coherence tomography
- FFA
fundus fluorescein angiography
- VF
visual field
- CNV
choroidal neovascularization
- SRF
subretinal fluid
- IRF
intraretinal fluid
- GA
geographic atrophy
- OCTA
optical coherence tomography angiography
References
- 1.Vasey B., Nagendran M., Campbell B., et al. Reporting guideline for the early-stage clinical evaluation of decision support systems driven by artificial intelligence: DECIDE-AI. Nat Med. May 2022;28(5):924–933. doi: 10.1038/s41591-022-01772-9. [DOI] [PubMed] [Google Scholar]
- 2.Ting D.S.W., Pasquale L.R., Peng L., et al. Artificial intelligence and deep learning in ophthalmology. Br J Ophthalmol. Feb 2019;103(2):167–175. doi: 10.1136/bjophthalmol-2018-313173. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 3.Gulshan V., Peng L., Coram M., et al. Development and validation of a deep learning algorithm for detection of diabetic retinopathy in retinal fundus photographs. JAMA. Dec 13 2016;316(22):2402–2410. doi: 10.1001/jama.2016.17216. [DOI] [PubMed] [Google Scholar]
- 4.Li Z., He Y., Keel S., et al. Efficacy of a deep learning system for detecting glaucomatous optic neuropathy based on color fundus photographs. Ophthalmology. Aug 2018;125(8):1199–1206. doi: 10.1016/j.ophtha.2018.01.023. [DOI] [PubMed] [Google Scholar]
- 5.Grassmann F., Mengelkamp J., Brandl C., et al. A deep learning algorithm for prediction of age-related eye disease study severity scale for age-related macular degeneration from color fundus photography. Ophthalmology. Sep 2018;125(9):1410–1420. doi: 10.1016/j.ophtha.2018.02.037. [DOI] [PubMed] [Google Scholar]
- 6.Li W., Yang Y., Zhang K. Dense anatomical annotation of slit-lamp images improves the performance of deep learning for the diagnosis of ophthalmic disorders. Nat Biomed Eng. Aug 2020;4(8):767–777. doi: 10.1038/s41551-020-0577-y. [DOI] [PubMed] [Google Scholar]
- 7.Bogunovic H., Montuoro A., Baratsits M., et al. Machine learning of the progression of intermediate age-related macular degeneration based on OCT imaging. Invest Ophthalmol Vis Sci. May 1 2017;58(6):Bio141–bio150. doi: 10.1167/iovs.17-21789. [DOI] [PubMed] [Google Scholar]
- 8.Yim J., Chopra R. Predicting conversion to wet age-related macular degeneration using deep learning. Nat Med. Jun. 2020;26(6):892–899. doi: 10.1038/s41591-020-0867-7. [DOI] [PubMed] [Google Scholar]
- 9.Cruz Rivera S., Liu X. Guidelines for clinical trial protocols for interventions involving artificial intelligence: the SPIRIT-AI extension. BMJ. Sep 2020;26(9):1351–1363. doi: 10.1136/bmj.m3210. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 10.Collins G.S., Moons K.G.M. Reporting of artificial intelligence prediction models. Lancet (London, England) Apr 20 2019;393(10181):1577–1579. doi: 10.1016/S0140-6736(19)30037-6. [DOI] [PubMed] [Google Scholar]
- 11.Sounderajah V., Ashrafian H. Developing specific reporting guidelines for diagnostic accuracy studies assessing AI interventions: the STARD-AI Steering Group. Nat Med. Jun 2020;26(6):807–808. doi: 10.1038/s41591-020-0941-1. [DOI] [PubMed] [Google Scholar]
- 12.González-Gonzalo C., Thee E.F., Klaver C.C.W., et al. Trustworthy AI: closing the gap between development and integration of AI systems in ophthalmic practice. Prog Retin Eye Res. Dec 10 2021 doi: 10.1016/j.preteyeres.2021.101034. [DOI] [PubMed] [Google Scholar]
- 13.Xie Y., Nguyen Q.D., Hamzah H., et al. Artificial intelligence for teleophthalmology-based diabetic retinopathy screening in a national programme: an economic analysis modelling study. Lancet Digit Health. May 2020;2(5):e240–e249. doi: 10.1016/S2589-7500(20)30060-1. [DOI] [PubMed] [Google Scholar]
- 14.Abràmoff M.D., Lou Y., Erginay A., et al. Improved automated detection of diabetic retinopathy on a publicly available dataset through integration of deep learning. Invest Ophthalmol Vis Sci. Oct 1 2016;57(13):5200–5206. doi: 10.1167/iovs.16-19964. [DOI] [PubMed] [Google Scholar]
- 15.Gargeya R., Leng T. Automated identification of diabetic retinopathy using deep learning. Ophthalmology. Jul 2017;124(7):962–969. doi: 10.1016/j.ophtha.2017.02.008. [DOI] [PubMed] [Google Scholar]
- 16.Ting D.S.W., Cheung C.Y., Lim G., et al. Development and validation of a deep learning system for diabetic retinopathy and related eye diseases using retinal images from multi-ethnic populations with diabetes. JAMA. Dec 12 2017;318(22):2211–2223. doi: 10.1001/jama.2017.18152. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 17.Pan X., Jin K., Cao J., et al. Multi-label classification of retinal lesions in diabetic retinopathy for automatic analysis of fundus fluorescein angiography based on deep learning. Graefes Arch Clin Exp Ophthalmol. Apr 2020;258(4):779–785. doi: 10.1007/s00417-019-04575-w. [DOI] [PubMed] [Google Scholar]
- 18.Gao Z., Jin K., Yan Y., et al. End-to-end diabetic retinopathy grading based on fundus fluorescein angiography images using deep learning. Graefes Arch Clin Exp Ophthalmol. May 2022;260(5):1663–1673. doi: 10.1007/s00417-021-05503-7. [DOI] [PubMed] [Google Scholar]
- 19.Ting D.S.W., Cheung C.Y., Nguyen Q., et al. Deep learning in estimating prevalence and systemic risk factors for diabetic retinopathy: a multi-ethnic study. NPJ Digit Med. 2019;2:24. doi: 10.1038/s41746-019-0097-x. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 20.Varadarajan A.V., Bavishi P., Ruamviboonsuk P., et al. Predicting optical coherence tomography-derived diabetic macular edema grades from fundus photographs using deep learning. Nat Commun. Jan 8 2020;11(1):130. doi: 10.1038/s41467-019-13922-8. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 21.Abràmoff M.D., Lavin P.T., Birch M., et al. Pivotal trial of an autonomous AI-based diagnostic system for detection of diabetic retinopathy in primary care offices. NPJ Digit Med. 2018;1:39. doi: 10.1038/s41746-018-0040-6. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 22.Bellemo V., Lim Z.W., Lim G., et al. Artificial intelligence using deep learning to screen for referable and vision-threatening diabetic retinopathy in Africa: a clinical validation study. The Lancet. Digital health. May 2019;1(1):e35–e44. doi: 10.1016/S2589-7500(19)30004-4. [DOI] [PubMed] [Google Scholar]
- 23.Liu H., Li L., Wormstone I.M., et al. Development and validation of a deep learning system to detect glaucomatous optic neuropathy using fundus photographs. JAMA Ophthalmol. Sep 12 2019 doi: 10.1001/jamaophthalmol.2019.3501. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 24.Chang J., Lee J., Ha A., et al. Explaining the rationale of deep learning glaucoma decisions with adversarial examples. Ophthalmology. Jan 2021;128(1):78–88. doi: 10.1016/j.ophtha.2020.06.036. [DOI] [PubMed] [Google Scholar]
- 25.Medeiros F.A., Jammal A.A., Thompson A.C. From machine to machine: an OCT-trained deep learning algorithm for objective quantification of glaucomatous damage in fundus photographs. Ophthalmology. Apr 2019;126(4):513–521. doi: 10.1016/j.ophtha.2018.12.033. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 26.Medeiros F.A., Jammal A.A., Mariottoni E.B. Detection of progressive glaucomatous optic nerve damage on fundus photographs with deep learning. Ophthalmology. Jul 28 2020 doi: 10.1016/j.ophtha.2020.07.045. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 27.Xu Y., Hu M., Liu H., et al. A hierarchical deep learning approach with transparency and interpretability based on small samples for glaucoma diagnosis. NPJ Digit Med. Mar 11 2021;4(1):48. doi: 10.1038/s41746-021-00417-4. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 28.Sun S., Ha A., Kim Y.K., et al. Dual-input convolutional neural network for glaucoma diagnosis using spectral-domain optical coherence tomography. Br J Ophthalmol. Sep 12 2020 doi: 10.1136/bjophthalmol-2020-316274. [DOI] [PubMed] [Google Scholar]
- 29.Thompson A.C., Jammal A.A., Berchuck S.I., et al. Assessment of a segmentation-free deep learning algorithm for diagnosing glaucoma from optical coherence tomography scans. JAMA Ophthalmol. Apr 1 2020;138(4):333–339. doi: 10.1001/jamaophthalmol.2019.5983. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 30.Fu H., Baskaran M., Xu Y., et al. A deep learning system for automated angle-closure detection in anterior segment optical coherence tomography images. Am J Ophthalmol. Jul 2019;203:37–45. doi: 10.1016/j.ajo.2019.02.028. [DOI] [PubMed] [Google Scholar]
- 31.Zangwill L.M., Chan K., Bowd C., et al. Heidelberg retina tomograph measurements of the optic disc and parapapillary retina for detecting glaucoma analyzed by machine learning classifiers. Invest Ophthalmol Vis Sci. Sep 2004;45(9):3144–3151. doi: 10.1167/iovs.04-0202. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 32.Burgansky-Eliash Z., Wollstein G., Chu T., et al. Optical coherence tomography machine learning classifiers for glaucoma detection: a preliminary study. Invest Ophthalmol Vis Sci. Nov 2005;46(11):4147–4152. doi: 10.1167/iovs.05-0366. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 33.Phene S., Dunn R.C., Hammel N., et al. Deep learning and glaucoma specialists: the relative importance of optic disc features to predict glaucoma referral in fundus photographs. Ophthalmology. Dec 2019;126(12):1627–1639. doi: 10.1016/j.ophtha.2019.07.024. [DOI] [PubMed] [Google Scholar]
- 34.Asaoka R., Murata H., Hirasawa K., et al. Using deep learning and transfer learning to accurately diagnose early-onset glaucoma from macular optical coherence tomography images. Am J Ophthalmol. Feb 2019;198:136–145. doi: 10.1016/j.ajo.2018.10.007. [DOI] [PubMed] [Google Scholar]
- 35.Xu B.Y., Chiang M., Chaudhary S., et al. Deep learning classifiers for automated detection of gonioscopic angle closure based on anterior segment OCT images. Am J Ophthalmol. Dec 2019;208:273–280. doi: 10.1016/j.ajo.2019.08.004. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 36.Elze T., Pasquale L.R., Shen L.Q., et al. Patterns of functional vision loss in glaucoma determined with archetypal analysis. J R Soc Interface. Feb 6 2015;12(103) doi: 10.1098/rsif.2014.1118. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 37.Wen J.C., Lee C.S., Keane P.A., et al. Forecasting future Humphrey Visual Fields using deep learning. PLoS One. 2019;14(4) doi: 10.1371/journal.pone.0214875. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 38.Huang X., Jin K., Zhu J., et al. A structure-related fine-grained deep learning system with diversity data for universal glaucoma visual field grading. Front Med. 2022;9 doi: 10.3389/fmed.2022.832920. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 39.Burlina P.M., Joshi N., Pekala M., et al. Automated grading of age-related macular degeneration from color fundus images using deep convolutional neural networks. JAMA Ophthalmol. Nov 1 2017;135(11):1170–1176. doi: 10.1001/jamaophthalmol.2017.3782. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 40.Peng Y., Dharssi S., Chen Q., et al. DeepSeeNet: a deep learning model for automated classification of patient-based age-related macular degeneration severity from color fundus photographs. Ophthalmology. Apr 2019;126(4):565–575. doi: 10.1016/j.ophtha.2018.11.015. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 41.Burlina P.M., Joshi N., Pacheco K.D., et al. Use of deep learning for detailed severity characterization and estimation of 5-year risk among patients with age-related macular degeneration. JAMA Ophthalmol. Dec 1 2018;136(12):1359–1366. doi: 10.1001/jamaophthalmol.2018.4118. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 42.Yan Q., Weeks D.E., Xin H., et al. Deep-learning-based prediction of late age-related macular degeneration progression. Nat Mach Intell. Feb 2020;2(2):141–150. doi: 10.1038/s42256-020-0154-9. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 43.Kermany D.S., Goldbaum M., Cai W., et al. Identifying medical diagnoses and treatable diseases by image-based deep learning. Cell. Feb 22 2018;172(5):1122–1131. doi: 10.1016/j.cell.2018.02.010. e1129. [DOI] [PubMed] [Google Scholar]
- 44.Schlegl T., Waldstein S.M., Bogunovic H., et al. Fully automated detection and quantification of macular fluid in OCT using deep learning. Ophthalmology. Apr 2018;125(4):549–558. doi: 10.1016/j.ophtha.2017.10.031. [DOI] [PubMed] [Google Scholar]
- 45.Schmidt-Erfurth U., Vogl W.D., Jampol L.M., et al. Application of automated quantification of fluid volumes to anti-VEGF therapy of neovascular age-related macular degeneration. Ophthalmology. Sep 2020;127(9):1211–1219. doi: 10.1016/j.ophtha.2020.03.010. [DOI] [PubMed] [Google Scholar]
- 46.Moraes G., Fu D.J., Wilson M., et al. Quantitative analysis of OCT for neovascular age-related macular degeneration using deep learning. Ophthalmology. May 2021;128(5):693–705. doi: 10.1016/j.ophtha.2020.09.025. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 47.Yan Y., Jin K., Gao Z., et al. Attention-based deep learning system for automated diagnoses of age-related macular degeneration in optical coherence tomography images. Med Phys. Sep 2021;48(9):4926–4934. doi: 10.1002/mp.15002. [DOI] [PubMed] [Google Scholar]
- 48.Zhang G., Fu D.J., Liefers B., et al. Clinically relevant deep learning for detection and quantification of geographic atrophy from optical coherence tomography: a model development and external validation study. Lancet Digit Health. Oct 2021;3(10):e665–e675. doi: 10.1016/S2589-7500(21)00134-5. [DOI] [PubMed] [Google Scholar]
- 49.Liefers B., Taylor P., Alsaedi A., et al. Quantification of key retinal features in early and late age-related macular degeneration using deep learning. Am J Ophthalmol. Jun 2021;226:1–12. doi: 10.1016/j.ajo.2020.12.034. [DOI] [PubMed] [Google Scholar]
- 50.Xu Z., Wang W., Yang J., et al. Automated diagnoses of age-related macular degeneration and polypoidal choroidal vasculopathy using bi-modal deep convolutional neural networks. Br J Ophthalmol. Apr 2021;105(4):561–566. doi: 10.1136/bjophthalmol-2020-315817. [DOI] [PubMed] [Google Scholar]
- 51.Jin K., Yan Y., Chen M., et al. Multimodal deep learning with feature level fusion for identification of choroidal neovascularization activity in age-related macular degeneration. Acta Ophthalmol. Mar 2022;100(2):e512–e520. doi: 10.1111/aos.14928. [DOI] [PubMed] [Google Scholar]
- 52.Cheung C.Y., Li H., Lamoureux E.L., et al. Validity of a new computer-aided diagnosis imaging program to quantify nuclear cataract from slit-lamp photographs. Invest Ophthalmol Vis Sci. Mar 10 2011;52(3):1314–1319. doi: 10.1167/iovs.10-5427. [DOI] [PubMed] [Google Scholar]
- 53.Long E., Chen J., Wu X., et al. Artificial intelligence manages congenital cataract with individualized prediction and telehealth computing. NPJ Digit Med. 2020;3:112. doi: 10.1038/s41746-020-00319-x. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 54.Li Z., Jiang J., Chen K., et al. Preventing corneal blindness caused by keratitis using artificial intelligence. Nat Commun. Jun 18 2021;12(1):3738. doi: 10.1038/s41467-021-24116-6. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 55.Ye X., Wang J., Chen Y., et al. Automatic screening and identifying myopic maculopathy on optical coherence tomography images using deep learning. Transl Vis Sci Technol. Nov 1 2021;10(13):10. doi: 10.1167/tvst.10.13.10. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 56.Yoo T.K., Ryu I.H., Lee G., et al. Adopting machine learning to automatically identify candidate patients for corneal refractive surgery. NPJ Digit Med. 2019;2:59. doi: 10.1038/s41746-019-0135-8. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 57.Wang L., Ding L., Liu Z., et al. Automated identification of malignancy in whole-slide pathological images: identification of eyelid malignant melanoma in gigapixel pathological slides using deep learning. Br J Ophthalmol. Mar 2020;104(3):318–323. doi: 10.1136/bjophthalmol-2018-313706. [DOI] [PubMed] [Google Scholar]
- 58.Burlina P., Paul W., Mathew P., et al. Low-shot deep learning of diabetic retinopathy with potential applications to address artificial intelligence bias in retinal diagnostics and rare ophthalmic diseases. JAMA Ophthalmol. Oct 1 2020;138(10):1070–1077. doi: 10.1001/jamaophthalmol.2020.3269. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 59.Burlina P., Paul W., Liu T.Y.A., et al. Detecting anomalies in retinal diseases using generative, discriminative, and self-supervised deep learning. JAMA Ophthalmology. Feb 1 2022;140(2):185–189. doi: 10.1001/jamaophthalmol.2021.5557. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 60.Diaz-Pinto A., Colomer A., Naranjo V., et al. Retinal image synthesis and semi-supervised learning for glaucoma assessment. IEEE Trans Med Imag. 2019;38(9):2211–2218. doi: 10.1109/TMI.2019.2903434. [DOI] [PubMed] [Google Scholar]
- 61.Liu S., Hong J., Lu X., et al. Joint optic disc and cup segmentation using semi-supervised conditional GANs. Comput Biol Med. 2019 Dec;115 doi: 10.1016/j.compbiomed.2019.103485. [DOI] [PubMed] [Google Scholar]
- 62.Zhao R., Chen X., Liu X., et al. Direct cup-to-disc ratio estimation for glaucoma screening via semi-supervised learning. IEEE J Biomed Health Inform. 2020 Apr;24(4):1104–1113. doi: 10.1109/JBHI.2019.2934477. [DOI] [PubMed] [Google Scholar]
- 63.Yip M.Y.T., Lim G. Technical and imaging factors influencing performance of deep learning systems for diabetic retinopathy. NPJ Digit Med. 2020;3:40. doi: 10.1038/s41746-020-0247-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 64.Jin K., Zhou M., Wang S., et al. Computer-aided diagnosis based on enhancement of degraded fundus photographs. Acta Ophthalmol. May 2018;96(3):e320–e326. doi: 10.1111/aos.13573. [DOI] [PubMed] [Google Scholar]
- 65.CONSORT-AI and SPIRIT-AI Steering Group Reporting guidelines for clinical trials evaluating artificial intelligence interventions are needed. Nat Med. Oct 2019;25(10):1467–1468. doi: 10.1038/s41591-019-0603-3. [DOI] [PubMed] [Google Scholar]
- 66.Liu X., Cruz Rivera S., Moher D., et al. Reporting guidelines for clinical trial reports for interventions involving artificial intelligence: the CONSORT-AI extension. The Lancet. Digital health. Oct 2020;2(10):e537–e548. doi: 10.1038/s41591-020-1034-x. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 67.Rivera S.C., Liu X., Chan A.W., et al. Guidelines for clinical trial protocols for interventions involving artificial intelligence: the SPIRIT-AI Extension. BMJ (Clinical research ed.) Sep 9 2020;370:m3210. doi: 10.1136/bmj.m3210. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 68.Collins G.S., Reitsma J.B., Altman D.G., et al. Transparent reporting of a multivariable prediction model for individual prognosis or diagnosis (TRIPOD): the TRIPOD statement. BMJ. Jan 2015:350. doi: 10.1136/bmj.g7594. [DOI] [PubMed] [Google Scholar]
- 69.Schulz K.F., Altman D.G., Moher D. CONSORT 2010 statement: updated guidelines for reporting parallel group randomised trials. PLoS Med. Mar 24 2010;7(3) doi: 10.1136/bmj.c332. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 70.Cohen J.F., Korevaar D.A., Altman D.G., et al. STARD 2015 guidelines for reporting diagnostic accuracy studies: explanation and elaboration. BMJ Open. Nov 14 2016;6(11) doi: 10.1136/bmjopen-2016-012799. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 71.Chan A.W., Tetzlaff J.M., Gøtzsche P.C., et al. SPIRIT 2013 explanation and elaboration: guidance for protocols of clinical trials. BMJ. Jan 8 2013;346 doi: 10.1136/bmj.e7586. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 72.Krleza-Jerić K., Chan A.W., Dickersin K., et al. Principles for international registration of protocol information and results from human trials of health related interventions: Ottawa statement (part 1) BMJ. Apr 23 2005;330(7497):956–958. doi: 10.1136/bmj.330.7497.956. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 73.Ibrahim H., Liu X., Rivera S.C., et al. Reporting guidelines for clinical trials of artificial intelligence interventions: the SPIRIT-AI and CONSORT-AI guidelines. Trials. Jan 6 2021;22(1):11. doi: 10.1186/s13063-020-04951-6. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 74.Tan T.E., Anees A., Chen C., et al. Retinal photograph-based deep learning algorithms for myopia and a blockchain platform to facilitate artificial intelligence medical research: a retrospective multicohort study. Lancet Digit Health. May 2021;3(5):e317–e329. doi: 10.1016/S2589-7500(21)00055-8. [DOI] [PubMed] [Google Scholar]
- 75.He J., Baxter S.L. The practical implementation of artificial intelligence technologies in medicine. Nat Med. Jan 2019;25(1):30–36. doi: 10.1038/s41591-018-0307-0. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 76.Dow E.R., Keenan T.D.L., Lad E.M., et al. From data to deployment: the collaborative community on ophthalmic imaging roadmap for artificial intelligence in age-related macular degeneration. Ophthalmology. May 2022;129(5):e43–e59. doi: 10.1016/j.ophtha.2022.01.002. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 77.Li J.O., Liu H., Ting D.S.J., et al. Digital technology, tele-medicine and artificial intelligence in ophthalmology: a global perspective. Prog Retin Eye Res. May 2021;82 doi: 10.1016/j.preteyeres.2020.100900. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 78.Ye T., Xue J. Psychosocial factors affecting artificial intelligence adoption in health care in China: cross-sectional study. J Med Internet Res. Oct 17 2019;21(10) doi: 10.2196/14316. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 79.Gianfrancesco M.A., Tamang S., Yazdany J., et al. Potential biases in machine learning algorithms using electronic health record data. JAMA Intern Med. Nov 1 2018;178(11):1544–1547. doi: 10.1001/jamainternmed.2018.3763. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 80.Tom E., Keane P.A., Blazes M., et al. Protecting data privacy in the age of AI-enabled ophthalmology. Transl Vis Sci Technol. Jul 2020;9(2):36. doi: 10.1167/tvst.9.2.36. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 81.Milea D., Najjar R.P. Artificial intelligence to detect Papilledema from ocular fundus photographs. N Engl J Med. Apr 30 2020;382(18):1687–1695. doi: 10.1056/NEJMoa1917130. [DOI] [PubMed] [Google Scholar]