Abstract
Objective:
In the pivotal clinical trial that led to Food and Drug Administration De Novo “approval” of the first fully autonomous artificial intelligence (AI) diabetic retinal disease diagnostic system, a reflexive dilation protocol was used. Using real-world deployment data before implementation of reflexive dilation, we identified factors associated with nondiagnostic results. These factors allow a novel predictive dilation workflow, where patients most likely to benefit from pharmacologic dilation are dilated a priori to maximize efficiency and patient satisfaction.
Methods:
Retrospective review of patients who were assessed with autonomous AI at Johns Hopkins Medicine (8/2020 to 5/2021). We constructed a multivariable logistic regression model for nondiagnostic results to compare characteristics of patients with and without diagnostic results, using adjusted odds ratio (aOR). P < .05 was considered statistically significant.
Results:
Of 241 patients (59% female; median age = 59), 123 (51%) had nondiagnostic results. In multivariable analysis, type 1 diabetes (T1D, aOR = 5.82, 95% confidence interval [CI]: 1.45-23.40, P = .01), smoking (aOR = 2.86, 95% CI: 1.36-5.99, P = .005), and age (every 10-year increase, aOR = 2.12, 95% CI: 1.62-2.77, P < .001) were associated with nondiagnostic results. Following feature elimination, a predictive model was created using T1D, smoking, age, race, sex, and hypertension as inputs. The model showed an area under the receiver-operator characteristics curve of 0.76 in five-fold cross-validation.
Conclusions:
We used factors associated with nondiagnostic results to design a novel, predictive dilation workflow, where patients most likely to benefit from pharmacologic dilation are dilated a priori. This new workflow has the potential to be more efficient than reflexive dilation, thus maximizing the number of at-risk patients receiving their diabetic retinal examinations.
Keywords: implementation science, diabetic retinal disease, nondiagnostic examination, pupillary dilation, artificial intelligence, machine learning, deep learning, clinical decision support
Introduction
Diabetic retinal disease (DRD) is a highly prevalent complication in patients with diabetes mellitus (DM) and is a leading cause of blindness and visual loss in working-age adults. 1 More than 100 million adults worldwide are currently diagnosed with DRD, and this patient population is projected to double over the next 20 years. 2 While early diagnosis and intervention can significantly improve patient outcomes and avoid progression to irreversible vision loss, 3 the current management of DRD is limited by the inadequate adherence to regular eye screening in people with diabetes on a population level. 4 DRD and its complications are particularly pronounced in underserved/under-represented populations. 5 Efforts to improve access include development of an autonomous AI system for the diagnosis of DRD (IDx-DR, Digital Diagnostics, Coralville, IA). 6 This breakthrough device detects Early Treatment Diabetic Retinopathy Study (ETDRS) level 35 and/or center involved and/or clinically significant macular edema (referred to as “more than mild diabetic retinopathy” [mtmDR]), and was the first fully autonomous diagnostic AI system approved by the Food and Drug Administration (FDA) in any medical fieldIt is meant to be deployed in a primary care setting to evaluate patients with diabetes with no known DRD.
To operate the autonomous AI system, minimally skilled operators take images of patients’ retinae using a nonmydriatic retinal camera (NW400, Topcon Medical Systems, Oakland, NJ), resulting in four images per patient (one centered on the optic disk and one centered on the fovea, per eye). While these images are being taken, the operator is guided by the AI system, and once the quality is sufficient, then a patient-level output is generated (mtmDR present, mtmDR absent, or nondiagnostic result). The AI system will attempt to determine whether mtmDR is present or not, only if the images are of sufficient quality to be deemed diagnostic in the first place. Clinic efficiency is maximized when the number of nondiagnostic results is low.
In the pivotal clinical trial that led to this autonomous AI system's FDA De Novo approval, the protocol included reflexive dilation: if the AI system reported no diagnostic result after three attempts, the patient was (reflexively) pharmacologically dilated. While 76% of subjects received a diagnostic result without pharmacologic dilation, another 20% did require dilation, leading to a reported 96% overall diagnostic result, which is part of the FDA labeling. Similarly, in a real-world deployment study of IDx-DR in Poland, Grzybowski et al found that the diagnostic rate was 78% without pharmacologic dilation. 7
Although the pivotal clinical trial and other studies showedthat diagnostic rates can be increased by a reflexive dilation protocol,6,8,9 such a protocol results in workflow inefficiencies and risks the potential of patients dropping out entirely, as multiple iterations of imaging are required before pharmacologic dilation can be administered. Therefore, during the initial deployment (August 2020 and May 2021) at Johns Hopkins Medicine, reflexive dilation was not implemented. This allowed us to collect data and thereby identify pre-existing factors associated with nondiagnostic results, to characterize those patients most likely to have nondiagnostic results without dilation before the first images are taken. The novel predictive dilation protocol, as compared with a reflexive dilation protocol, has the potential to increase workflow efficiency, reduce delays for operators and patients, and increase patient and clinical staff satisfaction, thereby maximizing the potential of AI-based DRD screening in the primary care setting.
Materials and Methods
Study Design
This study was approved by the Johns Hopkins University Institutional Review Board (IRB00269812) with a waiver of informed consent. This was a retrospective case–control study.
Population
During the initial deployment phase between August 2020 and May 2021, autonomous AI systems were implemented at four Johns Hopkins Medicine primary care clinics in Maryland, with the standard protocol conforming to FDA labeling, except that no reflexive (pharmacologic) dilation was used. Consecutive adult patients with diabetes but no known DRD and thus eligible for screening, who presented for a primary care visit, were assessed using the autonomous AI system. Patients with nondiagnostic images had their images re-taken for a maximum of three total attempts. If the autonomous system still determined the scans to be nondiagnostic after these three attempts, then this patient was determined to be “nondiagnostic.” No pharmacological dilation was performed on any of these patients. A patient’s images can be deemed “nondiagnostic” for two reasons: poor image quality or inadequate field of view. We analyzed the nondiagnostic subjects and excluded those subjects that were nondiagnostic due to field-of-view deficiencies (10 subjects)—these deficiencies were likely the result of operators’ errors and would not benefit from pharmacologic dilation.
Only adult patients evaluated during this initial pilot deployment phase between August 2020 and May 2021 were included in our study; that is, patients evaluated by the autonomous AI system after May 2021, after which reflexive dilation was implemented, were excluded. We extracted demographic and medical variables from the electronic health record, including age, sex, race (African American, Asian, Native Hawaiian or Pacific Islander, White, and other), ethnicity (Hispanic or non-Hispanic), type of diabetes, body mass index (BMI), HbA1c, history of hypertension, smoking status, and time between initial system deployment and examination date (as a proxy for a learning effect).
Outcome
The primary outcome was the adjusted odds ratio (aOR) for each factor associated with a nondiagnostic examination result.
Statistical Analysis
Continuous variables are presented as median (interquartile range/IQR), and categorical variables are presented as n (%). Mann-Whitney U and Pearson’s chi-square tests were used to compare differences between diagnostic and nondiagnostic groups, as appropriate.
Multivariable logistic regression analysis was used to identify a priori risk factors for nondiagnostic results. Ten such factors were considered: sex, ethnicity, race, smoking status, hypertension, type 1 DM, type 2 DM, BMI, age, and HbA1c. We included these factors for analysis for two reasons. First, these factors are readily retrievable at each primary care visit where autonomous AI screening is performed, from either the electronic health record (EHR) or, alternatively, the patient. Second, these factors have been associated with either poor retinal image quality or poor pupillary dilation in the published literature, and hence are likely to be associated with nondiagnostic results as well. Variables such as age, BMI, and HbA1c were treated as continuous, since they produced better-fitting models compared with categorical terms. A final logistic regression model was selected using backward stepwise selection based on the lowest Akaike’s information criterion (AIC). This methodology starts with building a model using all variables under consideration (the 10 listed previously). Variables are subsequently removed from the model, one at a time starting from the least significant variable, until a reduced model with an optimal AIC is found. Risk factors are reported as aORs with 95% CIs. Statistical significance was set at P< .05.
Predictive Dilation Model
Using the results of this analysis, a predictive dilation model was built using stepwise backward selection. Five-fold cross-validation was used to evaluate model performance in identifying patients who would have nondiagnostic results without dilation. The population was evaluated five times, each containing a random and mutually exclusive data partition (80% as the training fold and 20% as the validation fold). Area under the receiver-operator characteristics curve (AUC) was calculated as the primary outcome. The final AUC measurement is the result of averaging the AUC of each evaluation to yield a more realistic real-world estimate of model performance. We also report sensitivity/specificity pairs and corresponding 95% CIs at different test thresholds.
Results
Descriptive Statistics
Out of 241 patients (41% male, median age = 59 years) who underwent autonomous AI screening, 123 (51%) were nondiagnostic. Those with nondiagnostic imaging had a significantly higher median age (64 vs 53 years, P < .001) and lower median BMI (32.0 vs 35.5 kg/m2, P = .005), compared with those with a diagnostic result (Table 1). In addition, patients with nondiagnostic results were more likely to be current smokers (30% vs 14%, P = .002) and less likely to have type 2 diabetes (T2D) (104/123; 85% vs 110/118; 93%, P = .03). Sex, race, ethnicity, history of hypertension, HbA1c, and time between examination date and initial system deployment were not different between the two groups.
Table 1.
Demographic and Clinical Variables of Patients With and Without Diagnostic Results by the Autonomous AI System.
Variables | Total (n = 241) |
Diagnostic (n = 118, 49%) |
Nondiagnostic (n = 123, 51%) |
P |
---|---|---|---|---|
Male sex | 98 (41%) | 42 (36%) | 56 (46%) | .12 |
Age | 59 (50-67) | 53 (45-63) | 64 (55-69) | <.001 |
Race | .27 | |||
White | 86 (36%) | 48 (41%) | 38 (31%) | |
Black or African American | 132 (55%) | 61 (52%) | 71 (58%) | |
Asian | 11 (5%) | 3 (3%) | 8 (7%) | |
Native Hawaiian or Pacific Islander | 1 (0%) | 0 (0%) | 1 (1%) | |
Other or declined to answer | 11 (5%) | 6 (5%) | 5 (4%) | |
Hispanic Ethnicity | 10 (4%) | 6 (5%) | 4 (3%) | .48 |
BMI (kg/m2) | 33.4 (28.8-39.4) | 35.5 (29.5-41.7) | 32.0 (28.6-37.2) | .005 |
Current smoker | 53 (22%) | 16 (14%) | 37 (30%) | .002 |
Hypertension | 133 (55%) | 70 (59%) | 63 (51%) | .21 |
T1D | 16 (7%) | 5 (4%) | 11 (9%) | .14 |
T2D | 214 (89%) | 110 (93%) | 104 (85%) | .03 |
HbA1c (%) | 6.9 (6.3-7.9) | 7.0 (6.4-8.2) | 6.7 (6.2-7.7) | .08 |
Time between examination date and initial system deployment (days) | 47.6 (19.8-91.8) | 48.8 (21.2-92.6) | .87 |
Continuous variables: median (IQR). Categorical variables: n (%). Bolded P values are statistically significant.
Abbreviation: AI, artificial intelligence; BMI, body mass index; T1D, type 1 diabetes; T2D, type 2 diabetes.
In the multivariable logistic regression model, we identified three factors associated with nondiagnostic results: type 1 DM (aOR = 5.82, 95% CI: 1.45-23.40, P = .01), current smoking (2.86, 95% CI: 1.36-5.99, P = .005), and older age (every 10-year increase, aOR = 2.12, 95% CI: 1.62-2.77, P < .001) (Table 2). Nonwhite race (aOR = 1.85, 95% CI: 1.00-3.43, P = .05) and male sex (aOR = 1.80, 95% CI: 1.00-3.24, P = .05) trended toward statistical significance (Figure 1). As there were four different deployment sites with possible site-dependent variations, we conducted an exploratory clustering and mixed effects analysis with site as a random effect. The mixed effects analysis identified the same statistically significant factors associated with nondiagnostic results.
Table 2.
Multivariable Logistic Regression Model for Prediction of Nondiagnostic Results.
Variable | aOR | 95% CI | P |
---|---|---|---|
T1D | 5.82 | 1.45-23.40 | .01 |
Current smoker | 2.86 | 1.36-5.99 | .005 |
Age (every 10-year increase) | 2.12 | 1.62-2.77 | <.001 |
Nonwhite race | 1.85 | 1.00-3.43 | .05 |
Male sex | 1.80 | 1.00-3.24 | .05 |
Hypertension | 0.53 | 0.29-0.98 | .04 |
Abbreviations: aOR, adjusted odds ratio; CI, confidence interval.
Figure 1.
Factors associated with nondiagnostic examinations. Receiver-operator characteristics (ROC) curves for five-fold internal cross-validation using 50% cutoff value; average area under the receiver-operator characteristics curve (AUC) = 0.76.
Abbreviation: DR, diabetic retinopathy; T1D, type 1 diabetes.
Predictive Dilation Model
Based on the results of the multivariate analysis, we used six priors (Table 2) to build the predictive dilation model. The model’s AUC was 0.76 (Figure 2), and the sensitivities and specificities of this model at different probability cutoffs were reported in Table 3. At a 50% cutoff value, the sensitivity and specificity were 77% and 68%, respectively.
Figure 2.
Five-fold cross-validated ROC curves for the multivariable predictive dilation model. Each dotted line represents a single fold, while the solid red line represents the mean across all folds.
Abbreviations: cvAUC, cross-validated area under the receiver-operator characteristics curve; ROC, receiver-operator characteristics.
Table 3.
Sensitivity, Specificity, and AUC, With Corresponding 95% Confidence Intervals, as Well as Projected # Unnecessary Dilations (i.e., False Positives) and Projected # Missed Dilations (i.e., False Negatives) in a Theoretical Sample Size of 100 Patients.
Prediction cutoff | Sensitivity | Specificity | AUC | # Unnecessary dilations (false positive) |
# Missed dilations (false negative) |
---|---|---|---|---|---|
30% | 0.91 (0.87, 0.92) | 0.38 (0.34, 0.43) | 0.63 (0.58, 0.68) | 30 | 5 |
40% | 0.86 (0.82, 0.88) | 0.53 (0.48, 0.57) | 0.70 (0.64, 0.75) | 23 | 7 |
50% | 0.77 (0.71, 0.79) | 0.68 (0.63, 0.71) | 0.73 (0.67, 0.78) | 16 | 12 |
60% | 0.55 (0.51, 0.60) | 0.81 (0.77, 0.84) | 0.67 (0.61, 0.73) | 9 | 23 |
70% | 0.40 (0.36, 0.44) | 0.92 (0.89, 0.94) | 0.65 (0.60, 0.70) | 4 | 31 |
80% | 0.23 (0.20, 0.27) | 0.95 (0.91, 0.96) | 0.58 (0.54, 0.62) | 2 | 39 |
Our multivariable model for predicting nondiagnostic results at different classification cutoff values. For example, a 50% cutoff value means that patients with a >50% probability of nondiagnosis will be classified as “nondiagnostic.”
Abbreviation: AUC, area under the curve.
Discussion
In our study, we determined that there are significant priors associated with lower diagnosability: type 1 DM, current smoker, and older age. As these priors are straightforward to determine, they are central in our novel predictive dilation workflow, in which patients at a high risk of nondiagnostic results are identified immediately and can then be dilated proactively, prior to the first retinal images being taken. The autonomous AI system (IDx-DR) utilized in our study is the first FDA-approved fully autonomous system in any medical field, 6 and several studies have evaluated the gradeability of consecutively obtained retinal images.7,10-13 However, a detailed analysis of patient factors related to nondiagnostic images was not conducted in any of these studies.
Real-world experience in deploying AI systems for automatic DRD detection has proven to be challenging. Clinical workflows are highly variable across different sites, and AI system implementation is limited by pre-existing logistics. For example, in a white article on AI-based DRD screening deployment in Thailand, Google researchers found issues with gradeability due to poor image quality (21% of patients), as in our study, to be a major limitation in AI implementation. 14 Nondiagnostic retinal images hinder widespread deployment and adoption of this technology, since it is resource intensive and frustrating for patients to retake several sets of images. Also, the resources and time deployed to take nondiagnostic images will be “wasted,” as these patients will ultimately need to be re-examined by humans. While the risk of complications from pharmacologic dilation is extremely low, 15 it is inconvenient for patients, as it increases light sensitivity and may prevent them from being able to drive after the examination.
In our multivariable logistic regression model, type 1 DM and age are among the strongest predictors for nondiagnostic imaging. Interestingly, compared with the pivotal clinical trial of IDx-DR, 6 our median age (59 years in both studies) and percentage of patients with type 1 DM (roughly 7% in both studies) were similar. However, our study population had nearly double the proportion of nonwhite patients (64% vs 38%), and nonwhite race showed a trend toward nondiagnostic images in our multivariable logistic regression model. Perhaps, the high percentage of nonwhite patients in our population was one of the reasons for our higher nondiagnostic rate (without reflexive dilation), as compared with the pivotal clinical trial.
While it is beyond the scope of this study to determine the causal relationship between our identified patient-level risk factors and nondiagnostic results, we investigated whether the associations were plausible from a pathophysiologic point of view. Type 1 DM patients typically have longer duration of DM, as compared with type 2 DM patients, which likely translates into more advanced systemic microvascular and nerve damage. It has been speculated that autonomic neuropathy affecting the pupillary sphincter and dilator muscles lead to small pupils in patient with diabetes under low luminance conditions. 16 Similarly, smoking affects autonomic function. Specifically, smoking reduces sympathetic nerve activity, 17 which drives pupillary dilation in humans. Regarding age, studies have shown that an older age corresponds to a significantly increased rate of ungradable retinal images during DRD screening, 18 and this correlation is likely due to the fact that older patients have smaller pupils that could lead to poor image quality. 19 To this point, our institution’s experience with implementation of IDx-DR in the pediatric population (the current study only included adult patients) demonstrated diagnosability of 97.5% without dilation, and in this population of 310 youth, age, sex, and race/ethnicity were not associated with increased imaging attempts. 20
We have created a “plug-and-play” multivariable logistic regression model to enable predictive dilation. The model requires six inputs: age, DM type, smoking status, race, sex, and hypertension status. The resulting novel predictive dilation workflow had an AUC of 0.76. This is exciting as it is a clinical decision support tool created for real-life clinical practice, to predict which patients will benefit from pharmacological dilation before the first image is taken. Its six priors are readily retrievable at a wide range of primary care settings. In addition, we reported different sensitivities and specificities of the model at different user-dependent cutoff thresholds (Table 3), which will allow individual clinics to set their own minimum thresholds for initiating predictive dilation protocols.
This workflow has the potential to improve clinical workflow efficiency. Specifically, a patient predicted by the model to have a high chance of a nondiagnostic result can be dilated and will be dilated by the time he/she undergoes autonomous AI screening. In contrast, although the traditional reflexive dilation model (per current FDA labeling) is capable of increasing diagnostic rate (from 76% to 96% in the pivotal clinical trial and from 49% to roughly 80% at our institution), efficiency is lost and will inevitably prolong the visit duration and create inconvenience for the patient and clinical team.
The strength of our study lies in the real-world implementation and the diverse patient population, in both community-based and hospital-based primary care clinics within our integrated healthcare system. Our patient population had a higher portion of minority patients, as compared with other studies that investigated the usability and diagnostic rate of the IDx-DR system, including its pivotal clinical trial. The weaknesses of our study include its retrospective nature and relatively small sample size. Another limitation of our study is that the model did not control for operator or operator experience. However, a learning effect was not present, as there was no difference in time between examination date and initial system deployment between the nondiagnostic and diagnostic groups. We purposefully did not include field-of-view deficiencies in the analysis to account for this. As the next step, we plan to use our data to train other common machine learning models, such as random forests and boosting, test our models prospectively in a clinical setting, and to compare the efficiency of clinics that utilize a reflexive dilation protocol versus clinics that utilize a proactive dilation protocol.
Conclusions
In summary, we show that nondiagnostic results in real-world point-of-care DRD detection using autonomous AI can be predicted, and the resulting predictive dilation workflow has the potential to increase screening rates and improve access to early treatment for DRD, and thereby preventing visual loss and blindness in these vulnerable populations.
Footnotes
Abbreviations: AI, artificial intelligence; AIC, Akaike’s information criterion; aOR, adjusted odds ratio; AUC, area under receiver-operator characteristics curve; BMI, body mass index; CI, confidence interval; DM, diabetes mellitus; DRD, diabetic retinal disease; ETDRS, Early Treatment Diabetic Retinopathy Study; FDA, Food and Drug Administration; IQR, interquartile range; mtmDR, more than mild diabetic retinopathy.
Authors’ Contributions: Conceptualization—TYAL
Methodology—BLS, CC, RG, JHL, JW, MDA, TYAL
Formal analysis—BLS, CC, RG, JHL
Investigation—all authors
Data curation—BLS, KV, TYAL
Writing—Original draft—BLS
Writing—Review & Editing—all authors
Visualization—BLS
Supervision—TYAL, MDA, RMW, RC
Project administration—TYAL
Data Availability: The data set generated during this study that was used to generate the predictive model parameters is available upon reasonable request from the corresponding author TYAL.
The author(s) declared the following potential conflicts of interest with respect to the research, authorship, and/or publication of this article: RMW receives research funding from Dexcom, Inc. and Boehringer Ingelheim, unrelated to this research. MDA reports the following conflicts of interest: Investor, Director, Consultant, and Digital Diagnostics; patents and patent applications assigned to the University of Iowa and Digital Diagnostics that are relevant to the subject matter of this article; Chair Healthcare AI Coalition, Washington DC; member, American Academy of Ophthalmology (AAO) AI Committee; member, AI Workgroup Digital Medicine Payment Advisory Group (DMPAG). The remaining authors declare no competing financial or nonfinancial interests.
Funding: The author(s) received no financial support for the research, authorship, and/or publication of this article.
Conference Presentations: This project was presented as a podium presentation at the Association for Research in Vision & Ophthalmology 2022 Annual Meeting, May 1-5, 2022, Denver, CO, USA.
ORCID iDs: Benjamin L. Shou
https://orcid.org/0000-0003-2825-3301
Jae Hyoung Lee
https://orcid.org/0000-0002-6303-086X
Michael D. Abramoff
https://orcid.org/0000-0002-3490-0037
References
- 1. Fong DS, Aiello L, Gardner TW, et al. Diabetic retinopathy. Diabetes Care. 2003;26(1):226-229. doi: 10.2337/DIACARE.26.1.226 [DOI] [PubMed] [Google Scholar]
- 2. Teo ZL, Tham YC, Yu M, et al. Global prevalence of diabetic retinopathy and projection of burden through 2045: systematic review and meta-analysis. Ophthalmology. 2021;128(11):1580-1591. doi: 10.1016/J.OPHTHA.2021.04.027 [DOI] [PubMed] [Google Scholar]
- 3. Bragge P, Gruen RL, Chau M, Forbes A, Taylor HR. Screening for presence or absence of diabetic retinopathy: a meta-analysis. Arch Ophthalmol. 2011;129(4):435-444. doi: 10.1001/ARCHOPHTHALMOL.2010.319 [DOI] [PubMed] [Google Scholar]
- 4. Murchison AP, Hark L, Pizzi LT, et al. Nonadherence to eye care in people with diabetes. BMJ Open Diabetes Res Care. 2017;5(1):e000333. doi: 10.1136/BMJDRC-2016-000333 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 5. Kuo S, Fleming BB, Gittings NS, et al. Trends in care practices and outcomes among Medicare beneficiaries with diabetes. Am J Prev Med. 2005;29(5):396-403. doi: 10.1016/J.AMEPRE.2005.08.010 [DOI] [PubMed] [Google Scholar]
- 6. Abràmoff MD, Lavin PT, Birch M, Shah N, Folk JC. Pivotal trial of an autonomous AI-based diagnostic system for detection of diabetic retinopathy in primary care offices. NPJ Digit Med. 2018;1:39. doi: 10.1038/s41746-018-0040-6 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 7. Grzybowski A, Brona P. A pilot study of autonomous artificial intelligence-based diabetic retinopathy screening in Poland. Acta Ophthalmol. 2019;97(8):e1149-e1150. doi: 10.1111/AOS.14132 [DOI] [PubMed] [Google Scholar]
- 8. Deb-Joardar N, Germain N, Thuret G, et al. Screening for diabetic retinopathy by ophthalmologists and endocrinologists with pupillary dilation and a nonmydriatic digital camera. Am J Ophthalmol. 2005;140(5):814-821. doi: 10.1016/J.AJO.2005.05.051 [DOI] [PubMed] [Google Scholar]
- 9. Banaee T, Ansari-Astaneh MR, Pourreza H, et al. Utility of 1% tropicamide in improving the quality of images for tele-screening of diabetic retinopathy in patients with dark irides. Ophthalmic Epidemiol. 2017;24(4):217-221. doi: 10.1080/09286586.2016.1274039 [DOI] [PubMed] [Google Scholar]
- 10. van der Heijden AA, Abramoff MD, Verbraak F, van Hecke MV, Liem A, Nijpels G. Validation of automated screening for referable diabetic retinopathy with the IDx-DR device in the Hoorn Diabetes Care System. Acta Ophthalmol. 2018;96(1):63-68. doi: 10.1111/AOS.13613 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 11. Verbraak FD, Abramoff MD, Bausch GCF, et al. Diagnostic accuracy of a device for the automated detection of diabetic retinopathy in a primary care setting. Diabetes Care. 2019;42(4):651-656. doi: 10.2337/DC18-0148 [DOI] [PubMed] [Google Scholar]
- 12. Shah A, Clarida W, Amelon R, et al. Validation of automated screening for referable diabetic retinopathy with an autonomous diagnostic artificial intelligence system in a Spanish population. J Diabetes Sci Technol. 2021;15(3):655-663. doi: 10.1177/1932296820906212 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 13. Grzybowski A, Brona P. Analysis and comparison of two artificial intelligence diabetic retinopathy screening algorithms in a pilot study: IDx-DR and retinalyze. J Clin Med. 2021;10(11):2352. doi: 10.3390/JCM10112352 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14. Beede E, Baylor E, Hersch F, et al. A human-centered evaluation of a deep learning system deployed in clinics for the detection of diabetic retinopathy. Paper presented at the 2020 CHI Conference on Human Factors in Computing Systems; April 2020; Honolulu, HI. doi: 10.1145/3313831.3376718 [DOI] [Google Scholar]
- 15. Patel KH, Javitt JC, Tielsch JM, et al. Incidence of acute angle-closure glaucoma after pharmacologic mydriasis. Am J Ophthalmol. 1995;120(6):709-717. doi: 10.1016/S0002-9394(14)72724-2 [DOI] [PubMed] [Google Scholar]
- 16. Negi A, Vernon SA. An overview of the eye in diabetes. J R Soc Med. 2003;96(6):266-272. doi: 10.1258/JRSM.96.6.266 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 17. Niedermaier ON, Smith ML, Beightol LA, Zukowska-Grojec Z, Goldstein DS, Eckberg DL. Influence of cigarette smoking on human autonomic function. Circulation. 1993;88(2):562-571. doi: 10.1161/01.CIR.88.2.562 [DOI] [PubMed] [Google Scholar]
- 18. Scanlon PH, Foy C, Malhotra R, Aldington SJ. The influence of age, duration of diabetes, cataract, and pupil size on image quality in digital photographic retinal screening. Diabetes Care. 2005;28(10):2448-2453. [DOI] [PubMed] [Google Scholar]
- 19. Guillon M, Dumbleton K, Theodoratos P, Gobbe M, Wooley CB, Moody K. The effects of age, refractive status, and luminance on pupil size. Optom Vis Sci. 2016;93(9):1093-1100. doi: 10.1097/OPX.0000000000000893 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 20. Wolf RM, Liu TYA, Thomas C, et al. The SEE study: safety, efficacy, and equity of implementing autonomous artificial intelligence for diagnosing diabetic retinopathy in youth. Diabetes Care. 2021;44(3):781-787. doi: 10.2337/DC20-1671 [DOI] [PubMed] [Google Scholar]