Skip to main content
JMIR Medical Informatics logoLink to JMIR Medical Informatics
. 2021 Dec 13;9(12):e27363. doi: 10.2196/27363

Machine Learning Algorithms to Detect Subclinical Keratoconus: Systematic Review

Howard Maile 1,#, Ji-Peng Olivia Li 2,#, Daniel Gore 2, Marcello Leucci 2, Padraig Mulholland 1,2,3, Scott Hau 2, Anita Szabo 1, Ismail Moghul 2, Konstantinos Balaskas 2, Kaoru Fujinami 1,2,4,5, Pirro Hysi 6,7, Alice Davidson 1, Petra Liskova 8,9, Alison Hardcastle 1, Stephen Tuft 1,2, Nikolas Pontikos 1,2,
Editors: Rita Kukafka, Gunther Eysenbach
Reviewed by: Ayan Chatterjee, Randa Salah Mahmoud
PMCID: PMC8713097  PMID: 34898463

Abstract

Background

Keratoconus is a disorder characterized by progressive thinning and distortion of the cornea. If detected at an early stage, corneal collagen cross-linking can prevent disease progression and further visual loss. Although advanced forms are easily detected, reliable identification of subclinical disease can be problematic. Several different machine learning algorithms have been used to improve the detection of subclinical keratoconus based on the analysis of multiple types of clinical measures, such as corneal imaging, aberrometry, or biomechanical measurements.

Objective

The aim of this study is to survey and critically evaluate the literature on the algorithmic detection of subclinical keratoconus and equivalent definitions.

Methods

For this systematic review, we performed a structured search of the following databases: MEDLINE, Embase, and Web of Science and Cochrane Library from January 1, 2010, to October 31, 2020. We included all full-text studies that have used algorithms for the detection of subclinical keratoconus and excluded studies that did not perform validation. This systematic review followed the PRISMA (Preferred Reporting Items for Systematic Reviews and Meta-Analyses) recommendations.

Results

We compared the measured parameters and the design of the machine learning algorithms reported in 26 papers that met the inclusion criteria. All salient information required for detailed comparison, including diagnostic criteria, demographic data, sample size, acquisition system, validation details, parameter inputs, machine learning algorithm, and key results are reported in this study.

Conclusions

Machine learning has the potential to improve the detection of subclinical keratoconus or early keratoconus in routine ophthalmic practice. Currently, there is no consensus regarding the corneal parameters that should be included for assessment and the optimal design for the machine learning algorithm. We have identified avenues for further research to improve early detection and stratification of patients for early treatment to prevent disease progression.

Keywords: artificial intelligence, machine learning, cornea, keratoconus, corneal tomography, subclinical, corneal imaging, decision support systems, corneal disease, keratometry

Introduction

Background

Keratoconus is a bilateral ectatic disease of the cornea that can cause visual loss through corneal distortion and scarring [1,2]. The prevalence of keratoconus varies from 1 in 375 people in Northern Europe [3] to as high as 1 in 48 in some ethnic groups [4,5], with studies suggesting a higher incidence in Middle-Eastern, West Indian, and Asian populations with faster progression [6-8]. The onset of the disease typically occurs after puberty, with subsequent progression at a variable rate over 2 to 3 decades [6]. A recent meta-analysis found that patients <17 years are likely to progress more than 1.5 D in Kmax over 12 months, and those with steeper Kmax of more than 55 D are likely to have at least 1.5 D Kmax progression [6].

As the disease advances, corneal distortion can reach a stage where spectacle-corrected vision is inadequate, and patients must rely on soft or rigid contact lenses to achieve good functional vision [9]. However, contact lenses are not always tolerated, and visual impairment can severely affect quality of life [10,11]. In the natural course of the disease, approximately 20% of the patients are offered a corneal transplant to improve their vision but at the risk of postoperative complications (eg, microbial keratitis and inflammation), potential allograft rejection, and transplant failure [7,12,13]. Most individuals with keratoconus are identified because of the symptoms of visual disturbance or an increase in astigmatism at refraction. Therefore, it is inevitable that most individuals with keratoconus are detected at a stage when visual deterioration has already occurred [14].

The detection of keratoconus at an earlier stage has become increasingly relevant since the introduction of corneal collagen cross-linking (CXL). This is a photochemical treatment of the cornea with UV-A light following the application of riboflavin (vitamin B2), which can arrest the progression of keratoconus in 98.3% of the eyes even in relatively advanced cases [15-20]. The benefit of early treatment to minimize visual loss is clear, and there is evidence that it is cost-effective [21-23], but the mechanism to improve early diagnosis by community-based optometrists is challenging because asymptomatic patients with subclinical disease are unlikely to seek review [14]. Improved detection will probably require improved access or efficient community screening with expensive imaging equipment [24].

Machine learning is a branch of artificial intelligence centered on writing a software capable of learning from data in an autonomous fashion by minimizing a loss function or maximizing the likelihood [25]. It can be broadly classified as either supervised or unsupervised learning [26]. In supervised learning, the algorithm is trained with input data labeled with a desired output so that it can predict an output from unlabeled input data [27]. In comparison, in unsupervised learning, the algorithm is not trained using labeled data. Instead, the algorithm is used to identify patterns or clusters in the data [28]. When applied to the field of keratoconus detection, machine learning may be used to analyze a large number of corneal parameters that can be derived from corneal imaging as well as other clinical and biometric measures such as visual acuity and refraction to predict the disease [29]. It can also be applied directly to imaging data to work at the pixel level [30]. Deep learning, a specific branch of machine learning, uses artificial neural networks (NNs) with multiple layers to process input data [31]. It is particularly well suited to the segmentation or classification of corneal images [32]. Both machine learning and deep learning may facilitate superior diagnostic ability that, when implemented as automated screening tools, could result in significant advances in case detection, mitigating both the cost of new imaging hardware and the burden on ophthalmic health care professionals [33]. In addition, through unsupervised learning, it may be possible to discover previously unknown disease subtypes or features [34,35].

Unlike diabetic retinopathy, which uses a widely adopted diagnostic grading system (Early Treatment Diabetic Retinopathy Study) [36] and in which the diagnosis of early disease is based on the presence of discrete entities on the retina (eg, microaneurysms), the diagnostic grading of subclinical keratoconus has not yet reached the same level of consensus [37]. Frequently used grading systems such as Amsler-Krumeich [38] and ABCD [39] do not specifically include a grade for subclinical keratoconus. More detailed information about keratoconus grading systems is available in Multimedia Appendix 1 [37-43].

Case Definition for Keratoconus

Several terms describe the early stage of keratoconus before vision is affected, including forme fruste keratoconus (FFKC), keratoconus suspect, subclinical keratoconus, and preclinical keratoconus. The most commonly used terms are FFKC and subclinical keratoconus, but there is no consensus on their definition [44].

We have included all papers that contain an identifiable subgroup of eyes with any of the aforementioned definitions because of the overlap in the nomenclature and lack of evidence as to which, if any, pose a particular risk for progression to clinical keratoconus. We excluded papers that only consider eyes with established keratoconus.

Objectives

The aim of this study is to critically evaluate the literature on the algorithmic detection of subclinical keratoconus and its equivalent definitions. Advanced keratoconus is relatively easy to diagnose clinically, such that developing machine learning algorithms to identify advanced disease has limited utility. Therefore, we directed this review to publications that have included detection of subclinical keratoconus because identifying these individuals would allow for early treatment with CXL to reduce the likelihood of disease progression and visual loss. We have structured our review both around the different types of available input data (parameters, indices, and corneal imaging systems) and the machine learning algorithms for keratoconus detection. In addition, we investigated the validation methodology within each study and assessed the potential for bias.

Research Questions

Our specific research questions are as follows:

  1. Research question 1: What input data types have been used within subclinical keratoconus detection algorithms and how have they performed?

  2. Research question 2: What machine learning algorithms have been used for subclinical keratoconus detection and how have they performed?

  3. Research question 3: How was algorithm validation handled among the selected manuscripts?

Methods

Search Strategy

We conducted a literature review of the evidence for the utility of machine learning applied to the detection of keratoconus published between January 1, 2010, and October 31, 2020. The PRISMA (Preferred Reporting Items for Systematic Reviews and Meta-Analyses) Statement 2009 criteria [45] was followed to search 4 bibliographic databases: MEDLINE, Embase, Web of Science, and Cochrane Library using keyword search on their title, abstract, and keywords. The review was not registered, and no protocol was prepared.

We used the following keyword search for literature review in bibliographic databases: ((keratoconus) OR (cornea* protru*) OR (cornea* ectasia)) AND ((algorithm) OR (machine learn*) OR (deep learn*) OR (artificial intelligence) OR (detect*) OR (diagnos*) OR (screen*) OR (examin*) OR (analys*) OR (investigat*) OR (identif*) OR (discover*) OR (interpret*) OR (test*))

Inclusion and Exclusion Criteria

We included studies that investigated the detection of early keratoconus or included a subgroup of patients with early disease, as defined by one of the following terms: subclinical keratoconus, FFKC, preclinical keratoconus, suspected keratoconus, unilateral keratoconus (normal fellow eye), and asymmetric ectasia (normal fellow eye) and any definition considered equivalent to the aforementioned terms. The studies should have reported the performance of their model on a data set that was separate from the training data set (often called a validation or a test set). This includes splitting of the data set into training and test sets (eg, 70% training and 30% testing), K-fold cross-validation (an extension of simple splitting, but the process is repeated K times, eg, when K=10, partition the data set into 90% for training the model and 10% for testing, and the process is repeated 10 times by choosing a different 10% partition each time for testing), or evidence of a validation study where the aim is to assess a previously derived model on a new data set (also known as an external validation). Finally, the full-text article should be available, and only papers published in English were considered.

We excluded papers based on the detection of early keratoconus defined as Amsler-Krumeich stages 1 or 2, as this represents established keratoconus with both clinical and topographical features [46].

Data Synthesis

On the basis of the inclusion criteria, 2 reviewers (HM and JPOL) screened the initial results. These results were then screened for the exclusion criteria by HM and NP. The PRISMA diagram is presented in Figure 1. Any disagreements in meeting the inclusion or exclusion criteria were resolved by discussion. Once the set of articles was finalized, 2 reviewers (HM and JPOL) analyzed each article and extracted the following information in a master table presented in Multimedia Appendix 2 [14,47-71]: author and year, title, system, sample source, country, age, gender, number of eyes for each group, diagnosis details, validation details, input details, input types, method, classification groups, sensitivity, specificity, accuracy, precision, area under the receiver operating characteristic curve (AUC), and source code availability. We summarized the most important information for all the results in Table 1. The main effect measures sought were sensitivity and specificity. If these statistics were not directly available from the article, they were calculated manually using their standard definitions [72]. To visually compare the results, we plotted the sensitivity and specificity across all studies for diagnostic criteria and detection systems in Multimedia Appendix 3.

Figure 1.

Figure 1

Filtering steps taken to accept or exclude studies in the systematic review.

Table 1.

Summary of the 26 published studies that included the use of machine learning for the detection of subclinical keratoconus.

Study System Number of eyes Fellow eyea Input types Method Results (%)


Normal Subclinical keratoconus


Sensitivity Specificity
Arbelaez et al [47] Sirius 1259 426 No Elevation, keratometry, pachymetry, and aberrometry SVMb 92 97.7
Saad et al [48] Orbscan 69 34 No Pachymetry, keratometry, elevation, and Displacement DAc 92 96
Smadja et al [49] GALILEI 177 47 Yes Keratometry, pachymetry, elevation, aberrometry, demographic, and indices DTd 93.6 97.2
Ramos-Lopez et al [50] CSO topography system 50 24 No Elevation and displacement Linear regression 33 78
Cao et al [14] Pentacam 39 49 No Keratometry, pachymetry, and demographic RFe, SVM, K-nearest neighbors, LoRf, DA, Lasso regression, DT, and NNg 94 90
Buhren et al [51] Orbscan IIz 245 32 No Keratometry, pachymetry, aberrometry, and elevation DA 78.1 83.3
Chan et al [52] Orbscan IIz 104 24 Yes Pachymetry, keratometry, elevation, and displacement DA 70.8 98.1
Kovacs et al [53] Pentacam 60 15 Yes Keratometry, pachymetry, elevation, indices, and displacement NN 90 90
Saad et al [54] OPD-scan 114 62 Yes Keratometry, aberrometry, and indices DA 63 82
Ruiz Hidalgo et al [55] Pentacam HR 194 67 No Keratometry, pachymetry, and aberrometry SVM 79.1 97.9
Ruiz Hidalgo et al [56] Pentacam HR 44 23 No Keratometry, pachymetry, and indices SVM 61 75
Xu et al [57] Pentacam HR 147 77 Yes Pachymetry, elevation, and keratometry DA 83.7 84.5
Ambrosio et al [58] Pentacam+Corvis ST 480 94 Yes Pachymetry, elevation, keratometry, and Biomechanical RF, SVM, and LoR 90.4 96
Sideroudi et al [59] Pentacam 50 55 No Keratometry LoR 91.7 100
Francis et al [60] Corvis ST 253 62 Yes Biomechanical LoR 90 91
Yousefi et al [61] SS-1000 CASIA 1970 796 No Elevation, pachymetry, and aberrometry Unsupervised 88 14
Lopes et al [62] Pentacam HR 2980 188 Yes Pachymetry, elevation, indices, and displacement DA, SVM, naive Bayes, NN, and RF 85.2 96.6
Steinberg et al [63] Pentacam+Corvis ST 105 50 Yes Pachymetry, elevation, keratometry, and biomechanical RF 63 83
Issarti et al [64] Pentacam 312 90 Yes Elevation and pachymetry NN 97.8 95.6
Chandapura et al [65] RCTVue+Pentacam 221 72 Yes Keratometry, elevation, pachymetry, aberrometry, and indices RF 77.2 95.6
Xie at al [66] Pentacam HR 1368 202 No Heat maps CNNh 76.5 98.2
Kuo et al [67] TMS-4+Pentacam+Corvis ST 170 28 No Heat maps CNN 28.5 97.2
Shi et al [68] Pentacam+ultrahigh resolution optical coherence tomography 55 33 Yes Keratometry, elevation, pachymetry, indices, and demographic NN 98.5 94.7
Toprak et al [69] MS-39 66 50 Yes Keratometry, pachymetry, and displacement LoR 94 98.5
Issarti et al [70] Pentacam HR 304 117 Yes Elevation and Pachymetry NN 85.2 70
Lavric et al [71] SS-1000 CASIA 1970 791 No Keratometry, pachymetry, and aberrometry 25 machine learning methods compares 89.5 96

aFellow eye indicates whether the study defined subclinical keratoconus as the fellow eye of an individual with apparently unilateral keratoconus, with no clinical or topographical features of keratoconus.

bSVM: support vector machine.

cDA: discriminant analysis.

dDT: decision tree.

eRF: random forest.

fLoR: logistic regression.

gNN: neural network.

hCNN: convolutional neural network.

Bias Assessment

When assessing bias within the included studies, we used a tailored version of the QUADAS (Quality Assessment of Diagnostic Accuracy Studies)-2 tool [73], which consists of 4 domains: patient selection, index test, reference standard, flow, and timing. The 26 studies were assessed by 3 reviewers (HM, JPOL, and NP) such that each study was assessed by at least 2 reviewers.

Results

Overview

We identified 1998 potentially relevant papers published between 2010 and 2020. After filtering, we included 26 articles in our qualitative analysis. Table 1 summarizes these results, and a more extensive version can be found in Multimedia Appendix 2. To address research question 1, the results are discussed in terms of their input data. Charts displaying aggregate sensitivity and specificity can be found in Multimedia Appendix 3. To address research question 2, the results are considered in terms of the machine learning algorithms. Figures 2 and 3 present organizational diagrams of data categorization and machine learning algorithms, respectively. To maintain consistency, we opted to use the term subclinical keratoconus throughout regardless of the nomenclature used by the original authors. The original term is included in parenthesis, and details of the exact definition can be found in Multimedia Appendix 2.

Figure 2.

Figure 2

Organizational diagram of relevant data types reported to be used for the detection of subclinical keratoconus.

Figure 3.

Figure 3

Organizational diagram of relevant machine learning algorithms used for the detection of subclinical keratoconus.

Research Question 1: What Input Data Types Have Been Used Within Subclinical Keratoconus Detection Algorithms and How Have They Performed?

This section is subdivided according to the input data types used for the detection of subclinical keratoconus, as presented in the organizational chart in Figure 2.

Aberrometry

Aberrometry was used to detect subclinical keratoconus in 31% (8/26) of the papers [47-49,51,55,61,65,71]. Aberrations are produced by imperfections in the optical quality of the refracting surface of the eye, including the cornea and the lens. Higher-order aberrations (HOAs) are measured from the distortion of a plane wavefront of light passing through the optics of the eye. However, HOAs can also be derived indirectly from the measurement of any distortion (eg, elevation) of the corneal surfaces. They can be described as a set of Zernike polynomials or with Fourier analysis. Using the Zernike method, aberrations can be subclassified as lower-order aberrations and HOAs. Lower-order aberrations include simple defocus (myopia or hyperopia) and regular astigmatism, which account for approximately 90% of the refractive error of the normal eye [74]. The most clinically relevant HOAs are spherical aberration, coma, and trefoil that cannot be corrected by glasses or a soft contact lens. In keratoconus, the irregular distortion of the front and back surfaces of the cornea causes visually significant HOAs. Arbelaez et al [47] analyzed these parameters in their subclinical keratoconus detection model and included a weighted sum of HOAs (known as the Baiocchi-Calossi-Versaci index) and the root mean square of HOAs. Moreover, 5 other studies also used derived Zernike aberrometry data [48,49,51,65,71].

Corneal Imaging Data and Derived Parameters

Overview

Corneal images were used to detect subclinical keratoconus in 96% (25/26) of the papers. There are various acquisition techniques, including Scheimpflug optics (Pentacam [Oculus GmbH] or Sirius [CSO]), anterior segment optical coherence tomography (AS-OCT; MS-39 [CSO], or CASIA [Tomey]), and horizontal slit-scanning systems such as Orbscan II (Bausch & Lomb). These systems incorporate a software that processes the images to derive numerical indices or secondary images, such as heat maps, to visualize various aspects of corneal shape. These parameters can be classified as measurements of the corneal surface radius of curvature (keratometry), elevation or depression of a point on the corneal surface from the mean (elevation map), corneal thickness (pachymetry), or displacement from the apex of the cornea. Figure 4 illustrates the main parameter types in a schematic diagram. Figure 5 shows an example of the Pentacam heat map for an eye with subclinical keratoconus. See Multimedia Appendix 4 for an example of advanced keratoconus (fellow eye for the same patient).

Figure 4.

Figure 4

Schematic diagram illustrating the 4 basic corneal parameters that can be measured using corneal imaging. (A) pachymetry. (B) displacement: distance between the apex of the cornea and the point of minimum thickness. (C) and (D) represent 2 methods of calculating the best-fit sphere (BFS). In (C) the BFS is fitted to both the normal peripheral posterior surface (blue) and the abnormal anterior protrusion of the central posterior surface (green). In (D) the BFS is fitted to only the normal peripheral posterior surface (blue) excluding the abnormal central posterior surface (green), leading to a larger relative elevation than in (C). (E) the smallest radius of curve of the astigmatic corneal surface corresponds to the largest refractive power (Kmax) and the largest radius of curve corresponds to the smallest refractive power (Kmin). CCT: central corneal thickness.

Figure 5.

Figure 5

Heat maps of a subclinical keratoconus eye derived from Scheimpflug corneal imaging using the Pentacam HR device. The axial/sagittal map (A) depicts the curvature of the anterior corneal surface in dioptres and shows mild inferior steepening, while the pachymetry map (C) shows thinning in the same region. The front and back elevation maps (B and D, respectively) show a moderate increase in inferior elevation. BFS: best-fit sphere; OS: left eye.

In the following subsections, we briefly discuss the use of quantitative measures derived from corneal imaging when used in isolation or in combination with machine learning models.

Keratometry Parameters

Keratometric data are one of the most commonly used parameters in the literature, with 69% (18/26) of the papers incorporating keratometry as one of the parameters in their model [14,47-49,51-54,56,57,59,61,65,68,71,75]. Keratometric parameters measure the radius of curvature of the anterior or posterior corneal surfaces. Examples include the meridian with the minimum corneal radius of curvature (corresponding to Kmax) and maximum curvature (corresponding to Kmin). When looking at individual keratometric parameters derived using Fourier analysis for subclinical keratoconus detection, Sideroudi et al [59] achieved a predictive accuracy of over 90% using higher-order irregularities, asymmetry, and regular astigmatism, primarily in the corneal periphery.

Elevation Parameters

Overall, 62% (16/26) of the papers incorporated elevation parameters in their analysis [47-53,57,58,61-65,68,70]. Elevation represents points above or below the BFS of the corneal surface measured in microns (Figure 4). For the posterior cornea, this is measured either as the divergence from the best fit of the whole posterior corneal diameter or as the divergence from the best fit of the annulus of the peripheral posterior corneal surface outside the central 4 mm [76]. The latter method, the Belin-Ambrosio map, better describes the central corneal elevation, which is a feature of keratoconus. Values can be presented as either color-coded maps or individual parameters such as maximum anterior elevation, maximum posterior elevation, or derived data such as aberrometry.

Posterior corneal curvature consistently outperforms other parameters in the discrimination of subclinical keratoconus [47,49,56,57]. Its inclusion increases the sensitivity of a support vector machine (SVM) from 75.2% to 92% and precision from 57.4% to 78.8% but has a limited impact on specificity [47]. Posterior corneal curvature, measured using a Pentacam (Scheimpflug) device and analyzed using an SVM, was also found to be an important parameter for sensitivity and much less so for specificity and AUC [14]. Similarly, using the Galilei (Scheimpflug) device, the posterior asphericity asymmetry index was found to be the variable with the most discriminatory power when differentiating normal from subclinical keratoconus, followed by corneal volume [49]. Conversely, analysis of anterior surface topographical parameters and aberrometry using the random forest algorithm did not discriminate subclinical keratoconus (very asymmetric ectasia-normal topography) from normal eyes [65].

Saad et al [54] showed that combining parameters obtained from the anterior corneal curvature corneal wavefront and Placido-derived indices lead to a better discriminative ability between normal and subclinical keratoconus eyes (FFKC) over a Placido-only–based algorithm.

Displacement Parameters

A total of 23% (6/26) of the papers used displacement parameters in their analysis [48,50,52,53,62,75]. These represent measures such as the displacement of the point of minimum corneal thickness from the corneal apex. Of these, 3 papers used the displacement of the thinnest point from the geometric center of the cornea in their model [48,52,77]. Kovacs et al [53] used the vertical and horizontal decentration of the thinnest point and found them to be the best parameters to discriminate normal fellow eyes of keratoconus from control eyes using an NN.

Pachymetry Parameters

Overall, 77% (20/26) of the papers used pachymetry data in their model, making it one of the most commonly used parameters in the literature [14,47-49,51-53,55-58,61-65,69-71]. Pachymetry is the thickness of the cornea, measured using either ultrasound or imaging techniques. Simple examples include central corneal thickness and the thinnest point of the cornea. A reduction in the thickness of the central cornea is a fundamental biomarker of keratoconus [2].

Summary Indices

In total, 23% (6/26) of the papers used summary indices in their model [14,49,53,65,68,78]. In addition to single-parameter measurements (eg, central corneal thickness), tomographic systems such as the Pentacam can combine measurements to compute derived indices that estimate the regularity of corneal shape. Basic indices such as the index of surface variance, index of vertical asymmetry, or index of height asymmetry are formed from multiple data points. Composite indices are formed from other indices and data points. Examples include the Keratoconus index, keratoconus percentage index, and Belin/Ambrosio enhanced ectasia display (BAD-D). In a recent study, Shi et al [68] used 6 indices from the Pentacam along with keratometric, elevation, and pachymetric parameters derived from the Pentacam and ultrahigh resolution optical coherence tomography to create an NN classifier to discriminate between normal and subclinical keratoconus eyes. Using 50 normal eyes, 38 eyes with keratoconus, and 33 eyes with subclinical keratoconus, they achieved 98.5% sensitivity and 94.7% specificity. However, the results require further validation because of the small number of eyes in this group. Furthermore, the authors did not include a comparison between existing detection metrics, such as BAD-D.

Heat Maps

A total of 8% (2/26) of the papers used heat maps in their detection model [66,67]. Modalities such as Scheimpflug and AS-OCT capture images at various corneal meridians and subsequently use these data to derive the heat maps that facilitate visual interpretation of the data, although there is extrapolation of the data in areas between the imaged meridians. For example, the Pentacam can translate the raw images into several types of color heat maps (eg, axial curvature, posterior or anterior elevation, and regional pachymetry) based on the same original tomography data set. Prediction models applied to images often use convolutional NNs (CNNs), and studies applying these methods are discussed in detail in the next section addressing research question 2. To the best of our knowledge, no system has used raw pixel values from Scheimpflug or AS-OCT images directly when detecting subclinical keratoconus.

Biomechanical Data

Overall, 12% (3/26) of the papers incorporated biomechanical data in their analysis [58,60,63]. Corneal biomechanics refers to the distortion response of the cornea to an applied force. The Ocular Response Analyzer (Reichert Ophthalmic Instruments) uses a puff of air directed to the cornea, and the deformation response is measured. Two common indices have been reported: corneal hysteresis and corneal resistance factor. However, there is disagreement regarding their utility in the diagnosis of keratoconus [79,80]. Another device using the same principle is the Corvis ST (Oculus Optikgeräte GmbH), which uses a high-speed Scheimpflug camera to measure distortion in cross-sectional images. Numerous studies have described the application of machine learning to analyze biomechanical data, but very few validated their results; therefore, they have been excluded from this review. Ambrosio et al [58] combined Pentacam and Corvis ST data to create the Tomographic and Biomechanical Index, and this was followed up with a validation study [63]. Francis et al [60] used biomechanical data from the Corvis ST device when diagnosing keratoconus and achieved very high sensitivity (99.5%) and specificity (100%). However, when validating their model, they have only discriminated between 2 groups—a group combining subclinical keratoconus and keratoconus eyes, and a group of normal eyes. This represents an easier problem than including a distinction between normal and subclinical keratoconus eyes.

Demographic Risk Factors

A total of 15% (4/26) of the papers chose to include demographic data, such as age or sex, in their model [14,49,62,68]. Cao et al [14] demonstrated that sex was an important parameter in a minimum set that achieved the highest AUC using the random forest method, although their data set was small (49 subclinical keratoconus and 39 control eyes). Ethnicity, a major association with disease prevalence, risk of progression, disease severity, and acute corneal hydrops in Asian and Black populations [81], was not included in any model, although some studies have examined single ethnicities [66]. Ethnicity as a parameter should be considered by future investigators. No studies included other risk factors such as atopy and eye rubbing as model parameters, and these should be considered in future studies.

Research Question 2: What Machine Learning Algorithms Have Been Used for Subclinical Keratoconus Detection and How Have They Performed?

In most cases, researchers have used combinations of parameters and indices within machine learning algorithms to diagnose subclinical keratoconus. This section is subdivided according to the machine learning techniques that were applied. Figure 3 presents an organizational diagram of the relevant machine learning algorithms. There are several other algorithms, but discussion of these is beyond the scope of the review, and we have chosen to include only the methods found in our Results section.

Neural Networks

Overview

NNs consist of a series of interconnected layers of neurons and are thus loosely modeled on the structure within the human brain. Each neuron computes a nonlinear function of its inputs, and the network is trained until the output aligns optimally with the ground truth labels. Kovacs et al [53] used a combination of 15 keratometric, pachymetric, and elevation parameters in an NN classifier to discriminate healthy corneas from fellow eyes of patients with unilateral keratoconus. The patient data included 60 normal eyes from 30 patients, 60 bilateral keratoconus eyes from 30 patients, and 15 normal eyes from patients presenting with unilateral keratoconus. When classifying the normal eyes of the patients with unilateral keratoconus with clinical grading as a reference, they achieved 90% sensitivity and 90% specificity. They took a novel approach of training on both the eyes of the patients, which allowed them to incorporate the effect of any intereye asymmetry when detecting unilateral keratoconus. Shi et al [68] combined keratometric, elevation, and pachymetric parameters derived from Pentacam images and ultrahigh resolution optical coherence tomography to create an NN classifier for discriminating normal from subclinical keratoconus eyes. Using Pentacam elevation and pachymetry maps within a hybrid NN model, Issarti et al [64] demonstrated superiority over other common diagnostic indices such as BAD-D and topographical keratoconus classification.

Convolutional Neural Networks

When images are used for analysis, NNs with a large number of processing layers such as CNNs are often employed because of their ability to make inferences from 2D or 3D data structures through deep learning [82]. For example, Xie et al [66] used data from 1368 normal eyes, 202 eyes with early keratoconus, 389 eyes with more advanced keratoconus, and 369 eyes with subclinical (suspected) keratoconus to develop an automatic classifier. They achieved 76.5% sensitivity and 98.2% specificity when classifying subclinical keratoconus. However, the heat maps used were produced by Pentacam; therefore, it should be noted that the technique may not be transferable to other systems or even future Pentacam software iterations. Kuo et al [67] included 150 normal, 170 keratoconus, and 28 subclinical eyes in their study and used the Tomey TMS-4 topography system to produce corneal heat maps and trained 3 different CNN architectures (VGG16, InceptionV3, and ResNet152). When attempting to identify the 28 subclinical keratoconus eyes, they applied the VGG16 model and achieved barely satisfactory results with an accuracy of 28.5% when a threshold of 50% was applied. These results suggest that subclinical keratoconus cannot yet be detected with high sensitivity using CNNs on heat map images.

Decision Trees

The classification of data in a decision tree uses a binary decision at each node in the tree to determine the branch to take next. Starting from the root, the classification is determined by following each branch to its terminal node. Smadja et al [49] used a decision tree to classify normal, keratoconus, and subclinical (FFKC) keratoconus eyes. They enrolled 177 normal eyes, 148 keratoconus eyes, and 47 subclinical eyes. They used 55 parameters (including curvature, elevation, corneal wavefront, corneal power, pachymetry, and age) collected from the Galilei dual Scheimpflug camera, achieving 93.6% sensitivity and 97.2% specificity when classifying subclinical from normal. Cao et al [14] also evaluated a decision tree algorithm for classifying subclinical keratoconus but achieved lower sensitivity (82%) and specificity (78%). They attributed the comparatively inferior performance to the fact that Smadja et al [49] used additional machine-specific indices that they did not have access to.

Random Forests

Random forests combine a large number of decision trees into a single model [83]. Lopes et al [62] compared this method with other methods (naive Bayes, NNs, SVMs, and discriminant analysis) by training models on 71 post–laser-assisted in situ keratomileusis (LASIK) ectasia eyes, 298 post-LASIK eyes without ectasia, and 183 eyes with keratoconus. They included keratometry, pachymetry, elevation, and various Pentacam indices. The models were validated on an external data set containing 298 normal eyes (stable LASIK), 188 keratoconus eyes (very asymmetric ectasia-ectatic), and 188 subclinical eyes (very asymmetrical ectasia-normal topography). The latter 2 groups were collected from the same set of patients. They found that the random forest model performed best when detecting subclinical eyes with an 85.2% sensitivity. This accuracy is lower than that of other comparable studies, which is probably caused by their inclusion of external validation rather than an inferior model. The authors also note that their model classifies among 3 groups, whereas other related studies (such as that by Arbelaez et al [49]) only classify between 2 groups (eg, subclinical vs normal). This important distinction is expanded upon in the Discussion section.

Discriminant Analysis

Discriminant analysis uses a linear combination of variables that optimally separate 2 or more classes of data. Xu et al [57] used this method to classify eyes as either normal, subclinical keratoconus, or keratoconus. In total, 147 normal eyes, 139 eyes with keratoconus, and 77 eyes with subclinical keratoconus were included in the training set and verified on a separate set of 97 normal and 49 subclinical keratoconus eyes. They applied the Zernike fitting method to corneal pachymetry and elevation data derived from the Pentacam and achieved an AUC of 92.8% when discriminating subclinical keratoconus. Saad et al [54] also used discriminant analysis to classify eyes as either subclinical (FFKC) keratoconus or normal. They used a combination of wavefront aberrometry and Placido disc indices in their model with a total of 8 parameters using the OPD-Scan (Nidek Co Ltd). The model was trained on 114 normal and 62 subclinical eyes and validated on 93 normal and 82 subclinical eyes. Using training data only, the model achieved 89% sensitivity and 92% specificity, but when applied to the validation set, the accuracy dropped significantly to 63% sensitivity and 82% specificity. This highlights the need for external validation when reporting the performance of the detection algorithms.

Support Vector Machines

SVMs translate data into a higher-dimensional space where a dividing line (known as a hyperplane) separates the data such that the distance between the hyperplane and any given data point is maximized [26]. When 8 different machine learning algorithms were compared for classifying subclinical keratoconus on the same data set, SVMs achieved the highest sensitivity (94%) [14]. Arbelaez et al [47] achieved even higher sensitivities using SVMs with a large data set of 1259 normal eyes and 426 with subclinical keratoconus. They used 200 eyes from each group for training and the remainder for testing, achieving 92% sensitivity and 97.7% specificity. Ruiz Hidalgo et al [56] used 25 topographic or tomographic Pentacam-derived parameters to verify their SVM model. They included 131 patients in their study and provided results for 2 classifications from separate hospitals: Antwerp University Hospital and Rothschild Foundation, Paris. When classifying the 4 groups (keratoconus, subclinical, normal, and postrefractive surgery), the sensitivity for subclinical keratoconus detection was 61% compared with that of the Antwerp University Hospital classification and 100% compared with that of the Rothschild classification. This was a comprehensive validation study that compared multiple methods with 2 subjective reference standards. Only a small number of subclinical keratoconus cases (approximately 20) were included in this study, and a larger study is required to verify these results.

Logistic Regression

Logistic regression is commonly used to perform classification from a set of independent variables [26]. It transforms its output using the sigmoid function to return a probability that can then be thresholded for classification. When classifying subclinical keratoconus, 3 studies used this technique exclusively [59,60,75]. Sideroudi et al [59] used logistic regression to explore the diagnostic capacity of Fourier-derived posterior keratometry parameters (spherical component, regular astigmatism, asymmetry, and irregular astigmatism) extracted from Pentacam Scheimpflug images. They included 50 normal eyes, 80 eyes with keratoconus, and 55 with subclinical keratoconus (defined as a clinically normal eye with abnormal topography, where the fellow eye has advanced keratoconus) and validated their model on 30% of the data set. Their model attained 91.7% sensitivity and 100% specificity when classifying between subclinical keratoconus and normal eyes. Although these results are among the best reported, the study has yet to be validated using an external data set. Other studies implemented logistic regression as part of a wider comparison of machine learning algorithms [14,58].

Comparative Studies

Few studies have applied multiple machine learning algorithms to the same data set. Cao et al [14] tested 8 machine learning algorithms on the same data set of 39 normal control eyes and 49 eyes with subclinical keratoconus. Age, sex, and 9 corneal parameters from the Pentacam tomography were used, and the authors found that random forest, SVM, and K-nearest neighbors had the best performance. Random forests had the highest AUC of 0.97, SVM had the highest sensitivity (94%), and K-nearest neighbors had the best specificity (90%). Although they verified their results with 10-fold cross-validation, it would be instructive to repeat the analysis on a larger external data set. Ambrosio et al [58] also performed an analysis across algorithms including logistic regression, SVMs, and random forests to classify between 4 groups: normal, keratoconus, very asymmetrical ectasia-ectatic, and subclinical keratoconus (very asymmetric ectasia-normal). They used both Scheimpflug tomography and biomechanical data and included 480 normal eyes, 204 eyes with keratoconus, 72 eyes classified as very asymmetrical ectasia-ectatic, and 94 subclinical keratoconus eyes. When considering subclinical keratoconus, the random forest model performed the best, with 90.4% sensitivity and 96% specificity. The final model was named the Tomography and Biomechanical Index and was validated by leave-one-out cross-validation, resulting in as many models as there were subjects (N=850). Lopes et al [62] also performed a comparative analysis and found that random forests performed best when trying to classify 3 groups of eyes (including subclinical eyes). Lavric et al [71] provided the largest comparative study for detecting subclinical keratoconus. The authors included 1970 normal eyes, 390 eyes with keratoconus, and 791 subclinical (FFKC) keratoconus eyes in their study and used keratometric, pachymetric, and aberrometric data from the CASIA AS-OCT system in their analysis across 25 different machine learning algorithms. When they classified the 3 groups simultaneously, they found that the most accurate method was SVM, which attained 89.5% sensitivity for the detection of subclinical keratoconus, and the results were validated using 10-fold cross-validation. The limitations of this study include the use of the CASIA ectasia screening index (ESI) for the classification of the severity of keratoconus, which may not agree with clinical diagnosis, and that the analyzed parameters are closely tied to the CASIA device, which limits generalizability to other systems.

Unsupervised Learning

Unsupervised learning represents a distinct approach to the detection of subclinical keratoconus by attempting to identify groups of similar eyes without prelabeled data. Yousefi et al [61] used a 2-step approach that combined dimensionality reduction and density-based clustering to cluster a cohort of 3156 eyes categorized according to the ESI index as either normal, keratoconus, and subclinical (FFKC) keratoconus. They included 420 topography, elevation, and pachymetry parameters, and the algorithm produced 4 clusters of eyes with similar characteristics. When comparing their results with a reference standard (ESI), the model did not create a distinct grouping that separated the subclinical eyes from other eyes (sensitivity 88% and specificity 14%), suggesting poor correlation when compared with ESI alone. Furthermore, they did not compare their results with clinically labeled data.

Research Question 3: How Was Algorithm Validation Handled Among the Selected Manuscripts?

Although most studies performed internal validation by splitting the original data set into training and test sets, we identified 5 replication papers that validated a published model on a new data set [48,51,52,56,63]. Ruiz Hidalgo et al [56] verified their SVM technique presented in 2016 [55]. The authors found that when using the Antwerp University Hospital classification, there was approximately 18% decrease in sensitivity, whereas when using the Rothschild classification, there was approximately 21% increase in sensitivity. These discrepancies highlight the problems associated with subjective classification and the absence of ground truth. Furthermore, when multiple groups were included in the analysis; that is, normal, keratoconus, subclinical keratoconus, and postrefractive surgery eyes, it was noted that the accuracy decreased from 93.1% in discriminating normal from FFKC to 88.8%. However, this paper presented the most comprehensive methodology because the authors not only verified their results on a new sample population with multiple target classes but also compared their results with other methods and included 2 subjective reference standards.

Buhren et al [51] validated their model defined in 2010 [84]. When comparing their discriminant function derived from anterior and posterior corneal surface wavefront data, they reported approximately 22% decrease in sensitivity and approximately 9% decrease in specificity. This decrease was likely caused by overfitting in the original study. Saad et al [48] and Chan et al [52] validated the same discriminant analysis model presented by Saad et al [77]. Saad et al [48] reported sensitivity (92%) and specificity (96%), roughly in line with their previous study, which indicates that their method is reliable and does not suffer from overfitting. Chan et al [52] validated the original model in patients from different ethnic backgrounds (Asian). They reported approximately 21% decrease in sensitivity, which they attributed to overfitting in the original study; however, their specificity was almost equivalent. Steinberg et al [63] validated the work presented by Ambrosio et al [58]. They reported approximately 27% decrease in sensitivity and approximately 13% decrease in specificity when applying the same thresholds.

Bias Assessment

In general, patient selection was found to have a high risk of bias (19/26, 73% of studies) because most studies were case-control (thus susceptible to selection bias) and did not use consecutive or random samples. Multimedia Appendix 5 [14,47-71] contains the results of applying the QUADAS-2 tool when considering the risk of bias. The index test was also generally found to have a high risk of bias (21/26, 81% of studies) because of the lack of external validation. As there is no gold standard for subclinical keratoconus diagnosis, we could not assess the bias for the reference standard; therefore, all papers were marked as unclear. Finally, patient flow was found to have a low risk of bias (21/26, 81% of studies) because although chronological information was sparse, the same analysis was usually applied to all patients.

Discussion

Research Question 1: What Input Data Types Have Been Used Within Subclinical Keratoconus Detection Algorithms and How Have They Performed?

The data most commonly used for building subclinical keratoconus detection algorithms are numeric keratometry or pachymetry parameters; hence, according to our review, algorithms based on these tend to have the highest performance. These parameters are derived from a variety of imaging systems and devices and are then incorporated into different combinations to build a classification system or an index. Inevitably, individual systems produce parameters that may not be comparable across devices, and for proprietary reasons, the raw data are generally not available to derive these parameters. Therefore, comparison or replication across systems is difficult. Heat maps provide a visual representation of either corneal elevation, pachymetry, or curvature, which are helpful for the visual interpretation of results. However, heat maps require interpolation or extrapolation of data, which may introduce inaccuracies when included in the model. To the best of our knowledge, there are no studies that have analyzed actual pixel-level corneal imaging data (Scheimpflug or AS-OCT), probably because access to these data is restricted to commercial machines such as the Pentacam, which impedes bulk export to train machine learning algorithms.

We also noted that many studies do not incorporate details of patient demographics and associated diseases, such as age, sex, ethnicity, and atopy, which can influence the risk of developing keratoconus. Incorporating these data into these models may help define the population to which an algorithm applies, particularly as there are phenotypic indices that an algorithm can identify from images that humans cannot identify by manual inspection [85].

Research Question 2: What Machine Learning Algorithms Have Been Used for Subclinical Keratoconus Detection and How Have They Performed?

Subclinical keratoconus studies typically involve univariate or multivariate analyses. For univariate studies, receiver operating characteristic analysis is performed, as each parameter is included to quantify their diagnostic ability. However, because none of the univariate studies we identified performed an out-of-sample validation, they were all excluded. For multivariate studies, machine learning is used to create a detection model using multiple parameters. These algorithms have already demonstrated comparable performance to experienced ophthalmologists in the identification of retinopathy of prematurity [86] and retinal disease progression [87]. Machine learning–based research into the detection of subclinical keratoconus has largely focused on supervised learning techniques, such as decision trees, SVM, logistic regression, discriminant analysis, NNs, and CNNs. Logistic regression may be superior to NNs when parameters from a single imaging modality are considered [14,68], with a potentially greater role for NNs when a large number of potentially interacting parameters are combined, such as for multiple imaging modalities [68]. Unsupervised learning has also been evaluated for the detection of subclinical keratoconus, although it relies on identifying patterns in large amounts of data; hence, it may not translate to a different data set of a different size and with different properties. In addition, with the exception of the study by Yousefi [61], none of the papers provided access to the source code for their algorithms or a description of the hyperparameters, which makes it difficult to reproduce and validate the results with external data sets.

Research Question 3: How Was Algorithm Validation Handled Among the Selected?

We excluded papers that did not include a validation arm for the study, and the vast majority of initially identified studies did not appropriately validate their results. For any type of automatic classifier, validating the results on a data set distinct from the trained set is critical in determining the generalizability of the model to other data sets. With the exception of the studies by Saad et al [48] and Hidalgo et al [56], it is clear that studies attempting to validate a prior method reported significant decreases in sensitivity and specificity in comparison with their original results. This shows that even when techniques such as cross-validation are performed, the best method for validation is an independent out-of-sample data set, and its absence may introduce bias. Ideally, this external data set would be larger and more representative of the general population.

Strengths and Limitations

The primary strength of this study is that we present a comprehensive review of all studies published in English between January 1, 2010, and October 31, 2020, on the use of machine learning for the detection of subclinical keratoconus. Our focus on the detection of subclinical keratoconus addresses an important unmet clinical need for an effective machine-based technique to identify keratoconus at its earliest stage. This would move us closer to potential screening without significant demands on clinicians and clinical services. Subclinical disease diagnosis is more challenging than the detection of advanced disease, where the opportunity to prevent progression has already been lost. In this respect, our review builds on recent clinical trials of CXL to prevent keratoconus progression in children and young adults [15,88,89]. To present a balanced and comprehensive overview, we have combined the expertise of computer scientists (HM and NP) familiar with the development of machine learning for clinical medicine with the input from clinicians (JPOL, DG, and ST) who are experienced in keratoconus management. We have considered and compared the literature in terms of both clinical input data and machine learning methodology, which allows the reader to gain a wider perspective of the problem.

However, there are limitations to our search methods and inclusion criteria. As with any systematic review, articles that did not include the relevant key terms or were not appropriately indexed by the literature databases may have been missed. When considering our inclusion criteria based on subclinical disease, some studies may have been missed because of a lack of consensus on definition. In addition, where there was no form of validation, we excluded the study; thus, our results represent only the articles that have some degree of generalizability.

A further limitation is the difficulty in comparing the performance of the approaches described in the manuscripts; direct comparisons were not possible because of the variation of multiple study design factors such as subclinical disease definition, parameter choice, data set source, and machine learning algorithm. Finally, a limitation regarding case definition that applies to all studies is the uncertainty in the relationship between subclinical keratoconus and other nonprogressive abnormalities of corneal shape.

Challenges and Future Directions

Our systematic review identified several challenges from the literature and avenues for future research.

Case Definition, Gold Standard, and Ground Truth

Precise comparisons between the results of publications are problematic because of the ambiguous definition of early keratoconus and the absence of a gold standard examination technique. The most common definition of subclinical keratoconus is an eye with topographic findings that is at least suspicious of keratoconus and with confirmed keratoconus in the fellow eye. FFKC is usually defined as an eye that has both normal topography and slit-lamp examination but with keratoconus in the fellow eye [44]. With this differentiation, subclinical keratoconus will be easier to detect than FFKC, and studies using the former definition are likely to produce more accurate results because the problem becomes easier to solve. The problems of making statistical comparisons in the absence of a gold standard have been discussed extensively by Umemneku et al [90]. The authors suggest that latent class analysis, composite reference standards, or expert panel analysis may be appropriate in these circumstances.

Even if a precise definition of early subclinical keratoconus was established, the absence of ground truth data is relevant when evaluating the precision of data acquisition. For example, measurements of keratoconus taken by different operators or repeated on different days may lead to variations in the results. Flynn et al [91] found that keratometric measurements from Scheimpflug images (Pentacam) were more reproducible in early keratoconus (mean central K ≤53 D) compared with those in more advanced keratoconus (mean central K>53 D), although a cohort with subclinical keratoconus was not included. In contrast, Yang et al [92] found that biomechanical parameters (Corvis ST) had acceptable repeatability in both normal and keratoconus eyes.

Another issue we identified when comparing studies was the variation in the number of groups that were classified. The studies often started with multiple groups (usually 3, eg, FFKC, keratoconus, and normal); however, 21 papers chose to report their accuracy results from a model trained to classify between just 2 groups (eg, FFKC and normal), whereas 5 papers reported results for classifying between all groups. Classifying all groups is a more realistic clinical scenario, but it presents a more challenging problem because the features of the different groups can overlap. Complete details of the number of groups associated with the accuracy results are presented in Multimedia Appendix 2.

Study Size and Statistical Power

The size of the study is critical when developing a reliable detection system. In particular, the accuracy of machine learning models is directly related to the amount of training data. When considering eyes with subclinical keratoconus, only 2 studies included more than 500 eyes [61,71]. None of the papers included a priori power calculations to estimate the size of the cohort to be studied.

Study Design

None of the reported studies evaluated the performance of their method against masked observers; thus, they may introduce detection bias. The initial classification is often made by considering the fellow eye with keratoconus as a factor in the decision-making process, whereas the algorithm does not have this information. Hence, it would be interesting to design a study where, having already decided on the ground truth diagnosis, a new clinician is asked to evaluate the eye using the same information as the algorithm (ie, only the images or parameters). This situation is closest to real-life screening where a prospective patient (without a history of keratoconus in either eye) is examined for risk of keratoconus.

Subclinical keratoconus is, by definition, the least affected eye of highly asymmetrical keratoconus. An assumption is that any parameters of subclinical disease that differ from the values for normal corneas are the result of keratoconus. However, it has not been demonstrated prospectively that all eyes in such a cohort will progress to the clinical disease state. Although true unilateral keratoconus is thought not to exist [37], this has not been proven, and it is possible that some eyes with subclinical keratoconus are not at risk of progression and that some of the abnormal parameters in this group are not the result of keratoconus. It would be valuable to conduct a prospective study in which eyes that do not develop clinical keratoconus over time are used as lower-risk examples.

External Validation and Generalizability to Real-world Data

To be useful, it is essential that a detection algorithm can generalize beyond the limited data set from which the model was developed and benchmarked, which requires external validation in out-of-sample data sets. The creation of a large open-source data set of keratoconus images could serve as a reference standard to develop a benchmark for external validation. We also recommend that journals adhere to the Transparent Reporting of a Multivariable Prediction Model for Individual Prognosis or Diagnosis guidelines so that all published methods are externally validated. When generalizing to external data sets the source and quality of the data should be considered. Data from a referral hospital may not represent the general population, who might be the target for screening programs, with an underrepresentation of eyes with mild disease.

Other Challenges

There are several other considerations, such as keratoconus progression and the translation of a detection algorithm into a medical device that can be implemented in the real world, but these issues are beyond the scope of this review. Nevertheless, these points are discussed in Multimedia Appendix 6 [37,39,93-102].

New Avenues of Research

On the basis of the results of this review, there is a need for further fundamental research, particularly for analysis based on the raw pixel values of corneal imaging rather than only derived parameters. Furthermore, a multimodal solution could be developed by combining these raw images with other parameters, such as biomechanical, demographic, and genetic data. Demographic data such as age, sex, ethnicity, and allergic eye disease are known risk factors for progressive keratoconus, and a family history of keratoconus is also a risk factor that should also be included in diagnostic algorithms. Environmental risk factors, including eye rubbing, have been associated with keratoconus progression, although eye rubbing is difficult to quantify. A genetic predisposition to keratoconus is supported by heritability studies in twins, linkage analysis in families, and population-level genome-wide association studies [103]. From these studies, genetic risk scores have been derived, which could be included in machine learning models for the detection of subclinical keratoconus. Ideally, a prospective study should be performed in a large cohort of young (<30 years of age) patients with subclinical keratoconus to monitor disease progression. Training should be conducted on large data sets with the explicit aim of detecting subclinical keratoconus, and the resulting model should be externally validated on a new data set. Finally, a range of machine learning techniques should be applied to the same data set along with detailed comparison statistics.

Conclusions

We have conducted the most comprehensive review to date on machine learning algorithms for the detection of subclinical keratoconus. Early detection of keratoconus to enable treatment and prevent sight loss is a public health priority, and the use of machine learning algorithms has the potential to make the diagnostic process more efficient and widely available. We have summarized the relevant publications in terms of their input data and the choice of algorithm and identified whether studies performed appropriate validation. We have identified the challenges of obtaining accurate data sets for training machine learning algorithms and the need for a consistent, objective, and agreed definition of subclinical keratoconus. New avenues of research have been identified that combine multimodal source data with biomechanical, demographic, and genomic data. Defining disease progression and modeling progression to the point where there is sight loss are areas that may benefit from further research. We believe this up-to-date review is important to enable researchers, clinicians, and public health policymakers to understand the current state of the research and provide guidance for future health service planning.

Acknowledgments

HM is funded by Moorfields Eye Charity (GR001147). NP is funded by Moorfields Eye Charity Career Development Award (R190031A). Moorfields Eye Charity is supported in part by the National Institute for Health Research Biomedical Research Centre based at Moorfields Eye Hospital National Health Service Foundation Trust and University College London Institute of Ophthalmology. PL is supported by Progress Q26/LF1 and UNCE 204064. ST and DG acknowledge that a proportion of their financial support is from the Department of Health through the award made by the National Institute for Health Research to Moorfields Eye Hospital National Health Service Foundation Trust and University College London Institute of Ophthalmology for a Specialist Biomedical Research Centre for Ophthalmology.

The views expressed are those of the authors and not necessarily those of the National Health Service, the National Institute for Health Research, or the UK Department of Health and Social Care.

Abbreviations

AS-OCT

anterior segment optical coherence tomography

AUC

area under the receiver operating characteristic curve

BAD-D

Belin/Ambrosio enhanced ectasia display

BFS

best-fit sphere

CNN

convolutional neural network

CXL

corneal collagen cross-linking

ESI

ectasia screening index

FFKC

forme fruste keratoconus

HOA

higher-order aberration

LASIK

laser-assisted in situ keratomileusis

NN

neural network

PRISMA

Preferred Reporting Items for Systematic Reviews and Meta-Analyses

QUADAS

Quality Assessment of Diagnostic Accuracy Studies

SVM

support vector machine

Multimedia Appendix 1

Grading systems and indices.

Multimedia Appendix 2

Supplementary details for the 26 published studies that included the use of machine learning for the detection of subclinical keratoconus.

Multimedia Appendix 3

Sensitivity and specificity were plotted for the 26 published studies that included the use of machine learning for the detection of subclinical keratoconus. In the left-hand chart, the results were grouped by diagnosis criteria. A: clinically normal, topographically abnormal. B: Fellow eye of diagnosed keratoconus, clinically normal, topographically normal. C: Fellow eye of diagnosed keratoconus, clinically normal, topographically abnormal. Although there is no obvious pattern relating to diagnostic criteria, the largest outliers belong to group A, suggesting that using a fellow eye with keratoconus may lead to a better detection system. In the right-hand chart, the results are grouped according to the imaging system. No obvious pattern can be seen in the results, suggesting that the choice of imaging system is unrelated to the detection system accuracy.

Multimedia Appendix 4

Heat maps of an advanced keratoconus eye derived from Scheimpflug corneal imaging using the Pentacam device.

Multimedia Appendix 5

Results of applying the QUADAS (Quality Assessment of Diagnostic Accuracy Studies)-2 bias assessment tool including responses to tailored signaling questions.

Multimedia Appendix 6

Other challenges, such as keratoconus progression and translational considerations, have been identified for the detection of subclinical keratoconus using machine learning.

Footnotes

Conflicts of Interest: None declared.

References

  • 1.Mas Tur V, MacGregor C, Jayaswal R, O'Brart D, Maycock N. A review of keratoconus: diagnosis, pathophysiology, and genetics. Surv Ophthalmol. 2017;62(6):770–83. doi: 10.1016/j.survophthal.2017.06.009.S0039-6257(17)30046-2 [DOI] [PubMed] [Google Scholar]
  • 2.Davidson AE, Hayes S, Hardcastle AJ, Tuft SJ. The pathogenesis of keratoconus. Eye (Lond) 2014 Feb;28(2):189–95. doi: 10.1038/eye.2013.278. http://europepmc.org/abstract/MED/24357835 .eye2013278 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3.Godefrooij DA, de Wit GA, Uiterwaal CS, Imhof SM, Wisse RP. Age-specific incidence and prevalence of keratoconus: a nationwide registration study. Am J Ophthalmol. 2017 Mar;175:169–72. doi: 10.1016/j.ajo.2016.12.015.S0002-9394(16)30613-4 [DOI] [PubMed] [Google Scholar]
  • 4.Chan E, Chong EW, Lingham G, Stevenson LJ, Sanfilippo PG, Hewitt AW, Mackey DA, Yazar S. Prevalence of keratoconus based on Scheimpflug imaging: the Raine study. Ophthalmology. 2021 Apr;128(4):515–21. doi: 10.1016/j.ophtha.2020.08.020.S0161-6420(20)30838-1 [DOI] [PubMed] [Google Scholar]
  • 5.Papaliʼi-Curtin AT, Cox R, Ma T, Woods L, Covello A, Hall RC. Keratoconus prevalence among high school students in New Zealand. Cornea. 2019 Nov;38(11):1382–9. doi: 10.1097/ICO.0000000000002054. [DOI] [PubMed] [Google Scholar]
  • 6.Ferdi AC, Nguyen V, Gore DM, Allan BD, Rozema JJ, Watson SL. Keratoconus natural progression: a systematic review and meta-analysis of 11 529 eyes. Ophthalmology. 2019 Jul;126(7):935–45. doi: 10.1016/j.ophtha.2019.02.029.S0161-6420(18)33287-1 [DOI] [PubMed] [Google Scholar]
  • 7.Tuft SJ, Moodaley LC, Gregory WM, Davison CR, Buckley RJ. Prognostic factors for the progression of keratoconus. Ophthalmology. 1994 Mar;101(3):439–47. doi: 10.1016/s0161-6420(94)31313-3.S0161-6420(94)31313-3 [DOI] [PubMed] [Google Scholar]
  • 8.Pearson AR, Soneji B, Sarvananthan N, Sandford-Smith JH. Does ethnic origin influence the incidence or severity of keratoconus? Eye (Lond) 2000 Aug;14 ( Pt 4):625–8. doi: 10.1038/eye.2000.154. [DOI] [PubMed] [Google Scholar]
  • 9.Downie LE, Lindsay RG. Contact lens management of keratoconus. Clin Exp Optom. 2015 Jul;98(4):299–311. doi: 10.1111/cxo.12300. doi: 10.1111/cxo.12300. [DOI] [PubMed] [Google Scholar]
  • 10.Steinberg J, Bußmann N, Frings A, Katz T, Druchkiv V, Linke SJ. Quality of life in stable and progressive 'early-stage' keratoconus patients. Acta Ophthalmol. 2021 Mar;99(2):e196–201. doi: 10.1111/aos.14564. [DOI] [PubMed] [Google Scholar]
  • 11.Saunier V, Mercier A, Gaboriau T, Malet F, Colin J, Fournié P, Malecaze F, Touboul D. Vision-related quality of life and dependency in French keratoconus patients: impact study. J Cataract Refract Surg. 2017 Dec;43(12):1582–90. doi: 10.1016/j.jcrs.2017.08.024.S0886-3350(17)30760-5 [DOI] [PubMed] [Google Scholar]
  • 12.Gore DM, Watson MP, Tuft SJ. Permanent visual loss in eyes with keratoconus. Acta Ophthalmol. 2014 May;92(3):e244–5. doi: 10.1111/aos.12253. doi: 10.1111/aos.12253. [DOI] [PubMed] [Google Scholar]
  • 13.Kelly T, Williams KA, Coster DJ, Australian Corneal Graft Registry Corneal transplantation for keratoconus: a registry study. Arch Ophthalmol. 2011 Jun;129(6):691–7. doi: 10.1001/archophthalmol.2011.7.archophthalmol.2011.7 [DOI] [PubMed] [Google Scholar]
  • 14.Cao K, Verspoor K, Sahebjada S, Baird PN. Evaluating the performance of various machine learning algorithms to detect subclinical keratoconus. Trans Vis Sci Tech. 2020 Apr 24;9(2):24. doi: 10.1167/tvst.9.2.24. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15.Wittig-Silva C, Chan E, Islam FM, Wu T, Whiting M, Snibson GR. A randomized, controlled trial of corneal collagen cross-linking in progressive keratoconus: three-year results. Ophthalmology. 2014 Apr;121(4):812–21. doi: 10.1016/j.ophtha.2013.10.028.S0161-6420(13)01004-X [DOI] [PubMed] [Google Scholar]
  • 16.Caporossi A, Mazzotta C, Baiocchi S, Caporossi T. Long-term results of riboflavin ultraviolet a corneal collagen cross-linking for keratoconus in Italy: the Siena eye cross study. Am J Ophthalmol. 2010 Apr;149(4):585–93. doi: 10.1016/j.ajo.2009.10.021.S0002-9394(09)00808-3 [DOI] [PubMed] [Google Scholar]
  • 17.O'Brart DP, Chan E, Samaras K, Patel P, Shah SP. A randomised, prospective study to investigate the efficacy of riboflavin/ultraviolet A (370 nm) corneal collagen cross-linkage to halt the progression of keratoconus. Br J Ophthalmol. 2011 Nov;95(11):1519–24. doi: 10.1136/bjo.2010.196493.bjo.2010.196493 [DOI] [PubMed] [Google Scholar]
  • 18.Gore DM, Shortt AJ, Allan BD. New clinical pathways for keratoconus. Eye (Lond) 2013 Mar;27(3):329–39. doi: 10.1038/eye.2012.257. http://europepmc.org/abstract/MED/23258309 .eye2012257 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19.Koller T, Mrochen M, Seiler T. Complication and failure rates after corneal crosslinking. J Cataract Refract Surg. 2009 Aug;35(8):1358–62. doi: 10.1016/j.jcrs.2009.03.035.S0886-3350(09)00481-7 [DOI] [PubMed] [Google Scholar]
  • 20.Gore DM, Leucci MT, Koay S, Kopsachilis N, Nicolae MN, Malandrakis MI, Anand V, Allan BD. Accelerated pulsed high-fluence corneal cross-linking for progressive keratoconus. Am J Ophthalmol. 2021 Jan;221:9–16. doi: 10.1016/j.ajo.2020.08.021.S0002-9394(20)30447-5 [DOI] [PubMed] [Google Scholar]
  • 21.Salmon HA, Chalk D, Stein K, Frost NA. Cost effectiveness of collagen crosslinking for progressive keratoconus in the UK NHS. Eye (Lond) 2015 Nov;29(11):1504–11. doi: 10.1038/eye.2015.151. http://europepmc.org/abstract/MED/26315704 .eye2015151 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22.Lindstrom RL, Berdahl JP, Donnenfeld ED, Thompson V, Kratochvil D, Wong C, Falvey H, Lytle G, Botteman MF, Carter JA. Corneal cross-linking versus conventional management for keratoconus: a lifetime economic model. J Med Econ. 2021;24(1):410–20. doi: 10.1080/13696998.2020.1851556. [DOI] [PubMed] [Google Scholar]
  • 23.Godefrooij DA, Mangen MJ, Chan E, O'Brart DP, Imhof SM, de Wit GA, Wisse RP. Cost-effectiveness analysis of corneal collagen crosslinking for progressive keratoconus. Ophthalmology. 2017 Oct;124(10):1485–95. doi: 10.1016/j.ophtha.2017.04.011.S0161-6420(16)32228-X [DOI] [PubMed] [Google Scholar]
  • 24.Kreps EO, Claerhout I, Koppen C. Diagnostic patterns in keratoconus. Cont Lens Anterior Eye. 2021 Jun;44(3):101333. doi: 10.1016/j.clae.2020.05.002.S1367-0484(20)30103-X [DOI] [PubMed] [Google Scholar]
  • 25.Lee A, Taylor P, Kalpathy-Cramer J, Tufail A. Machine learning has arrived! Ophthalmology. 2017 Dec;124(12):1726–8. doi: 10.1016/j.ophtha.2017.08.046.S0161-6420(17)31563-4 [DOI] [PubMed] [Google Scholar]
  • 26.Hastie T, Tibshirani R, Friedman J. The Elements of Statistical Learning: Data Mining, Inference, and Prediction. USA: Springer; 2009. [Google Scholar]
  • 27.Russell S, Norvig P. Artificial Intelligence: A Modern Approach. USA: Prentice Hall; 2020. [Google Scholar]
  • 28.Tong Y, Lu W, Yu Y, Shen Y. Application of machine learning in ophthalmic imaging modalities. Eye Vis (Lond) 2020;7:22. doi: 10.1186/s40662-020-00183-6. https://eandv.biomedcentral.com/articles/10.1186/s40662-020-00183-6 .183 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 29.Lin SR, Ladas JG, Bahadur GG, Al-Hashimi S, Pineda R. A review of machine learning techniques for keratoconus detection and refractive surgery screening. Semin Ophthalmol. 2019;34(4):317–26. doi: 10.1080/08820538.2019.1620812. [DOI] [PubMed] [Google Scholar]
  • 30.Bishop C. Pattern Recognition and Machine Learning (Information Science and Statistics) USA: Springer-Verlag; 2007. [Google Scholar]
  • 31.Goodfellow I, Bengio Y, Courville A. Deep Learning (Adaptive Computation and Machine Learning Series) USA: MIT Press; 2017. p. 800. [Google Scholar]
  • 32.Ting DS, Pasquale LR, Peng L, Campbell JP, Lee AY, Raman R, Tan GS, Schmetterer L, Keane PA, Wong TY. Artificial intelligence and deep learning in ophthalmology. Br J Ophthalmol. 2019 Feb;103(2):167–75. doi: 10.1136/bjophthalmol-2018-313173. http://bjo.bmj.com/lookup/pmidlookup?view=long&pmid=30361278 .bjophthalmol-2018-313173 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 33.Korot E, Guan Z, Ferraz D, Wagner SK, Zhang G, Liu X, Faes L, Pontikos N, Finlayson SG, Khalid H, Moraes G, Balaskas K, Denniston AK, Keane PA. Code-free deep learning for multi-modality medical image classification. Nat Mach Intell. 2021 Mar 01;3(4):288–98. doi: 10.1038/s42256-021-00305-2. [DOI] [Google Scholar]
  • 34.Wang Y, Zhao Y, Therneau TM, Atkinson EJ, Tafti AP, Zhang N, Amin S, Limper AH, Khosla S, Liu H. Unsupervised machine learning for the discovery of latent disease clusters and patient subgroups using electronic health records. J Biomed Inform. 2020 Feb;102:103364. doi: 10.1016/j.jbi.2019.103364. https://linkinghub.elsevier.com/retrieve/pii/S1532-0464(19)30284-9 .S1532-0464(19)30284-9 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 35.Pontikos N. Normalisation and Clustering Methods Applied to Association Studies in Type 1 Diabetes. UK: Cambridge University; 2015. [Google Scholar]
  • 36.- Grading diabetic retinopathy from stereoscopic color fundus photographs--an extension of the modified Airlie House classification. ETDRS report number 10. Early Treatment Diabetic Retinopathy Study Research Group. Ophthalmology. 1991 May;98(5 Suppl):786–806.S0161-6420(13)38012-9 [PubMed] [Google Scholar]
  • 37.Gomes JA, Tan D, Rapuano CJ, Belin MW, Ambrósio R, Guell JL, Malecaze F, Nishida K, Sangwan VS, Group of Panelists for the Global Delphi Panel of KeratoconusEctatic Diseases Global consensus on keratoconus and ectatic diseases. Cornea. 2015 Apr;34(4):359–69. doi: 10.1097/ICO.0000000000000408.00003226-201504000-00001 [DOI] [PubMed] [Google Scholar]
  • 38.Amsler M. Kératocône classique et kératocône fruste; arguments unitaires. Ophthalmologica. 1946;111(2-3):96–101. doi: 10.1159/000300309. [DOI] [PubMed] [Google Scholar]
  • 39.Belin MW, Duncan JK. Keratoconus: the ABCD grading system. Klin Monbl Augenheilkd. 2016 Jun;233(6):701–7. doi: 10.1055/s-0042-100626. [DOI] [PubMed] [Google Scholar]
  • 40.Independent population validation of the belinambrosio enhanced ectasia display implications for keratoconus studies and screening. Enhanced Screening for Ectasia Detection based on Scheimpflug Tomography and Biomachanical evaluations. 2014. [2021-11-22]. https://tinyurl.com/5yywh3fm .
  • 41.Hashemi H, Beiranvand A, Yekta A, Maleki A, Yazdani N, Khabazkhoob M. Pentacam top indices for diagnosing subclinical and definite keratoconus. J Curr Ophthalmol. 2016 Mar;28(1):21–6. doi: 10.1016/j.joco.2016.01.009. https://linkinghub.elsevier.com/retrieve/pii/S2452-2325(15)30016-0 .S2452-2325(15)30016-0 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 42.Pentacam®. OCULUS. [2021-11-22]. https://www.pentacam.com/int/opticianoptometrist-without-pentacamr/models/pentacamr/core-functions.html .
  • 43.Ocular Response Analyzer® G3. Reichert Technologies. [2021-11-22]. https://www.reichert.com/products/ocular-response-analyzer-g3 .
  • 44.Henriquez MA, Hadid M, Izquierdo L. A systematic review of subclinical keratoconus and forme fruste keratoconus. J Refract Surg. 2020 Apr 01;36(4):270–9. doi: 10.3928/1081597X-20200212-03. [DOI] [PubMed] [Google Scholar]
  • 45.Moher D, Liberati A, Tetzlaff J, Altman DG, PRISMA Group Preferred reporting items for systematic reviews and meta-analyses: the PRISMA statement. PLoS Med. 2009 Jul 21;6(7):e1000097. doi: 10.1371/journal.pmed.1000097. https://dx.plos.org/10.1371/journal.pmed.1000097 . [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 46.Krumeich JH, Daniel J, Knülle A. Live-epikeratophakia for keratoconus. J Cataract Refract Surg. 1998 Apr;24(4):456–63. doi: 10.1016/s0886-3350(98)80284-8.S0886-3350(98)80284-8 [DOI] [PubMed] [Google Scholar]
  • 47.Arbelaez MC, Versaci F, Vestri G, Barboni P, Savini G. Use of a support vector machine for keratoconus and subclinical keratoconus detection by topographic and tomographic data. Ophthalmology. 2012 Nov;119(11):2231–8. doi: 10.1016/j.ophtha.2012.06.005.S0161-6420(12)00513-1 [DOI] [PubMed] [Google Scholar]
  • 48.Saad A, Gatinel D. Validation of a new scoring system for the detection of early forme of keratoconus. Int J Keratoconus Ectatic Corneal Dis. 2012 May;1(2):100–8. doi: 10.5005/JP-JOURNALS-10025-1019. [DOI] [Google Scholar]
  • 49.Smadja D, Touboul D, Cohen A, Doveh E, Santhiago MR, Mello GR, Krueger RR, Colin J. Detection of subclinical keratoconus using an automated decision tree classification. Am J Ophthalmol. 2013 Aug;156(2):237–46.e1. doi: 10.1016/j.ajo.2013.03.034.S0002-9394(13)00222-5 [DOI] [PubMed] [Google Scholar]
  • 50.Ramos-López D, Martínez-Finkelshtein A, Castro-Luna GM, Burguera-Gimenez N, Vega-Estrada A, Piñero D, Alió JL. Screening subclinical keratoconus with placido-based corneal indices. Optom Vis Sci. 2013 Apr;90(4):335–43. doi: 10.1097/OPX.0b013e3182843f2a. [DOI] [PubMed] [Google Scholar]
  • 51.Bühren J, Schäffeler T, Kohnen T. Validation of metrics for the detection of subclinical keratoconus in a new patient collective. J Cataract Refract Surg. 2014 Feb;40(2):259–68. doi: 10.1016/j.jcrs.2013.07.044.S0886-3350(13)01387-4 [DOI] [PubMed] [Google Scholar]
  • 52.Chan C, Ang M, Saad A, Chua D, Mejia M, Lim L, Gatinel D. Validation of an objective scoring system for forme fruste keratoconus detection and post-lasik ectasia risk assessment in asian eyes. Cornea. 2015 Sep;34(9):996–1004. doi: 10.1097/ICO.0000000000000529. [DOI] [PubMed] [Google Scholar]
  • 53.Kovács I, Miháltz K, Kránitz K, Juhász É, Takács Á, Dienes L, Gergely R, Nagy ZZ. Accuracy of machine learning classifiers using bilateral data from a Scheimpflug camera for identifying eyes with preclinical signs of keratoconus. J Cataract Refract Surg. 2016 Feb;42(2):275–83. doi: 10.1016/j.jcrs.2015.09.020.S0886-3350(15)01368-1 [DOI] [PubMed] [Google Scholar]
  • 54.Saad A, Gatinel D. Combining placido and corneal wavefront data for the detection of forme fruste keratoconus. J Refract Surg. 2016 Aug 01;32(8):510–6. doi: 10.3928/1081597X-20160523-01. [DOI] [PubMed] [Google Scholar]
  • 55.Ruiz Hidalgo I, Rodriguez P, Rozema JJ, Ní Dhubhghaill S, Zakaria N, Tassignon M, Koppen C. Evaluation of a machine-learning classifier for keratoconus detection based on scheimpflug tomography. Cornea. 2016 Jun;35(6):827–32. doi: 10.1097/ICO.0000000000000834. [DOI] [PubMed] [Google Scholar]
  • 56.Ruiz Hidalgo I, Rozema JJ, Saad A, Gatinel D, Rodriguez P, Zakaria N, Koppen C. Validation of an objective keratoconus detection system implemented in a scheimpflug tomographer and comparison with other methods. Cornea. 2017 Jun;36(6):689–95. doi: 10.1097/ICO.0000000000001194. [DOI] [PubMed] [Google Scholar]
  • 57.Xu Z, Li W, Jiang J, Zhuang X, Chen W, Peng M, Wang J, Lu F, Shen M, Wang Y. Characteristic of entire corneal topography and tomography for the detection of sub-clinical keratoconus with Zernike polynomials using Pentacam. Sci Rep. 2017 Nov 28;7(1):16486. doi: 10.1038/s41598-017-16568-y. doi: 10.1038/s41598-017-16568-y.10.1038/s41598-017-16568-y [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 58.Ambrósio R, Lopes BT, Faria-Correia F, Salomão MQ, Bühren J, Roberts CJ, Elsheikh A, Vinciguerra R, Vinciguerra P. Integration of scheimpflug-based corneal tomography and biomechanical assessments for enhancing ectasia detection. J Refract Surg. 2017 Jul 01;33(7):434–43. doi: 10.3928/1081597X-20170426-02. https://journals.healio.com/doi/abs/10.3928/1081597X-20170426-02?url_ver=Z39.88-2003&rfr_id=ori:rid:crossref.org&rfr_dat=cr_pub%3dpubmed . [DOI] [PubMed] [Google Scholar]
  • 59.Sideroudi H, Labiris G, Georgantzoglou K, Ntonti P, Siganos C, Kozobolis V. Fourier analysis algorithm for the posterior corneal keratometric data: clinical usefulness in keratoconus. Ophthalmic Physiol Opt. 2017 Jul;37(4):460–6. doi: 10.1111/opo.12386. [DOI] [PubMed] [Google Scholar]
  • 60.Francis M, Pahuja N, Shroff R, Gowda R, Matalia H, Shetty R, Remington Nelson EJ, Sinha Roy A. Waveform analysis of deformation amplitude and deflection amplitude in normal, suspect, and keratoconic eyes. J Cataract Refract Surg. 2017 Oct;43(10):1271–80. doi: 10.1016/j.jcrs.2017.10.012.S0886-3350(17)30642-9 [DOI] [PubMed] [Google Scholar]
  • 61.Yousefi S, Yousefi E, Takahashi H, Hayashi T, Tampo H, Inoda S, Arai Y, Asbell P. Keratoconus severity identification using unsupervised machine learning. PLoS One. 2018 Nov 6;13(11):e0205998. doi: 10.1371/journal.pone.0205998. https://dx.plos.org/10.1371/journal.pone.0205998 .PONE-D-18-15784 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 62.Lopes BT, Ramos IC, Salomão MQ, Guerra FP, Schallhorn SC, Schallhorn JM, Vinciguerra R, Vinciguerra P, Price FW, Price MO, Reinstein DZ, Archer TJ, Belin MW, Machado AP, Ambrósio R. Enhanced tomographic assessment to detect corneal ectasia based on artificial intelligence. Am J Ophthalmol. 2018 Nov;195:223–32. doi: 10.1016/j.ajo.2018.08.005.S0002-9394(18)30445-8 [DOI] [PubMed] [Google Scholar]
  • 63.Steinberg J, Siebert M, Katz T, Frings A, Mehlan J, Druchkiv V, Bühren J, Linke SJ. Tomographic and biomechanical scheimpflug imaging for keratoconus characterization: a validation of current indices. J Refract Surg. 2018 Dec 01;34(12):840–7. doi: 10.3928/1081597X-20181012-01. [DOI] [PubMed] [Google Scholar]
  • 64.Issarti I, Consejo A, Jiménez-García M, Hershko S, Koppen C, Rozema JJ. Computer aided diagnosis for suspect keratoconus detection. Comput Biol Med. 2019 Jun;109:33–42. doi: 10.1016/j.compbiomed.2019.04.024.S0010-4825(19)30132-5 [DOI] [PubMed] [Google Scholar]
  • 65.Chandapura R, Salomão MQ, Ambrósio R, Swarup R, Shetty R, Sinha Roy A. Bowman's topography for improved detection of early ectasia. J Biophotonics. 2019 Oct;12(10):e201900126. doi: 10.1002/jbio.201900126. [DOI] [PubMed] [Google Scholar]
  • 66.Xie Y, Zhao L, Yang X, Wu X, Yang Y, Huang X, Liu F, Xu J, Lin L, Lin H, Feng Q, Lin H, Liu Q. Screening candidates for refractive surgery with corneal tomographic-based deep learning. JAMA Ophthalmol. 2020 May 01;138(5):519–26. doi: 10.1001/jamaophthalmol.2020.0507. http://europepmc.org/abstract/MED/32215587 .2763363 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 67.Kuo B, Chang W, Liao T, Liu F, Liu H, Chu H, Chen W, Hu F, Yen J, Wang I. Keratoconus screening based on deep learning approach of corneal topography. Transl Vis Sci Technol. 2020 Sep 25;9(2):53. doi: 10.1167/tvst.9.2.53. https://tvst.arvojournals.org/article.aspx?doi=10.1167/tvst.9.2.53 .TVST-20-2403 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 68.Shi C, Wang M, Zhu T, Zhang Y, Ye Y, Jiang J, Chen S, Lu F, Shen M. Machine learning helps improve diagnostic ability of subclinical keratoconus using Scheimpflug and OCT imaging modalities. Eye Vis (Lond) 2020 Sep 10;7:48. doi: 10.1186/s40662-020-00213-3. https://eandv.biomedcentral.com/articles/10.1186/s40662-020-00213-3 .213 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 69.Toprak I, Cavas F, Velázquez JS, Alio Del Barrio JL, Alio JL. Subclinical keratoconus detection with three-dimensional (3-D) morphogeometric and volumetric analysis. Acta Ophthalmol. 2020 Dec;98(8):e933–42. doi: 10.1111/aos.14433. [DOI] [PubMed] [Google Scholar]
  • 70.Issarti I, Consejo A, Jiménez-García M, Kreps EO, Koppen C, Rozema JJ. Logistic index for keratoconus detection and severity scoring (Logik) Comput Biol Med. 2020 Jul;122:103809. doi: 10.1016/j.compbiomed.2020.103809.S0010-4825(20)30176-1 [DOI] [PubMed] [Google Scholar]
  • 71.Lavric A, Popa V, Takahashi H, Yousefi S. Detecting keratoconus from corneal imaging data using machine learning. IEEE Access. 2020 Aug 12;8:149113–21. doi: 10.1109/access.2020.3016060. [DOI] [Google Scholar]
  • 72.Parikh R, Mathai A, Parikh S, Chandra Sekhar G, Thomas R. Understanding and using sensitivity, specificity and predictive values. Indian J Ophthalmol. 2008;56(1):45–50. doi: 10.4103/0301-4738.37595. http://www.ijo.in/article.asp?issn=0301-4738;year=2008;volume=56;issue=1;spage=45;epage=50;aulast=Parikh . [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 73.Whiting PF, Rutjes AW, Westwood ME, Mallett S, Deeks JJ, Reitsma JB, Leeflang MM, Sterne JA, Bossuyt PM, QUADAS-2 Group QUADAS-2: a revised tool for the quality assessment of diagnostic accuracy studies. Ann Intern Med. 2011 Oct 18;155(8):529–36. doi: 10.7326/0003-4819-155-8-201110180-00009. https://www.acpjournals.org/doi/abs/10.7326/0003-4819-155-8-201110180-00009?url_ver=Z39.88-2003&rfr_id=ori:rid:crossref.org&rfr_dat=cr_pub%3dpubmed .155/8/529 [DOI] [PubMed] [Google Scholar]
  • 74.Lawless MA, Hodge C. Wavefront's role in corneal refractive surgery. Clin Exp Ophthalmol. 2005 Apr;33(2):199–209. doi: 10.1111/j.1442-9071.2005.00994.x.CEO994 [DOI] [PubMed] [Google Scholar]
  • 75.Toprak I, Vega A, Alió Del Barrio JL, Espla E, Cavas F, Alió JL. Diagnostic value of corneal epithelial and stromal thickness distribution profiles in forme fruste keratoconus and subclinical keratoconus. Cornea. 2021 Jan;40(1):61–72. doi: 10.1097/ICO.0000000000002435.00003226-202101000-00011 [DOI] [PubMed] [Google Scholar]
  • 76.Belin MW, Khachikian SS. Keratoconus / ectasia detection with the oculus pentacam: belin / ambrósio enhanced ectasia display. Highlights Ophthalmol. 2008;35(6):5–12. doi: 10.5005/jp/books/11830_8. [DOI] [Google Scholar]
  • 77.Saad A, Gatinel D. Topographic and tomographic properties of forme fruste keratoconus corneas. Invest Ophthalmol Vis Sci. 2010 Nov;51(11):5546–55. doi: 10.1167/iovs.10-5369.iovs.10-5369 [DOI] [PubMed] [Google Scholar]
  • 78.Rozema JJ, Rodriguez P, Ruiz Hidalgo I, Navarro R, Tassignon M, Koppen C. SyntEyes KTC: higher order statistical eye model for developing keratoconus. Ophthalmic Physiol Opt. 2017 May;37(3):358–65. doi: 10.1111/opo.12369. [DOI] [PubMed] [Google Scholar]
  • 79.Hashemi H, Beiranvand A, Yekta A, Asharlous A, Khabazkhoob M. Biomechanical properties of early keratoconus: suppressed deformation signal wave. Cont Lens Anterior Eye. 2017 Apr;40(2):104–8. doi: 10.1016/j.clae.2016.12.004.S1367-0484(16)30198-9 [DOI] [PubMed] [Google Scholar]
  • 80.Galletti JD, Ruiseñor Vázquez PR, Fuentes Bonthoux F, Pförtner T, Galletti JG. Multivariate analysis of the ocular response analyzer's corneal deformation response curve for early keratoconus detection. J Ophthalmol. 2015;2015:496382. doi: 10.1155/2015/496382. doi: 10.1155/2015/496382. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 81.Barsam A, Petrushkin H, Brennan N, Bunce C, Xing W, Foot B, Tuft S. Acute corneal hydrops in keratoconus: a national prospective study of incidence and management. Eye (Lond) 2015 Apr;29(4):469–74. doi: 10.1038/eye.2014.333. http://europepmc.org/abstract/MED/25592120 .eye2014333 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 82.Li Z, Liu F, Yang W, Peng S, Zhou J. A survey of convolutional neural networks: analysis, applications, and prospects. IEEE Trans Neural Netw Learn Syst. 2021 Jun 10;PP doi: 10.1109/TNNLS.2021.3084827. (forthcoming) [DOI] [PubMed] [Google Scholar]
  • 83.Breiman L. Random forests. Mach Learn. 2001 Oct;45:5–32. doi: 10.1023/A:1010933404324. [DOI] [Google Scholar]
  • 84.Bühren J, Kook D, Yoon G, Kohnen T. Detection of subclinical keratoconus by using corneal anterior and posterior surface aberrations and thickness spatial profiles. Invest Ophthalmol Vis Sci. 2010 Jul;51(7):3424–32. doi: 10.1167/iovs.09-4960.iovs.09-4960 [DOI] [PubMed] [Google Scholar]
  • 85.Poplin R, Varadarajan AV, Blumer K, Liu Y, McConnell MV, Corrado GS, Peng L, Webster DR. Prediction of cardiovascular risk factors from retinal fundus photographs via deep learning. Nat Biomed Eng. 2018 Mar;2(3):158–64. doi: 10.1038/s41551-018-0195-0.10.1038/s41551-018-0195-0 [DOI] [PubMed] [Google Scholar]
  • 86.Scruggs BA, Chan RV, Kalpathy-Cramer J, Chiang MF, Campbell JP. Artificial intelligence in retinopathy of prematurity diagnosis. Transl Vis Sci Technol. 2020 Feb 10;9(2):5. doi: 10.1167/tvst.9.2.5. https://tvst.arvojournals.org/article.aspx?doi=10.1167/tvst.9.2.5 .TVST-19-2010 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 87.Yim J, Chopra R, Spitz T, Winkens J, Obika A, Kelly C, Askham H, Lukic M, Huemer J, Fasler K, Moraes G, Meyer C, Wilson M, Dixon J, Hughes C, Rees G, Khaw PT, Karthikesalingam A, King D, Hassabis D, Suleyman M, Back T, Ledsam JR, Keane PA, De Fauw J. Predicting conversion to wet age-related macular degeneration using deep learning. Nat Med. 2020 Jun;26(6):892–9. doi: 10.1038/s41591-020-0867-7.10.1038/s41591-020-0867-7 [DOI] [PubMed] [Google Scholar]
  • 88.Chowdhury K, Dore C, Burr JM, Bunce C, Raynor M, Edwards M, Larkin DF. A randomised, controlled, observer-masked trial of corneal cross-linking for progressive keratoconus in children: the KERALINK protocol. BMJ Open. 2019 Sep 12;9(9):e028761. doi: 10.1136/bmjopen-2018-028761. https://bmjopen.bmj.com/lookup/pmidlookup?view=long&pmid=31515418 .bmjopen-2018-028761 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 89.Larkin DF, Chowdhury K, Burr JM, Raynor M, Edwards M, Tuft SJ, Bunce C, Caverly E, Doré C, KERALINK Trial Study Group Effect of corneal cross-linking versus standard care on keratoconus progression in young patients: the KERALINK randomized controlled trial. Ophthalmology. 2021 Nov;128(11):1516–26. doi: 10.1016/j.ophtha.2021.04.019. https://linkinghub.elsevier.com/retrieve/pii/S0161-6420(21)00297-9 .S0161-6420(21)00297-9 [DOI] [PubMed] [Google Scholar]
  • 90.Umemneku Chikere CM, Wilson K, Graziadio S, Vale L, Allen AJ. Diagnostic test evaluation methodology: a systematic review of methods employed to evaluate diagnostic tests in the absence of gold standard - An update. PLoS One. 2019 Oct;14(10):e0223832. doi: 10.1371/journal.pone.0223832. https://dx.plos.org/10.1371/journal.pone.0223832 .PONE-D-19-15721 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 91.Flynn TH, Sharma DP, Bunce C, Wilkins MR. Differential precision of corneal Pentacam HR measurements in early and advanced keratoconus. Br J Ophthalmol. 2016 Sep;100(9):1183–7. doi: 10.1136/bjophthalmol-2015-307201.bjophthalmol-2015-307201 [DOI] [PubMed] [Google Scholar]
  • 92.Yang K, Xu L, Fan Q, Zhao D, Ren S. Repeatability and comparison of new Corvis ST parameters in normal and keratoconus eyes. Sci Rep. 2019 Oct 25;9(1):15379. doi: 10.1038/s41598-019-51502-4. doi: 10.1038/s41598-019-51502-4.10.1038/s41598-019-51502-4 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 93.Ahn SJ, Kim MK, Wee WR. Topographic progression of keratoconus in the Korean population. Korean J Ophthalmol. 2013 Jun;27(3):162–6. doi: 10.3341/kjo.2013.27.3.162. https://ekjo.org/DOIx.php?id=10.3341/kjo.2013.27.3.162 . [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 94.Duncan JK, Belin MW, Borgstrom M. Assessing progression of keratoconus: novel tomographic determinants. Eye Vis (Lond) 2016 Mar 11;3:6. doi: 10.1186/s40662-016-0038-6. https://eandv.biomedcentral.com/articles/10.1186/s40662-016-0038-6 .38 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 95.Meyer JJ, Gokul A, Vellara HR, Prime Z, McGhee CN. Repeatability and Agreement of Orbscan II, Pentacam HR, and Galilei Tomography Systems in Corneas with keratoconus. Am J Ophthalmol. 2017 Mar;175:122–8. doi: 10.1016/j.ajo.2016.12.003.S0002-9394(16)30599-2 [DOI] [PubMed] [Google Scholar]
  • 96.Guilbert E, Saad A, Elluard M, Grise-Dulac A, Rouger H, Gatinel D. Repeatability of keratometry measurements obtained with three topographers in keratoconic and normal corneas. J Refract Surg. 2016 Mar;32(3):187–92. doi: 10.3928/1081597X-20160113-01. [DOI] [PubMed] [Google Scholar]
  • 97.Hashemi H, Yekta A, Khabazkhoob M. Effect of keratoconus grades on repeatability of keratometry readings: comparison of 5 devices. J Cataract Refract Surg. 2015 May;41(5):1065–72. doi: 10.1016/j.jcrs.2014.08.043.S0886-3350(15)00454-X [DOI] [PubMed] [Google Scholar]
  • 98.Kanellopoulos AJ, Moustou V, Asimellis G. Evaluation of visual acuity, pachymetry and anterior-surface irregularity in keratoconus and crosslinking intervention follow-up in 737 cases. Int J Kerat Ect Cor Dis. 2013:95–103. doi: 10.5005/jp-journals-10025-1060. [DOI] [Google Scholar]
  • 99.Kanellopoulos AJ, Asimellis G. Revisiting keratoconus diagnosis and progression classification based on evaluation of corneal asymmetry indices, derived from Scheimpflug imaging in keratoconic and suspect cases. Clin Ophthalmol. 2013;7:1539–48. doi: 10.2147/OPTH.S44741. doi: 10.2147/OPTH.S44741.opth-7-1539 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 100.Piñero DP, Alio JL, Tomás J, Maldonado MJ, Teus MA, Barraquer RI. Vector analysis of evolutive corneal astigmatic changes in keratoconus. Invest Ophthalmol Vis Sci. 2011 Jun 08;52(7):4054–62. doi: 10.1167/iovs.10-6856.iovs.10-6856 [DOI] [PubMed] [Google Scholar]
  • 101.Yousefi S, Kiwaki T, Zheng Y, Sugiura H, Asaoka R, Murata H, Lemij H, Yamanishi K. Detection of longitudinal visual field progression in glaucoma using machine learning. Am J Ophthalmol. 2018 Sep;193:71–9. doi: 10.1016/j.ajo.2018.06.007.S0002-9394(18)30271-X [DOI] [PubMed] [Google Scholar]
  • 102.Belghith A, Bowd C, Medeiros FA, Balasubramanian M, Weinreb RN, Zangwill LM. Glaucoma progression detection using nonlocal Markov random field prior. J Med Imaging (Bellingham) 2014 Oct;1(3):034504. doi: 10.1117/1.JMI.1.3.034504. http://europepmc.org/abstract/MED/26158069 .14103PR [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 103.Hardcastle AJ, Liskova P, Bykhovskaya Y, McComish BJ, Davidson AE, Inglehearn CF, Li X, Choquet H, Habeeb M, Lucas SE, Sahebjada S, Pontikos N, Lopez KE, Khawaja AP, Ali M, Dudakova L, Skalicka P, Van Dooren BT, Geerards AJ, Haudum CW, Faro VL, Tenen A, Simcoe MJ, Patasova K, Yarrand D, Yin J, Siddiqui S, Rice A, Farraj LA, Chen YI, Rahi JS, Krauss RM, Theusch E, Charlesworth JC, Szczotka-Flynn L, Toomes C, Meester-Smoor MA, Richardson AJ, Mitchell PA, Taylor KD, Melles RB, Aldave AJ, Mills RA, Cao K, Chan E, Daniell MD, Wang JJ, Rotter JI, Hewitt AW, MacGregor S, Klaver CC, Ramdas WD, Craig JE, Iyengar SK, O'Brart D, Jorgenson E, Baird PN, Rabinowitz YS, Burdon KP, Hammond CJ, Tuft SJ, Hysi PG. A multi-ethnic genome-wide association study implicates collagen matrix integrity and cell differentiation pathways in keratoconus. Communications Biology. 2021 Mar 01;4(1):266. doi: 10.1038/s42003-021-01784-0. doi: 10.1038/s42003-021-01784-0.10.1038/s42003-021-01784-0 [DOI] [PMC free article] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Multimedia Appendix 1

Grading systems and indices.

Multimedia Appendix 2

Supplementary details for the 26 published studies that included the use of machine learning for the detection of subclinical keratoconus.

Multimedia Appendix 3

Sensitivity and specificity were plotted for the 26 published studies that included the use of machine learning for the detection of subclinical keratoconus. In the left-hand chart, the results were grouped by diagnosis criteria. A: clinically normal, topographically abnormal. B: Fellow eye of diagnosed keratoconus, clinically normal, topographically normal. C: Fellow eye of diagnosed keratoconus, clinically normal, topographically abnormal. Although there is no obvious pattern relating to diagnostic criteria, the largest outliers belong to group A, suggesting that using a fellow eye with keratoconus may lead to a better detection system. In the right-hand chart, the results are grouped according to the imaging system. No obvious pattern can be seen in the results, suggesting that the choice of imaging system is unrelated to the detection system accuracy.

Multimedia Appendix 4

Heat maps of an advanced keratoconus eye derived from Scheimpflug corneal imaging using the Pentacam device.

Multimedia Appendix 5

Results of applying the QUADAS (Quality Assessment of Diagnostic Accuracy Studies)-2 bias assessment tool including responses to tailored signaling questions.

Multimedia Appendix 6

Other challenges, such as keratoconus progression and translational considerations, have been identified for the detection of subclinical keratoconus using machine learning.


Articles from JMIR Medical Informatics are provided here courtesy of JMIR Publications Inc.

RESOURCES