Skip to main content
Scientific Reports logoLink to Scientific Reports
. 2020 Mar 12;10:4623. doi: 10.1038/s41598-020-61519-9

Effects of Hypertension, Diabetes, and Smoking on Age and Sex Prediction from Retinal Fundus Images

Yong Dae Kim 1,2,#, Kyoung Jin Noh 1,#, Seong Jun Byun 1,#, Soochahn Lee 3, Tackeun Kim 4, Leonard Sunwoo 5, Kyong Joon Lee 5, Si-Hyuck Kang 6, Kyu Hyung Park 1, Sang Jun Park 1,
PMCID: PMC7067849  PMID: 32165702

Abstract

Retinal fundus images are used to detect organ damage from vascular diseases (e.g. diabetes mellitus and hypertension) and screen ocular diseases. We aimed to assess convolutional neural network (CNN) models that predict age and sex from retinal fundus images in normal participants and in participants with underlying systemic vascular-altered status. In addition, we also tried to investigate clues regarding differences between normal ageing and vascular pathologic changes using the CNN models. In this study, we developed CNN age and sex prediction models using 219,302 fundus images from normal participants without hypertension, diabetes mellitus (DM), and any smoking history. The trained models were assessed in four test-sets with 24,366 images from normal participants, 40,659 images from hypertension participants, 14,189 images from DM participants, and 113,510 images from smokers. The CNN model accurately predicted age in normal participants; the correlation between predicted age and chronologic age was R2 = 0.92, and the mean absolute error (MAE) was 3.06 years. MAEs in test-sets with hypertension (3.46 years), DM (3.55 years), and smoking (2.65 years) were similar to that of normal participants; however, R2 values were relatively low (hypertension, R2 = 0.74; DM, R2 = 0.75; smoking, R2 = 0.86). In subgroups with participants over 60 years, the MAEs increased to above 4.0 years and the accuracies declined for all test-sets. Fundus-predicted sex demonstrated acceptable accuracy (area under curve > 0.96) in all test-sets. Retinal fundus images from participants with underlying vascular-altered conditions (hypertension, DM, or smoking) indicated similar MAEs and low coefficients of determination (R2) between the predicted age and chronologic age, thus suggesting that the ageing process and pathologic vascular changes exhibit different features. Our models demonstrate the most improved performance yet and provided clues to the relationship and difference between ageing and pathologic changes from underlying systemic vascular conditions. In the process of fundus change, systemic vascular diseases are thought to have a different effect from ageing. Research in context. Evidence before this study. The human retina and optic disc continuously change with ageing, and they share physiologic or pathologic characteristics with brain and systemic vascular status. As retinal fundus images provide high-resolution in-vivo images of retinal vessels and parenchyma without any invasive procedure, it has been used to screen ocular diseases and has attracted significant attention as a predictive biomarker for cerebral and systemic vascular diseases. Recently, deep neural networks have revolutionised the field of medical image analysis including retinal fundus images and shown reliable results in predicting age, sex, and presence of cardiovascular diseases. Added value of this study. This is the first study demonstrating how a convolutional neural network (CNN) trained using retinal fundus images from normal participants measures the age of participants with underlying vascular conditions such as hypertension, diabetes mellitus (DM), or history of smoking using a large database, SBRIA, which contains 412,026 retinal fundus images from 155,449 participants. Our results indicated that the model accurately predicted age in normal participants, while correlations (coefficient of determination, R2) in test-sets with hypertension, DM, and smoking were relatively low. Additionally, a subgroup analysis indicated that mean absolute errors (MAEs) increased and accuracies declined significantly in subgroups with participants over 60 years of age in both normal participants and participants with vascular-altered conditions. These results suggest that pathologic retinal vascular changes occurring in systemic vascular diseases are different form the changes in spontaneous ageing process, and the ageing process observed in retinal fundus images may saturate at age about 60 years. Implications of all available evidence. Based on this study and previous reports, the CNN could accurately and reliably predict age and sex using retinal fundus images. The fact that retinal changes caused by ageing and systemic vascular diseases occur differently motivates one to understand the retina deeper. Deep learning-based fundus image reading may be a more useful and beneficial tool for screening and diagnosing systemic and ocular diseases after further development.

Subject terms: Biomarkers, Medical research

Introduction

Changes occur naturally in human retina and optic disc over the lifetime in a human16. The retina and optic disc, in addition to the final structures in itself in a visual system, share some physiological characteristics with the brain as they differentiate from diencephalon during embryonic development7. Therefore, the effect of normal ageing changes have been studied through histological or multimodal imaging techniques8. Age is the single most reliable surrogate as the milestone in ageing and growing; therefore, researchers have attempted to predict patients’ chronologic age from their medical examinations and images (e.g., X-rays, facial photographs, and DNA), aiming to reflect the status of a target organ and/or the whole body914. In particular, as retinal fundus images provide high-resolution in-vivo images of retinal vessels and parenchyma without any invasive procedure15, retinal fundus images have been used to detect target organ damage in vascular diseases (e.g., hypertension and diabetes mellitus [DM]), to screen retinal and optic disc diseases (e.g., age-related macular degeneration and glaucoma), and to predict cerebral/cardiovascular diseases1621. Recently, deep neural networks have revolutionised the field of medical image analysis including retinal fundus images; deep-learning algorithms presented discriminative performances comparable to those of an ophthalmologist in diagnosing diabetic retinopathy, age-related macular degeneration, and glaucoma, as well as in predicting age, sex, and presence of cardiovascular diseases2226. Previous study reported highly accurate results of a mean absolute error of 3.26 years for age prediction and receiver operating characteristic curve (AUC) of 0.97 for gender prediction25, however, there were several limitations including a low proportion of Asian, a lack of analysis of the differences in accuracy with age, and a lack of discussion of which parts of the fundus image were used for age prediction. Virtually nothing is known regarding how a neural network predicts age and sex through fundus images, and which compartments of retinal fundus images are crucial for discrimination and identification. In addition, many systemic diseases are known to be associated with retinal diseases, and studies on the retina changes in patients with systemic diseases are actively conducted. However, how systemic diseases change retina and optic nerves and how these changes differ from age-related changes are still unclear. That is, the common features and differences between pathologic changes in and the ageing process of the retinal fundus are not well known16.

The convolutional neural network (CNN) proposed by Krizhevsky et al.27 demonstrated high performance in image recognition, many models for medical image recognition tasks have been developed and showed good results28,29. We expected the CNN to find the proper feature for the task from the retinal fundus images and to produce accurate results. We investigated how well a CNN predicts age and sex through retinal fundus images, and subsequently investigated according to the participants’ age, sex, smoking status, as well as whether the presence of hypertension and DM. Furthermore, we investigated the role of retinal vessels in predicting age and sex by inpainting vessels from retinal fundus images.

Materials and Methods

Dataset Organisation

We used the retinal fundus images from the Seoul National University Bundang Hospital Retina Image Archive (SBRIA) after de-identification except the age, sex, and underlying diseases at the study date; details are described in our previous study26,30. We included 412,026 retinal fundus images from 155,449 participants obtained at the health promotion centre in Seoul National University Bundang Hospital (SNUBH) between June 1st, 2003, and June 30th, 2016, in which detailed information regarding the presence of hypertension, DM, and the smoking status was presented. Among them, 243,668 images from normal participants without hypertension, DM, and smoking history, 40,659 images from participants with hypertension, 14,189 images from participants with DM, and 113,510 images from participants with smoking exist. The number of participants and images in each group, and the number of participants and images with overlapping diseases are shown in the Supplementary Table S1. In addition, the number and percentage of retinal fundus images according to the age and sex of each test-set are shown in Supplementary Table S2. As the participants underwent fundus photographs at least 1-year or more intervals, the fundus photographs of the same patient captured on different dates were considered to be distinct from each for age prediction. Retinal fundus images were acquired using various fundus cameras (Kowa VX-10, Kowa VX-10a, Kowa nonmyd7, Canon CF60Uvi, Canon CR6-45NM). This study was approved by the Institutional Review Board (IRB) of the SNUBH (IRB no. B-1703-386-103), and the requirement of informed consent was waived from the IRB. The study complied with the guidelines of the Declaration of Helsinki.

Pre-processing and experimental setup

Every RGB channel of each input image was normalised to a z-score31. This ensures the classification results to be invariant of intensities and colour contrasts of the images, thereby enabling the model to make predictions solely based on the shape configurations of the fundi. The black background of the fundus images was excluded in normalisation.

The images from normal participants are distributed randomly into training-set, validation-set, and test-set of ratios 89% (216,866), 1% (2,436), and 10% (24,366) by uniform sampling, respectively. We developed and trained CNN for the prediction of age and sex using the training-set from normal participants and verified the models using the validation-set while training the models. We assessed the performance of the models using the test-set of 24,366 normal images (normal test-set) and the groups of images from participants having hypertension, DM, and smoking history, separately. By performing the same number of every epoch time, we obtained the sample mean of the resulting prediction probabilities. Figure 1 shows a flow chart for the experimental setup.

Figure 1.

Figure 1

Flow chart illustrating the processes of data partitioning, training, validation, and testing. SBRIA, Seoul National University Bundang Hospital Retina Image Archive; DM, diabetes mellitus.

Convolutional neural network (CNN) as regressor and classifier

Separate CNNs were trained for age and sex prediction, respectively. We used Pytorch (ver. 0.4.1, https://pytorch.org/) as the deep-learning library to implement the software to train, validate, and test the CNN. The network structure of the CNN adopted the structure of the residual network (ResNet), which had previously demonstrated high performance in a classification task32. ResNet has five versions (ResNet-18, 34, 50, 101, 152) depending on the depth of the convolution. We chose ResNet-152 because the deeper the convolution, the better performance. ResNet-152 has a total of 152 layers and its configuration is as follows. The set consists of a convolution layer, batch normalization, and ReLU (activation function) with 151 layers, and a fully connected last layer. Only the first convolution layer is set to 7×7 kernel with stride 2 and padding 3, and the kernel size of all subsequent convolution layers is 3×3. In addition, by setting stride 2 in 4 convolution layers, multi-scale can be considered without additional pooling layer. We used ResNet-152 as a backbone network, which was pre-trained in general image classification from ImageNet database. We used the parameters of the convolution layers in pre-trained network as our initial values, except for the fully connected last layer. The pre-trained network had learned various features through the huge amount of images from the ImageNet database. Therefore, the transfer learning method from the pre-trained network is faster than the scratched method and shows better performance33. While the overall structure is similar, some details were modified regarding the average pooling layer (changed kernel size from 7 to 16) and the fully connected layer (changed output dimension from 1000 to 1) to tailor the CNN to our data and desired output (Fig. 2). We define the unprocessed numerical output of the CNN as the predicted age, making the CNN for age prediction a direct regressor, while the output of the sigmoid function for the numerical output of the CNN is defined as the probability of predicted sex, making the CNN for sex prediction a classifier. As a loss function for training and error calculation, the smooth L1 loss function was adopted in the age prediction model because it has the appropriate properties of both regular features of L1 and L23436, and the binary cross-entropy function was adopted in the sex prediction models for binary classification. In addition, Adam was used as the optimization scheme (learning rate = 1e−5, beta1 = 0.9, and beta2 = 0.999)37. One epoch is defined as performing backpropagation once for all images in the total training-set, and both CNNs for age and sex prediction were learned through 10 epochs. The CNNs were verified for each epoch using the validation-set; finally, the best results were built into the prediction model.

Figure 2.

Figure 2

Schematics of convolutional neural network. The retinal image is an anonymous image.

Class activation map analysis

To identify the region used in CNN, we used the class activation map (CAM) technique to highlight the core regions that the network focuses on for the age and sex prediction via Bayesian approximation for estimation of uncertainty in predictions31,38,39.

Prediction of age and sex in retinal fundus images after inpainting retinal blood vessels

To determine the effect of retinal blood vessels in retinal fundus images on predicting age and sex, we created vessel-erased images. First, the blood vessel region was extracted using the scale-space approximated CNN (SSANet) that was previously reported by our group and demonstrated state of the art performance in retinal vessel segmentation40. Subsequently, we used the inpainting technique, which naturally fills holes or some regions in the image by extrapolation from the surrounding background. We applied the method previously reported by Telea et al.41, to inpaint the vascular regions, that is, to essentially erase the vessels from the retinal image. All images of the training-set, validation-set, and test-set were first pre-processed identically to the original training-set, then reconstructed as stated above, and our age and sex prediction models were trained, validated, and tested again in the same manner described above.

Statistical analysis

To analyse the relationship between predicted age and chronologic age, the absolute values of error and their distributions were obtained from each test-set including normal, hypertension, DM, and smoking status. To further assess the statistical significance of the performance of our age prediction model, we used linear regression to obtain their coefficients of determination (R2) and 95% confidence interval (CI). The mean absolute error (MAE) from the chronologic age was calculated for the testing-sets, for the age subsets that are divided by decade. Additionally, to obtain the accuracy for age prediction, each result was interpreted as correct if the predicted age was concordant with the chronologic age with certain error margins of ±1, ±3, and ±5 years. The analysis of variance (ANOVA) was conducted to compare the values of squared error according to the underlying vascular conditions. Tukey’s post-test was used for pairwise group comparison. The area under a receiver operating characteristic curve (AUC) and 95% CI were used to report the performance of sex prediction. All statistical analysis was performed using Python; in particular, 95% CI were estimated by the bootstrap technique42.

Results

Table 1 shows the baseline demographics of each dataset. The mean chronologic age (years ± standard deviation) was 46.70 ± 16.67 in the normal training-set, 46.63 ± 15.86 in the normal validation-set, and 46.64  ±  15.83 in the normal test-set. The mean chronologic age (years) in the fundus photographs from participants with underlying vascular conditions was 56.71  ±  9.66 in the hypertension test-set, 56.56  ±  9.75 in the DM test-set, and 49.35  ±  10.05 in the smoking test-set. All test-sets consisted of more males than females, especially in the smoking group (normal, male 55.1%; hypertension, male 77.4%; DM, male 74.1%; smoking, male 93.5%).

Table 1.

Baseline demographics of normal training-set and validation set, and test-sets with various conditions: normal, hypertension, diabetes, and smoking.

Sex Number of subjects Fundus photo image, N (%) Age, mean (SD, range)
Normal
Training-set All 84,526 216,866 46.70 (16.67, 0-105)
Male 44,714 (52.90%) 119,442 (55.08%) 46.68 (14.80, 0-105)
Female 39,812 (47.10%) 97,424 (44.92%) 46.73 (18.70, 1-98)
Validation-set All 2,397 2,436 46.63 (15.86, 1-92)
Male 1,341 (55.94%) 1,372 (56.32%) 46.46 (15.18, 2-87)
Female 1,056 (44.06%) 1,064 (43.68%) 46.84 (16.69, 1-92)
Test-set All 20,823 24,366 46.64 (15.83, 0-95)
Male 11,327 (54.40%) 13,427 (55.11%) 46.69 (14.90, 1-92)
Female 9,496 (45.60%) 10,939 (44.89%) 46.57 (16.91, 0-95)
Hypertension test-set All 12,168 40,659 56.71 (9.66, 20-90)
Male 7,762 (63.79%) 27,389 (77.36%) 55.19 (9.65, 22-90)
Female 4,406 (36.21%) 13,270 (32.64%) 59.83 (8.90, 20-86)
Diabetes test-set All 4,545 14,189 56.56 (9.75, 19-90)
Male 3,191 (70.21%) 10,509 (74.06%) 55.49 (9.40, 19-90)
Female 1,354 (29.79%) 36,80 (25.94%) 59.63 (10.06, 23-86)
Smoking test-set All 30,990 113,510 49.35 (10.05, 18-86)
Male 28,305 (91.34%) 106,116 (93.49%) 49.69 (9.92, 18-86)
Female 2,685 (8.66%) 7,394 (6.51%) 44.39 (10.64, 18-82)

Data are expressed by means and standard deviations.

The normal test-set of 24,366 images indicated a MAE of 3.06 (95% CI, 3.03-3.09) years. The MAE increased significantly in the hypertension test-set (3.46 [95% CI, 3.44-3.49] years) and DM test-set (3.55 [95% CI, 3.50-3.60] years), while it decreased in the smoking test-set (2.76 [95% CI, 2.75-2.77] years). Figure 3 presents a scatter plot of the linear relationship between chronologic age and predicted age in each test-set. Supplementary Fig. S1 presents a box plot of the squared error values ([predicted age – chronologic age]2) of the four groups. ANOVA showed significant difference at the P < 0.05 level for the squared error values of the four underlying vascular conditions, and post hoc comparisons also indicated that the means of squared error values for each vascular condition were different. (P < 0.05, Supplementary Table S3). The normal test-set demonstrated a fairly linear relationship (R2 = 0.92 [95% CI, 0.92-0.93]), and other test-sets demonstrated relatively low coefficients of determination: R2 of 0.74 (95% CI, 0.75-0.76), 0.75 (95% CI, 0.74-0.75), and 0.86 (95% CI, 0.86-0.86) in hypertension, DM, and smoking test-sets, respectively. Based on the assumption that age prediction was correct when the predicted and chronologic ages differed by an error margin of ±5 years, we observed an 82.8% correct prediction in the normal test-set. The accuracies of prediction were 77.6%, 77.0%, and 85.6% in the hypertension, DM, and smoking test-set, respectively. The accuracy for the prediction of age within ±1 year was less than 30% in all groups, and that within ±3 years was more than 55% in all groups (Table 2).

Figure 3.

Figure 3

Predicted and chronologic ages in the four test-sets: (a) normal, (b) hypertension, (c) diabetes, (d) smoking. The lines represent y = x values. Fair linear relationships were indicated in normal test-set (R2 = 0.92, 95% confidence interval [CI] 0.92-0.93), while the others indicated relatively loose relationship as follows: hypertension, R2 = 0.75, 95% CI 0.75–0.76; diabetes mellitus, R2 = 0.75, 95% CI 0.74–0.75; smoking, R2 = 0.86, 95% CI 0.86–0.86.

Table 2.

Mean absolute error (MAE) and its 95% confidence interval (CI); coefficient of determination (R2) and its 95% CI; accuracy in predicting age within maximum differences of ±1, ±3, and ±5 years in each test-set.

Test-set Mean Absolute Error, year (95% CI) Coefficient of determination, R2 (95% CI) Accuracy for age prediction by error margin
±1 year ±3 years ±5 years
Normal 3.06 (3.03-3.09) 0.92 (0.92-0.93) 24.00% 62.11% 82.82%
Hypertension 3.46 (3.44-3.49) 0.75 (0.75-0.76) 20.78% 56.33% 77.63%
Diabetes 3.55 (3.50, 3.60) 0.75 (0.74-0.75) 20.42% 55.09% 77.01%
Smoking 2.76 (2.75, 2.77) 0.86 (0.86-0.86) 25.25% 64.79% 85.59%

CI, Confidence interval.

When each test-set was divided into age subgroups by decades, the MAEs of the estimation in age groups between 20 and 59 years were below 3.0 years in all test-sets. Table 3 shows a fairly consistent MAE and good accuracy under 60 years old; however, an increased MAE and declined accuracy were exhibited in groups aged above 60 years old. Table 4 shows the MAEs and accuracies of the estimations according to three age categories (1: <30 years; 2: 30–59 years; 3: ≥60 years) in each test-set. The MAE in category 3 were higher than that in category 2 in all test-sets (normal, 4.96 vs. 2.60; hypertension, 4.75 vs. 2.69; DM, 4.74 vs. 2.84; smoking, 4.42 vs. 2.47). The accuracies for age prediction within ±5 years were more than 80% in categories 1 and 2, but less than 70% in category 3. Figure 4 shows the changes in accuracy of age prediction according to chronologic age subgroups (divided by 10 years) by error margins of ±1, ±3, and ±5 years in the four test-sets. The accuracy was the highest at 30–40 years of age in the normal and DM test-sets, and 20–30 years in the hypertension and smoking test-sets. The accuracies of age prediction declined gradually with increasing age and deteriorated after 60 years old. The accuracies of the age prediction did not differ according to sex in all groups. Representative images of the CAM heat-map in our age prediction model are shown in Fig. 5. It shows the regions that have higher influence on the prediction results in red, relative to the regions with lower influence, in blue. The CAM of our age prediction model indicated activation primarily in the vascular region.

Table 3.

Mean predicted age, mean absolute error (MAE) and its 95% confidence interval (CI); accuracy for predicting age within maximum differences of ±1, ±3, and ±5 years, respectively, in each test-set divided by 10-year age subgroup.

Test-set Age group Fundus photo images, N Mean chronologic age, year Mean predicted age, year Mean absolute error, year (95% CI) Accuracy for age prediction by error margin
±1 year ±3 years ±5 years
Normal 0-9 1,345 5.72 8.09 2.82 (2.70-2.95) 21.71% 61.49% 85.13%
10-19 473 13.30 11.51 3.02 (2.86-3.19) 18.39% 58.14% 81.82%
20-29 746 25.56 25.57 2.82 (2.64-3.01) 29.22% 67.56% 85.12%
30-39 3,317 35.67 36.01 2.35 (2.29-2.42) 29.57% 72.63% 91.23%
40-49 7,681 44.61 44.83 2.40 (2.36-2.45) 28.97% 70.95% 89.68%
50-59 6,360 53.92 53.46 2.97 (2.92-3.04) 22.81% 61.08% 83.82%
60-69 3,061 63.96 62.06 4.29 (5.18-4.40) 15.52% 45.18% 68.02%
70-79 1,169 73.47 69.82 6.08 (5.83-6.32) 9.24% 30.97% 51.84%
80-89 208 82.84 76.57 8.45 (7.68-9.25) 6.25% 18.27% 37.02%
90-99 6 91.33 82.20 9.51 0.00% 16.67% 33.33%
Hypertension 0-9 0
10-19 4 19.00 22.22 3.22 25.00% 50.00% 50.00%
20-29 118 26.66 27.77 2.14 (1.91-2.38) 26.27% 74.58% 93.22%
30-39 1,181 36.41 37.02 2.30 (2.20-2.40) 29.72% 72.06% 92.04%
40-49 7,959 45.46 46.08 2.43 (2.39-2.46) 28.22% 70.29% 89.45%
50-59 16,064 54.56 54.36 2.85 (2.81-2.88) 23.13% 62.28% 84.28%
60-69 11,226 64.02 61.62 3.85 (3.80-3.90) 16.15% 48.41% 71.97%
70-79 3,811 73.01 66.78 6.79 (6.66-6.92) 7.69% 24.17% 42.04%
80-89 294 81.81 69.27 12.61 (12.03-13.25) 1.02% 3.40% 9.18%
90-99 2 90.00 78.34 11.65 0.00% 0.00% 0.00%
Diabetes 0-9 0
10-19 0
20-29 26 25.00 26.93 2.38 (1.17-3.10) 30.77% 69.23% 80.77%
30-39 469 36.48 37.04 2.26 (2.13-2.40) 28.78% 72.49% 92.96%
40-49 2,947 45.51 46.18 2.64 (2.57-2.71) 26.94% 66.14% 86.97%
50-59 5,449 54.63 54.66 3.00 (2.95-3.06) 21.95% 59.17% 82.42%
60-69 3,869 63.98 62.02 3.75 (3.67-3.83) 16.90% 50.66% 73.79%
70-79 1,305 73.02 66.69 6.98 (6.74-7.22) 8.51% 24.67% 41.76%
80-89 122 81.51 69.63 11.90 (10.94-12.89) 0.00% 3.28% 13.11%
90-99 2 90.00 78.34 11.65 0.00% 0.00% 0.00%
Smoking 0-9 0
10-19 64 18.84 20.89 2.34 (2.03-2.66) 23.44% 64.06% 96.88%
20-29 2,345 26.33 26.86 2.05 (2.00-2.11) 31.00% 76.42% 95.18%
30-39 14,960 35.86 36.11 2.16 (2.14-2.19) 30.20% 74.29% 92.91%
40-49 41,915 44.74 44.95 2.30 (2.28-2.31) 29.07% 71.98% 91.07%
50-59 36,778 53.86 53.40 2.78 (2.76-2.80) 23.63% 62.44% 85.14%
60-69 13,887 63.70 61.24 3.80 (3.76-3.94) 16.28% 47.74% 71.92%
70-79 3,340 72.86 66.98 6.49 (6.36-6.63) 7.78% 24.85% 43.86%
80-89 221 81.61 69.43 12.21 (11.56-12.87) 1.36% 3.62% 9.05%
90-99 0

Consistent MAE and fair accuracy are shown under 60 years old; however, increased MAE and declined accuracy are shown in age subgroup above 60 years old.

CI, Confidence interval.

Table 4.

Mean predicted age, mean absolute error (MAE) and its 95% confidence interval (CI); accuracy in predicting age within maximum differences of ±1, ±3, and ±5 years, divided into three categories in each test-set as follows: (1) age under 30; (2) age 30–59; (3) age over 60.

Test-set Age group Fundus photo images, N Mean chronologic age, year Mean predicted age, year Mean absolute error, year (95% CI) Accuracy for age prediction by error margin
±1 year ±3 years ±5 years
Normal 0–29 2,564 12.89 13.81 2.86 (2.77–2.95) 23.28% 62.64% 84.52%
30–59 17,358 46.31 46.31 2.60 (2.57–2.63) 26.83% 67.66% 87.83%
60–99 4,444 67.38 64.81 4.96 (4.85–5.07) 13.41% 40.14% 62.26%
Hypertension 0–29 122 26.41 27.59 2.18 (1.94–2.42) 26.23% 73.77% 91.80%
30–59 2,5204 50.83 50.93 2.69 (2.66–2.71) 25.04% 65.26% 86.28%
60–99 15,333 66.60 63.05 4.75 (4.70–4.81) 13.75% 41.52% 63.31%
Diabetes 0–29 26 25.00 26.93 2.38 (1.71–3.10) 30.77% 69.23% 80.77%
30–59 8,865 50.64 50.91 2.84 (2.80–2.89) 23.97% 62.19% 84.49%
60–99 5,298 66.62 63.35 4.74 (4.64–4.84) 14.44% 43.15% 64.48%
Smoking 0–29 2,409 26.14 26.71 2.06 (2.01–2.12) 30.80% 76.09% 95.23%
30–59 93,653 46.90 46.86 2.47 (2.45–2.48) 27.11% 68.60% 89.03%
60–99 17,448 65.68 62.44 4.42 (4.38–4.47) 14.47% 42.80% 65.75%

A small MAE is shown under 3.0 (years) in categories 1 and 2, and a large MAE over 4.0 (years) in category 3. Accuracy for predicting age within ±5 years was higher than 80% in categories 1 and 2, but less than 70% in category 3.

CI, Confidence interval.

Figure 4.

Figure 4

Changes in accuracy of age prediction according to chronologic age subgroups (divided by 10 years) by error margins of ±1, ±3, and ±5 years in the four test-sets and inpainted vessel-erased images from all test-sets: (A) normal, (B) hypertension, (C) diabetes, (D) smoking, (E) inpainted normal images, (F) inpainted hypertension images, (G) inpainted diabetes images, (H) inpainted smoking images. The accuracies were the highest in the 20 s to 40 s in all test-sets, decline gradually with age, and decrease significantly after 60 years old. Differences according to sex were not obvious. In inpainted vessel-erased images, the accuracy was similar compared with original images in all test-sets.

Figure 5.

Figure 5

Representative class activation mapping (CAM) heat-map and its original image in age prediction model. It shows the regions that have higher influence on the prediction results in red, relative to the regions with lower influence, in blue. The CAM of the age prediction model indicated activation primarily in the vascular region.

The age prediction model trained by the vessel-erased images was assessed to obtain the MAE and accuracy in the vessel-erased images in all test-sets. The predicted age from the vessel-erased images indicated a similar MAE of 3.19 years (Table 5), and also showed similar changes with age (Fig. 4). In the retinal vessel-erased images, attention is still focused on the area where blood vessels are present (Fig. 6).

Table 5.

Analysis of reconstructed retinal vessel-erased images - mean predicted age, mean absolute error (MAE) and its 95% confidence interval (CI); accuracy for predicting age within maximum differences of ±1, ±3, and ±5 years, divided into three categories in each test-set as follows: (1) age under 30; (2) age 30–59; (3) age over 60.

Reconstructed test-set of retinal vessel-erased images Age group Fundus photo images, N Mean chronologic age, year Mean predicted age, year Mean absolute error, year (95% CI) Accuracy for age prediction by error margin
±1 year ±3 years ±5 years
Normal 0–29 2,564 12.89 14.58 5.92 (5.77, 6.07) 20.36% 59.13% 82.76%
30–59 17,358 46.31 46.54 6.70 (6.68, 6.73) 25.38% 64.93% 86.35%
60–99 4,444 67.38 63.89 4.32 (4.28, 4.36) 13.77% 38.73% 60.31%
Hypertension 0–29 122 26.41 30.60 4.46 (4.11, 4.84) 5.74% 29.51% 61.48%
30–59 25,204 50.83 52.98 3.99 (3.96, 4.03) 17.89% 48.87% 70.68%
60–99 15,333 66.60 64.24 4.57 (4.52, 4.62) 14.86% 42.65% 64.36%
Diabetes 0–29 26 25.00 30.48 5.58 (4.54, 6.70) 0.00% 15.38% 61.54%
30–59 8,865 50.64 52.79 4.01 (3.95, 4.07) 18.33% 48.34% 70.23%
60–99 5,298 66.62 64.47 4.58 (4.49, 4.66) 15.16% 42.90% 64.59%
Smoking 0–29 2,409 26.14 31.74 5.92 (5.77, 6.07) 8.18% 26.90% 47.07%
30–59 93,653 46.90 53.21 6.70 (6.68, 6.73) 9.24% 27.50% 44.91%
60–99 17,448 65.68 66.74 4.32 (4.28, 4.36) 15.22% 43.67% 66.10%

A small MAE is shown under 3.0 (years) in categories 1 and 2, and a large MAE over 4.0 (years) in category 3. Accuracy for predicting age within ±5 years was higher than 80% in categories 1 and 2, but less than 70% in category 3.

CI, Confidence interval.

Figure 6.

Figure 6

Class activation mapping (CAM) heat-map in original fundus photograph image and retinal vessel-erased image in age prediction model. CAM heat-map still focused on the retinal vascular arcade area, thus resulting in an obvious decline in the accuracy of age prediction, as shown in Table 5.

Our sex prediction model was trained and validated using the same training-set and validation-set that were tested in four test-sets; it demonstrated excellent accuracies, where the AUC was 0.97 in the normal test-set, and similar AUCs above 0.96 were shown in the other test-sets with underlying vascular conditions (hypertension, 0.96; DM, 0.96; smoking, 0.98). To confirm the significance of the fovea and retinal vessels in the prediction of sex, we generated inpainted images with erased fovea and retinal vessels, separately. The AUC was 0.881 (95% CI, 0.877–0.885) in the fovea-erased image and 0.682 (95% CI, 0.676–0.688) in the retinal vessel-erased image (Fig. 7). Representative images of the CAM heat-map in our sex prediction model are shown in Fig. 8. The CAM of our sex prediction model indicated various activations in the fovea, optic disc, and retinal vessel; in particular, the proximal vascular region was prominently activated in females.

Figure 7.

Figure 7

Receiver operating characteristic (ROC) curves in four test-sets and inpainting (vessel-erased and fovea-erased) images. The area under curve (AUC) was more than 0.96 in every test-set regardless of underlying condition, while the AUCs in the inpainting images were much smaller.

Figure 8.

Figure 8

Representative class activation mapping (CAM) heat-map in sex prediction model. The area painted in red is the predominant region for sex prediction, while the blue area is used less. The CAM heat-map indicated various focuses in fovea, optic disc, and retinal vessels; in particular, proximal retinal vascular arcade lesion was prominently activated in females.

Discussion

Historically, researchers have sought a relevant biomarker among medical images that can reflect chronologic ageing or biological ageing. Retinal fundus images are easy to capture, inexpensive, non-invasive, and provide high-resolution images of retinal blood vessels as well as of retina and optic nerve that change according to age and vary with sex. In this respect, we developed algorithms predicting chronologic age in retinal fundus images using deep neural networks, and the algorithms demonstrated fair performance in predicting age and sex. Interestingly, the performance was the best in participants aged 20–40 years, decreased as age increased, and poor in participants aged 60 years or over. In addition, the performance deteriorated in participants suffering hypertension or DM with increasing age. The performance of the sex prediction model demonstrated high accuracy and showed no inferiority in test-sets with underlying disease.

CNN have led to a series of breakthroughs for computer vision area including image classification, semantic segmentation, object detection, and bounding box regression27,4345. Network models such as VGG net, GoogLeNet, and ResNet have been widely used to date46. Among them, ResNet proposed the concept of residual blocks to improve performance that can be degraded in very deep networks, and solved the interference with convergence such as vanishing/exploding gradients4750. Through this, ResNet showed better performance compared to VGG net and GoogLeNet in deep network learning. ResNet also demonstrated high performance for medical image recognition tasks such as retinal vessel segmentation40. Therefore we used ResNet, which has the advantages mentioned above, to develop an age and sex prediction model using the retinal fundus images.

Deep-learning based age and sex estimation using retinal fundus images has been previously studied by other group. Poplin et al. reported the predictions of age, gender, and cardiovascular risk factors using fundus photograph data from UK Biobank and EyePACS25. In the algorithm validation using the UK Biobank validation dataset (n = 12,026 patients), the MAE was 3.26 (95% CI, 3.22-3.31) and R2 = 0.74 (95% CI, 0.73-0.75). In the algorithm validation using the EyePACS-2K validation dataset (n = 999 patients), the MAE was 3.42 (95% CI, 3.23-3.61) and the R2 was 0.82 (95% CI, 0.79-0.84). The MAE of the age prediction validation in our study was 3.06 (95% CI, 3.03–3.09) and the R2 was 0.92, which demonstrated a better or comparable performance than those of previous study. Although there are differences in the data set, the present study showed comparable results using a large number of Korean’s fundus images. The cross validation of each other’s dataset may be required to confirm the superiority of accuracy.

Vascular risk-scoring methods use equations based on large cohort studies, such as the US Framingham Heart and Offspring Studies51, the recent European Systematic Coronary Risk Evaluation project52, and the German Prospective Cardiovascular Münster study53. The Framingham score calculation, a well-known algorithm, considers age, sex, total and HDL cholesterol, systolic blood pressure, and smoking51. Additionally, various novel biomarkers such as intravascular ultrasound in left main coronary arteries54 and the ankle brachial index55 have been studied to predict vascular age and risk. Retinal images can also be used as a tool for such predictions, and this application has been demonstrated in our study using fundus photograph images and the CNN algorithm. In addition, retinal vessels and cerebral vessels exhibit similarity and in vivo direct observations of retinal vessels provide information about cerebral vessels56,57. Therefore, efforts to find and understand cerebral and systemic blood vessels through retinal fundus images are worthwhile. Furthermore, retinal images will be more interesting and valuable in systemic vascular diseases such as hypertension and DM.

Our age prediction model demonstrated good accuracies in ages under 60 years; however, in ages 60 or over, the performance deteriorated significantly. This is because ageing changes observed in retinal fundus occur continuously until the age of 60 years, and may saturate at approximately 60 years of age; subsequently, after the age of 60 years, the ageing changes may not be obvious with age. The CAM images of our age prediction model demonstrated activation primarily in the vascular region; this may be another evidence to explain the focus of the deep-learning CNN, however, similar results on vessel-erased images showed that not only blood vessels were used for prediction. In other words, optic disc, papillary vessels, and retinal parenchyma can be used to predict age. Our research was inspired by the effects of ageing on the human fundus, beyond reconfirmation of the possibility or capability of age prediction by deep learning. Interestingly, our age prediction model demonstrated similar MAEs in each test-set categorised by underlying diseases, implying that pathologic changes occurring in systemic vascular diseases are different form the changes in the ageing process. In other words, our results demonstrated that systemic vascular diseases such as hypertension, DM, and smoking resulted in various inconsistent changes in the retinal blood vessels and resulted in loosened relationships (increased coefficient of determination, R2). In addition, statistical analysis using ANOVA and post-hoc test for values of absolute errors in each test-set show significant differences in age prediction when the underlying disease is present.

The ageing process causes many changes in the retina and optic nerve that consist of several layers including RNFL, ILM, ONL, PRL, and RPE8,58. Retinal vascular changes include decrease in cellularity of peripheral capillaries and diminution of the number of capillaries of the fovea8,59. Thickening and hyalinisation of the vessel wall, and arteriosclerotic changes may also develop in retinal vessels with ageing8. In addition, the density of the choriocapillaris that provides nutritional support for the RPE and outer retina, decreases with ageing6062. In particular, the increase in flow deficit of the choriocapillaris is prominent in the central 1-mm circle of the macula62. The fundus autofluorescence from lipofuscin of RPE correlates with age, and indicates age-dependent changes in the fovea63. The reason for such changes is that the human body is always maintained through metabolism, and in this process, cells and tissues are damaged while by-products accumulate2,64. Hence, changes due to normal ageing and pathological changes share a similar feature and cannot be completely distinguished. In particular, the ageing process and hypertension are known to be associated with vascular network changes such as retinal vascular junctional bifurcation angles65. Therefore, the predicted age is expected to be higher than the chronologic age in the presence of hypertension. However, the MAE and difference between the predicted age and chronologic age in the hypertension test-set were similar to that of the normal test-set in our study. This suggests that changes in retinal vessels due to hypertension show some similar features to ageing, but not entirely mimicking normal ageing. DM is another chronic systemic vascular disease that causes changes in whole blood vessels including retinal vessels. The prevalence of any diabetic retinopathy in 35 population-based studies from 22,896 patients was 34.6% overall among diabetic patients66. Additionally, DM altered all blood vessels and affected the diameter of the retinal blood vessels; however, the branching angle was not significantly different from those of normal participants67. Several studies have reported that wider retinal venular diameters and narrower arteriolar diameters were associated with the presence of diabetic nephropathy and severe levels of diabetic retinopathy6870. In a type-1 DM cohort study, age (odds ratio [OR] per 10 years, 2.43 and 2.02) and retinopathy severity (OR per level, 1.14 and 1.21) were associated with focal retinal arteriolar narrowing and A/V nicking, respectively68. Another type-1 DM cohort study indicated that both wider venular diameters and smaller arteriolar diameters were predictors of the 16-year development of nephropathy, neuropathy, and proliferative retinopathy71. In a type-2 DM cohort study, smaller retinal arteriolar calibers exhibited associations with increasing age and mean arterial BP, and a larger retinal venular caliber was associated with increasing severities of retinopathy and cigarette smoking69. Because the level of diabetic retinopathy and tissue oxygen demand affect the changes in blood vessels69,72, retinal vessels in diabetic patients show various changes, thus resulting in a smaller R2 (coefficient of determination) in diabetic participants than normal participants. In other words, pathologic retinal changes result from various factors; our study suggests that ageing and pathological changes are not exactly the same. Unlike the other groups, the smoking group showed better performance on age prediction. There are several studies on the effects of smoking on retinal vessels73,74, but this does not explain how they perform better than normal. Moreover, there are spectral domain optical coherence tomography studies which shows the thickness of retinal layers in healthy chronic smokers was not significantly different to those of healthy individuals75,76. Sex, age, and confounding factors were likely to improve predictions, in other words, a large number of men over 30 years of age were included in our study, resulting in uniform results and higher R2 values.

Our prediction model predicted sex with good accuracy in all test-sets unlike the results of age prediction. A previous study indicated that the AUC of UK Biobank was 0.97 (95% CI, 0.966-0.971) and that of EyePACS-2K was 0.97 (95% CI, 0.96-0.98), which are comparable to our study25. Considering that the prediction accuracy is reduced significantly in the fovea-erased images and the accuracy is reduced significantly in the vessel-erased images, both the fovea and blood vessel are used for sex prediction where the blood vessel is the core region. Interestingly, the presence of underlying vascular conditions did not indicate a significant effect on sex prediction. This indicates that alterations in retinal blood vessels due to underlying vascular conditions did not exceed beyond sexual differences. Unexpectedly, our sex prediction model showed the highest AUC in the smoking group. This is probably due to the fact that the sex ratio of smokers is biased towards men (Supplementary Table S2). Sex prediction has been studied in forensic science typically, using bones or bone fragments to produce estimating formulas7779. However, these human structures are expected to exhibit significant differences between sexes, and they can be determined or measured by human examiners without a computer. Considering that sex identification in retinal fundus images proved to be almost impossible even when an inspection was performed by experienced ophthalmologists, our results suggest that deep-running is superior to human perception in image discrimination and identification.

Some limitations were present in this study. First, the type of underlying disease and the duration of the disease were not considered, and images showing any ocular disease were not included. Next, changes in retinal fundus images caused by lens yellowing, cataract, and cataract surgery might affect age prediction80; however, the consideration of lens status is lacking in this study. Subsequently, myopia and axial length could cause significant changes in the fundus and optic nerve81,82. It was reported that an increased axial length was associated with arteriolar and venular narrowing; however, the arteriovenous diameter ratio or vessel junctions were not affected significantly by the axial length81. It is unclear which components of the retinal vasculature were used to predict age and sex; at the least, the effects of arteriovenous diameter ratio and junctional exponents were considered to be less disturbed by the axial length81. Next, confounding factors may appear depending on the fundus photograph camera model or its manufacturer. Differences in image size and telecentricity were reported depending on the fundus imaging system83. However, the absolute size of the targets by each fundus imaging system may be different; nevertheless, the ratio of each target size may not be significant. Finally, our study has not been validated in other databases and only Koreans are included in the study. Differences in retina according to ethnicity have been reported84,85, therefore, further validation studies involving other database especially other ethnicities are warranted.

Our model demonstrates accurate and highly reliable age estimates especially in normal participants under age 60 years. Retinal fundus images from participants with underlying conditions (hypertension, DM, or smoking) indicated relatively low coefficients of determination (R2) between the predicted age and chronologic age, thus suggesting that the ageing process and pathologic vascular changes exhibit different features. Fundus-predicted sex indicated an accuracy of 0.96 of AUC score in all groups. Our CNN-based age and sex prediction model has demonstrated the most improved performance to date. Our research suggests that ageing and systemic vascular diseases have different effects on the retina. Further research on the clinical significance and application of our model to other population groups is needed.

Supplementary information

Acknowledgements

This study was supported by the Small Grant for Exploratory Research of the National Research Foundation of Korea (NRF), which is funded by the Ministry of Science, ICT, and Future Planning (NRF- 2018R1D1A1A09083241). The sponsor or funding organisation had no role in the design or conduct of this research.

Author contributions

Y.D.K., K.J.N., K.H.P. and S.J.P. developed the study concept and design. K.J.N., S.J.B. and S.J.P. extracted data from a database and trained the convolutional neural network. Y.D.K. wrote the manuscript. K.J.N. and S.J.B. prepared the figure and table. S.J.B. conducted statistical analysis. S.C.L, T.K., L.S., K.J.L., S.H.K., K.H.P. and S.J.P. critically revised the drafted manuscript. S.J.P. had full access to all the data in the study and was responsible for the integrity of the data and the accuracy of the data analysis. All authors have reviewed the manuscript.

Data availability

The datasets generated during and/or analysed during the current study are available from the corresponding author on reasonable request.

Competing interests

The authors declare no competing interests.

Footnotes

Publisher’s note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

These authors contributed equally: Yong Dae Kim, Kyoung Jin Noh and Seong Jun Byun.

Supplementary information

is available for this paper at 10.1038/s41598-020-61519-9.

References

  • 1.Dorey CK, Wu G, Ebenstein D, Garsd A, Weiter JJ. Cell loss in the aging retina. Relationship to lipofuscin accumulation and macular degeneration. Investigative Ophthalmology & Visual Science. 1989;30:1691–1699. [PubMed] [Google Scholar]
  • 2.Ardeljan D, Chan C-C. Aging is not a disease: Distinguishing age-related macular degeneration from aging. Progress in Retinal and Eye Research. 2013;37:68–89. doi: 10.1016/j.preteyeres.2013.07.003. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3.Dolman CL, McCormick AQ, Drance SM. Aging of the optic nerve. Archives of Ophthalmology. 1980;98:2053–2058. doi: 10.1001/archopht.1980.01020040905024. [DOI] [PubMed] [Google Scholar]
  • 4.Garway-Heath DF, Wollstein G, Hitchings RA. Aging changes of the optic nerve head in relation to open angle glaucoma. British Journal of Ophthalmology. 1997;81:840. doi: 10.1136/bjo.81.10.840. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5.Bowd C, et al. Imaging of the optic disc and retinal nerve fiber layer: the effects of age, optic disc area, refractive error, and gender. J. Opt. Soc. Am. A. 2002;19:197–207. doi: 10.1364/JOSAA.19.000197. [DOI] [PubMed] [Google Scholar]
  • 6.Kanamori AK, et al. Evaluation of the Effect of Aging on Retinal Nerve Fiber Layer Thickness Measured by Optical Coherence Tomography. Ophthalmologica. 2003;217:273–278. doi: 10.1159/000070634. [DOI] [PubMed] [Google Scholar]
  • 7.London A, Benhar I, Schwartz M. The retina as a window to the brain-from eye research to CNS disorders. Nat Rev Neurol. 2013;9:44–53. doi: 10.1038/nrneurol.2012.227. [DOI] [PubMed] [Google Scholar]
  • 8.Grossniklaus HE, Nickerson JM, Edelhauser HF, Bergman LA, Berglin L. Anatomic alterations in aging and age-related diseases of the eye. Invest Ophthalmol Vis Sci. 2013;54:ORSF23–27. doi: 10.1167/iovs.13-12711. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9.Roche AF, Eyman SL, Davila GH. Skeletal age prediction. The Journal of Pediatrics. 1971;78:997–1003. doi: 10.1016/S0022-3476(71)80430-4. [DOI] [PubMed] [Google Scholar]
  • 10.Lee H, et al. Fully Automated Deep Learning System for Bone Age Assessment. J Digit Imaging. 2017;30:427–441. doi: 10.1007/s10278-017-9955-8. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11.Tanner JM, Gibbons RD. A computerized image analysis system for estimating Tanner-Whitehouse 2 bone age. Horm Res. 1994;42:282–287. doi: 10.1159/000184210. [DOI] [PubMed] [Google Scholar]
  • 12.Kwon YH, Lobo NdV. Age Classification from Facial Images. Computer Vision and Image Understanding. 1999;74:1–21. doi: 10.1006/cviu.1997.0549. [DOI] [Google Scholar]
  • 13.Eidinger E, Enbar R, Hassner T. Age and Gender Estimation of Unfiltered Faces. IEEE Transactions on Information Forensics and Security. 2014;9:2170–2179. doi: 10.1109/TIFS.2014.2359646. [DOI] [Google Scholar]
  • 14.Vidaki A, et al. DNA methylation-based forensic age prediction using artificial neural networks and next generation sequencing. Forensic Science International: Genetics. 2017;28:225–236. doi: 10.1016/j.fsigen.2017.02.009. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15.Patton N, et al. Retinal vascular image analysis as a potential screening tool for cerebrovascular disease: a rationale based on homology between cerebral and retinal microvasculatures. Journal of Anatomy. 2005;206:319–348. doi: 10.1111/j.1469-7580.2005.00395.x. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16.Pose-Reino A, et al. Computerized measurement of retinal blood vessel calibre: description, validation and use to determine the influence of ageing and hypertension. J Hypertens. 2005;23:843–850. doi: 10.1097/01.hjh.0000163154.35577.8e. [DOI] [PubMed] [Google Scholar]
  • 17.Moss SE, Klein R, Kessler SD, Richie KA. Comparison between ophthalmoscopy and fundus photography in determining severity of diabetic retinopathy. Ophthalmology. 1985;92:62–67. doi: 10.1016/S0161-6420(85)34082-4. [DOI] [PubMed] [Google Scholar]
  • 18.Pirbhai A, Sheidow T, Hooper P. Prospective evaluation of digital non-stereo color fundus photography as a screening tool in age-related macular degeneration. Am J Ophthalmol. 2005;139:455–461. doi: 10.1016/j.ajo.2004.09.077. [DOI] [PubMed] [Google Scholar]
  • 19.Tielsch JM, et al. A population-based evaluation of glaucoma screening: the Baltimore Eye Survey. Am J Epidemiol. 1991;134:1102–1110. doi: 10.1093/oxfordjournals.aje.a116013. [DOI] [PubMed] [Google Scholar]
  • 20.Cheung CY, Chen C, Wong TY. Ocular Fundus Photography as a Tool to Study Stroke and Dementia. Semin Neurol. 2015;35:481–490. doi: 10.1055/s-0035-1563570. [DOI] [PubMed] [Google Scholar]
  • 21.Wong TY, et al. Retinal microvascular abnormalities and their relationship with hypertension, cardiovascular disease, and mortality. Surv Ophthalmol. 2001;46:59–80. doi: 10.1016/S0039-6257(01)00234-X. [DOI] [PubMed] [Google Scholar]
  • 22.Xiangyu C, Yanwu X, Damon Wing Kee W, Tien Yin W, Jiang L. Glaucoma detection based on deep convolutional neural network. Conf Proc IEEE Eng Med Biol Soc. 2015;2015:715–718. doi: 10.1109/EMBC.2015.7318462. [DOI] [PubMed] [Google Scholar]
  • 23.Grassmann F, et al. A Deep Learning Algorithm for Prediction of Age-Related Eye Disease Study Severity Scale for Age-Related Macular Degeneration from Color Fundus Photography. Ophthalmology. 2018;125:1410–1420. doi: 10.1016/j.ophtha.2018.02.037. [DOI] [PubMed] [Google Scholar]
  • 24.Gulshan V, et al. Development and Validation of a Deep Learning Algorithm for Detection of Diabetic Retinopathy in Retinal Fundus Photographs. JAMA. 2016;316:2402–2410. doi: 10.1001/jama.2016.17216. [DOI] [PubMed] [Google Scholar]
  • 25.Poplin R, et al. Prediction of cardiovascular risk factors from retinal fundus photographs via deep learning. Nature Biomedical Engineering. 2018;2:158–164. doi: 10.1038/s41551-018-0195-0. [DOI] [PubMed] [Google Scholar]
  • 26.Son Jaemin, Shin Joo Young, Kim Hoon Dong, Jung Kyu-Hwan, Park Kyu Hyung, Park Sang Jun. Development and Validation of Deep Learning Models for Screening Multiple Abnormal Findings in Retinal Fundus Images. Ophthalmology. 2020;127(1):85–94. doi: 10.1016/j.ophtha.2019.05.029. [DOI] [PubMed] [Google Scholar]
  • 27.Krizhevsky Alex, Sutskever Ilya, Hinton Geoffrey E. ImageNet classification with deep convolutional neural networks. Communications of the ACM. 2017;60(6):84–90. doi: 10.1145/3065386. [DOI] [Google Scholar]
  • 28.Spampinato C, Palazzo S, Giordano D, Aldinucci M, Leonardi R. Deep learning for automated skeletal bone age assessment in X-ray images. Med Image Anal. 2017;36:41–51. doi: 10.1016/j.media.2016.10.010. [DOI] [PubMed] [Google Scholar]
  • 29.Ginneken B, Setio A, Jacobs C, Ciompi F. Off-the-shelf convolutional neural network features for pulmonary nodule detection in computed tomography scans. 2015;2015:286–289. doi: 10.1109/ISBI.2015.7163869. [DOI] [Google Scholar]
  • 30.Park, S. J. et al. A Novel Fundus Image Reading Tool for Efficient Generation of a Multi-dimensional Categorical Image Database for Machine Learning Algorithm Training. J Korean Med Sci33 (2018). [DOI] [PMC free article] [PubMed]
  • 31.Jang Y, Son J, Park KH, Park SJ, Jung KH. Laterality Classification of Fundus Images Using Interpretable Deep Neural Network. J Digit Imaging. 2018;31:923–928. doi: 10.1007/s10278-018-0099-2. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 32.He K, Z. X., Ren S, Sun J. Deep Residual Learning for Image Recognition. arXiv preprint arXiv:151203385 (2015).
  • 33.Yu Y, et al. Deep Transfer Learning for Modality Classification of Medical Images. Information (Switzerland) 2017;8:91. doi: 10.3390/info8030091. [DOI] [Google Scholar]
  • 34.Huber P. Robust Estimation of a Location Parameter. Annals of Mathematical Statistics. 1964;35:73–101. doi: 10.1214/aoms/1177703732. [DOI] [Google Scholar]
  • 35.Girshick, R. Fast r-cnn. 10.1109/ICCV.2015.169 (2015).
  • 36.Ren, S., He, K., Girshick, R. & Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. (2016). [DOI] [PubMed]
  • 37.Kingma DP, B. J. A. A method for stochastic optimization. arXiv preprint arXiv:14126980 (2014).
  • 38.Selvaraju, R. R. et al. In 2017 IEEE International Conference on Computer Vision (ICCV) 618–626.
  • 39.Gal, Y. & Ghahramani, Z. In Proceedings of the 33rd International Conference on International Conference on Machine Learning - Volume 48, 1050–1059 (JMLR.org, New York, NY, USA, 2016).
  • 40.Noh Kyoung Jin, Park Sang Jun, Lee Soochahn. Scale-space approximated convolutional neural networks for retinal vessel segmentation. Computer Methods and Programs in Biomedicine. 2019;178:237–246. doi: 10.1016/j.cmpb.2019.06.030. [DOI] [PubMed] [Google Scholar]
  • 41.Telea A. An Image Inpainting Technique Based on the Fast Marching Method. Journal of Graphics Tools. 2004;9:23–34. doi: 10.1080/10867651.2004.10487596. [DOI] [Google Scholar]
  • 42.Harrell, F. E., Jr., Lee, K. L. & Mark, D. B. Multivariable prognostic models: issues in developing models, evaluating assumptions and adequacy, and measuring and reducing errors. Stat Med15, 361–387, doi:10.1002/(SICI)1097-0258(19960229)15:4<361::AID-SIM168>3.0.CO;2-4 (1996). [DOI] [PubMed]
  • 43.Lecun Y, et al. Backpropagation Applied to Handwritten Zip Code Recognition. Neural Computation. 1989;1:541–551. doi: 10.1162/neco.1989.1.4.541. [DOI] [Google Scholar]
  • 44.Sermanet, P. et al. OverFeat: Integrated Recognition, Localization and Detection using Convolutional Networks. International Conference on Learning Representations (ICLR) (Banff) (2013).
  • 45.Zeiler, M. & Fergus, R. Visualizing and Understanding Convolutional Neural Networks Vol. 8689 (2013).
  • 46.Canziani, A., Paszke, A. & Culurciello, E. An Analysis of Deep Neural Network Models for Practical Applications (2016).
  • 47.Balduzzi, D. et al. The Shattered Gradients Problem: If resnets are the answer, then what is the question? (2017).
  • 48.He, K., Zhang, X., Ren, S. & Sun, J. Delving Deep into Rectifiers: Surpassing Human-Level Performance on ImageNet Classification. IEEE International Conference on Computer Vision (ICCV 2015) 1502, 10.1109/ICCV.2015.123 (2015).
  • 49.Ioffe, S. & Szegedy, C. Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift (2015).
  • 50.He, K. & Sun, J. Convolutional neural networks at constrained time cost (2015).
  • 51.Anderson KM, Wilson PW, Odell PM, Kannel WB. An updated coronary risk profile. A statement for health professionals. Circulation. 1991;83:356–362. doi: 10.1161/01.CIR.83.1.356. [DOI] [PubMed] [Google Scholar]
  • 52.Conroy RM, et al. Estimation of ten-year risk of fatal cardiovascular disease in Europe: the SCORE project. Eur Heart J. 2003;24:987–1003. doi: 10.1016/S0195-668X(03)00114-3. [DOI] [PubMed] [Google Scholar]
  • 53.Assmann G, Cullen P, Schulte H. Simple scoring scheme for calculating the risk of acute coronary events based on the 10-year follow-up of the prospective cardiovascular Munster (PROCAM) study. Circulation. 2002;105:310–315. doi: 10.1161/hc0302.102575. [DOI] [PubMed] [Google Scholar]
  • 54.von Birgelen C, et al. Relationship Between Cardiovascular Risk as Predicted by Established Risk Scores Versus Plaque Progression as Measured by Serial Intravascular Ultrasound in Left Main Coronary Arteries. Circulation. 2004;110:1579–1585. doi: 10.1161/01.CIR.0000142048.94084.CA. [DOI] [PubMed] [Google Scholar]
  • 55.Tziomalos K, Athyros VG, Karagiannis A, Mikhailidis DP. The role of ankle brachial index and carotid intima-media thickness in vascular risk stratification. Curr Opin Cardiol. 2010;25:394–398. doi: 10.1097/HCO.0b013e328338c109. [DOI] [PubMed] [Google Scholar]
  • 56.Tso MO, Jampol LM. Pathophysiology of hypertensive retinopathy. Ophthalmology. 1982;89:1132–1145. doi: 10.1016/S0161-6420(82)34663-1. [DOI] [PubMed] [Google Scholar]
  • 57.Wardlaw JM, Dennis MS, Warlow CP, Sandercock PA. Imaging appearance of the symptomatic perforating artery in patients with lacunar infarction: occlusion or other vascular pathology? Ann Neurol. 2001;50:208–215. doi: 10.1002/ana.1082. [DOI] [PubMed] [Google Scholar]
  • 58.Gartner S, Henkind P. Aging and degeneration of the human macula. 1. Outer nuclear layer and photoreceptors. Br J Ophthalmol. 1981;65:23–28. doi: 10.1136/bjo.65.1.23. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 59.Kuwabara T, Cogan DG. Retinal vascular patterns. VII. Acellular change. Invest Ophthalmol. 1965;4:1049–1064. [PubMed] [Google Scholar]
  • 60.Nickla DL, Wallman J. The multifunctional choroid. Prog Retin Eye Res. 2010;29:144–168. doi: 10.1016/j.preteyeres.2009.12.002. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 61.Ramrattan RS, et al. Morphometric analysis of Bruch’s membrane, the choriocapillaris, and the choroid in aging. Invest Ophthalmol Vis Sci. 1994;35:2857–2864. [PubMed] [Google Scholar]
  • 62.Zheng F, et al. Age-Dependent Changes in the Macular Choriocapillaris of Normal Eyes Imaged with Swept-Source OCT Angiography. Am J Ophthalmol. 2019 doi: 10.1016/j.ajo.2018.12.025. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 63.von Ruckmann A, Fitzke FW, Bird AC. Fundus autofluorescence in age-related macular disease imaged with a laser scanning ophthalmoscope. Invest Ophthalmol Vis Sci. 1997;38:478–486. [PubMed] [Google Scholar]
  • 64.Zealley B, de Grey AD. Strategies for engineered negligible senescence. Gerontology. 2013;59:183–189. doi: 10.1159/000342197. [DOI] [PubMed] [Google Scholar]
  • 65.Stanton AV, et al. Vascular network changes in the retina with age and hypertension. J Hypertens. 1995;13:1724–1728. doi: 10.1097/00004872-199501000-00008. [DOI] [PubMed] [Google Scholar]
  • 66.Yau JW, et al. Global prevalence and major risk factors of diabetic retinopathy. Diabetes Care. 2012;35:556–564. doi: 10.2337/dc11-1909. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 67.Luo T, Gast TJ, Vermeer TJ, Burns SA. Retinal Vascular Branching in Healthy and Diabetic Subjects. Invest Ophthalmol Vis Sci. 2017;58:2685–2694. doi: 10.1167/iovs.17-21653. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 68.Klein R, et al. Retinal vascular abnormalities in persons with type 1 diabetes: the Wisconsin Epidemiologic Study of Diabetic Retinopathy: XVIII. Ophthalmology. 2003;110:2118–2125. doi: 10.1016/S0161-6420(03)00863-7. [DOI] [PubMed] [Google Scholar]
  • 69.Klein R, Klein BE, Moss SE, Wong TY, Sharrett AR. Retinal vascular caliber in persons with type 2 diabetes: the Wisconsin Epidemiological Study of Diabetic Retinopathy: XX. Ophthalmology. 2006;113:1488–1498. doi: 10.1016/j.ophtha.2006.03.028. [DOI] [PubMed] [Google Scholar]
  • 70.Tsai AS, et al. Differential association of retinal arteriolar and venular caliber with diabetes and retinopathy. Diabetes Res Clin Pract. 2011;94:291–298. doi: 10.1016/j.diabres.2011.07.032. [DOI] [PubMed] [Google Scholar]
  • 71.Broe R, et al. Retinal vessel calibers predict long-term microvascular complications in type 1 diabetes: the Danish Cohort of Pediatric Diabetes 1987 (DCPD1987) Diabetes. 2014;63:3906–3914. doi: 10.2337/db14-0227. [DOI] [PubMed] [Google Scholar]
  • 72.Meehan RT, et al. An automated method of quantifying retinal vascular responses during exposure to novel environmental conditions. Ophthalmology. 1990;97:875–881. doi: 10.1016/S0161-6420(90)32500-9. [DOI] [PubMed] [Google Scholar]
  • 73.Rose K, Flanagan JG, Patel SR, Cheng R, Hudson C. Retinal blood flow and vascular reactivity in chronic smokers. Invest Ophthalmol Vis Sci. 2014;55:4266–4276. doi: 10.1167/iovs.14-14022. [DOI] [PubMed] [Google Scholar]
  • 74.Omae T, Nagaoka T, Yoshida A. Effects of Habitual Cigarette Smoking on Retinal Circulation in Patients With Type 2 Diabetes. Invest Ophthalmol Vis Sci. 2016;57:1345–1351. doi: 10.1167/iovs.15-18813. [DOI] [PubMed] [Google Scholar]
  • 75.Duman R, Duman R, Sabaner MC, Cetinkaya E. Effect of smoking on the thickness of retinal layers in healthy smokers. Cutan Ocul Toxicol. 2017;36:366–369. doi: 10.1080/15569527.2017.1300162. [DOI] [PubMed] [Google Scholar]
  • 76.Ulas F, Celik F, Dogan U, Celebi S. Effect of smoking on choroidal thickness in healthy smokers. Curr Eye Res. 2014;39:504–511. doi: 10.3109/02713683.2013.850099. [DOI] [PubMed] [Google Scholar]
  • 77.Soltani S, Aghakhani K, Fallah F. Sex prediction potential of hyoid metric measurements in Iranian adults. Leg Med (Tokyo) 2017;25:6–10. doi: 10.1016/j.legalmed.2016.12.010. [DOI] [PubMed] [Google Scholar]
  • 78.Alves, N., Deana, N. F., Ceballos, F., Hernandez, P. & Gonzalez, J. Sex prediction by metric and non-metric analysis of the hard palate and the pyriform aperture. Folia Morphol (Warsz), 10.5603/FM.a2018.0109 (2018). [DOI] [PubMed]
  • 79.Navega D, Vicente R, Vieira DN, Ross AH, Cunha E. Sex estimation from the tarsal bones in a Portuguese sample: a machine learning approach. Int J Legal Med. 2015;129:651–659. doi: 10.1007/s00414-014-1070-5. [DOI] [PubMed] [Google Scholar]
  • 80.Guymer R, Luthert P, Bird A. Changes in Bruch’s membrane and related structures with age. Progress in Retinal and Eye Research. 1999;18:59–90. doi: 10.1016/S1350-9462(98)00012-3. [DOI] [PubMed] [Google Scholar]
  • 81.Patton N, et al. Effect of axial length on retinal vascular network geometry. Am J Ophthalmol. 2005;140:648–653. doi: 10.1016/j.ajo.2005.04.040. [DOI] [PubMed] [Google Scholar]
  • 82.Kim TW, et al. Optic disc change with incipient myopia of childhood. Ophthalmology. 2012;119(21–26):e21–23. doi: 10.1016/j.ophtha.2011.07.051. [DOI] [PubMed] [Google Scholar]
  • 83.Rudnicka AR, Burk RO, Edgar DF, Fitzke FW. Magnification characteristics of fundus imaging systems. Ophthalmology. 1998;105:2186–2192. doi: 10.1016/S0161-6420(98)91214-3. [DOI] [PubMed] [Google Scholar]
  • 84.Wong TY, et al. Retinal vascular caliber, cardiovascular risk factors, and inflammation: the multi-ethnic study of atherosclerosis (MESA) Invest Ophthalmol Vis Sci. 2006;47:2341–2350. doi: 10.1167/iovs.05-1539. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 85.Kelty PJ, et al. Macular thickness assessment in healthy eyes based on ethnicity using Stratus OCT optical coherence tomography. Invest Ophthalmol Vis Sci. 2008;49:2668–2672. doi: 10.1167/iovs.07-1000. [DOI] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Data Availability Statement

The datasets generated during and/or analysed during the current study are available from the corresponding author on reasonable request.


Articles from Scientific Reports are provided here courtesy of Nature Publishing Group

RESOURCES