Skip to main content
Scientific Reports logoLink to Scientific Reports
. 2025 Dec 30;16:3583. doi: 10.1038/s41598-025-33510-9

A deep-learning system for the assessment of coronary heart disease risk via scleral photographs

Yixuan Shi 1,#, Zhaoxuan Ding 2,#, Chuxiang Gao 3, Machao Li 1, Dongsheng Wei 2, Jianing Wang 1, Hao Ma 1, Li Ma 1, Xianbo Luo 1, Jiang Zhu 3,, Xiaoqing Zhang 2,4,, Guoliang Huang 1,5,
PMCID: PMC12847730  PMID: 41469469

Abstract

Cardiovascular disease is a major cause of death worldwide, especially the coronary heart disease (CHD). Scleral blood vessels provide information on the risk of CHD. Here, we report the development and validation of deep learning system that leverages scleral photographs for assessment of CHD risk, using diverse multi-age datasets that comprise more than 5000 images. Risk assessment of CHD measured by the system and by specialist doctors showed high agreement, with overall accuracy of 0.891 and AUC of 0.942. We further demonstrated that the trained deep learning system predominantly relied on vascular abnormalities as interpretable features for prediction, and in a subset of cases, it also captured pigmentation spots. These findings suggest that the model learns physiologically relevant cues linking scleral changes to CHD, thereby enhancing its clinical interpretability. Our findings motivate the development of clinically application explainable deep learning system for the assessment of CHD risk on the basis of the features of vessels and spots in scleral photographs.

Keywords: Coronary heart disease (CHD), Sclera, Deep learning, Convolutional neural network (CNN), U-Net++

Subject terms: Cardiovascular diseases, Scleral diseases

Introduction

Coronary heart disease (CHD) is a major global health concern, primarily caused by the accumulation of plaques in the coronary arteries, leading to restricted blood flow to the myocardium. According to the American Heart Association (AHA)1 and European Society of Cardiology (ESC)2,3, CHD ranks among the leading causes of death worldwide, accounting for millions of deaths annually. Beyond its high mortality rate, CHD significantly affects the quality of life of affected individuals. Common symptoms include angina and shortness of breath, while severe cases may progress to myocardial infarction4, potentially causing irreversible cardiac damage or fatality. The economic burden of CHD is significant, with high medical costs associated with treatment and prolonged management imposing significant pressure on healthcare systems and society. One of the primary challenges in CHD management is that its early stages are often asymptomatic5. However, plaque rupture can trigger acute myocardial infarction, emphasizing the critical importance of early diagnosis6. Identifying high-risk populations and implementing timely interventions can mitigate complications such as heart failure and malignant arrhythmias1. Early detection not only improves long-term survival rates but also enhances the overall well-being of affected individuals, underscoring its critical clinical value.

Traditional Chinese Medicine (TCM) has long held that the heart governs the eyes, and that pathological changes in cardiac function may manifest in ocular features such as eye discoloration, visible vessels, and changes in scleral or conjunctival appearance. Classical texts describe how heart-blood deficiency or stagnation may lead to observable changes in the eye area. Studies such as Gao et al.7. have begun to empirically examine eye-image features in patients with CHD. In modern medical research, there is growing evidence that eye or ocular vascular changes correlate with cardiovascular disease. For example, microvascular alterations in the retina measured via OCTA have been shown to relate to severity of coronary artery disease, left ventricular function, or number of diseased coronary vessels810. Structural changes in ocular tissues such as choroidal thickness have been associated with cardiovascular risk factors11 (e.g. blood pressure, left ventricular mass). Taken together, both TCM theory and accumulating empirical data suggest that ocular features—particularly vascular and appearance changes—may serve as non-invasive indicators of systemic cardiovascular pathology. This motivates present studys focusing on scleral imaging and interpretable features such as vascular abnormalities and pigmentation spots.

At present, scleral appearance is not utilized as a standard diagnostic tool for CHD in modern cardiology. Clinical practice primarily relies on well-established imaging modalities such as echocardiography, coronary computed tomography angiography (CCTA), or invasive coronary angiography. In contrast, Traditional Chinese Medicine (TCM) has historically incorporated ocular inspection as part of syndrome differentiation, where scleral vessel congestion, tortuosity, or pigmentation changes are interpreted as external manifestations of internal circulatory dysfunction, particularly involving the heart12. While these TCM-based observations are widely applied in clinical practice, they remain largely qualitative and have not been systematically validated or integrated into mainstream cardiology. Against this background, our study seeks to provide an objective and quantitative framework that bridges traditional clinical observations with modern deep learning techniques, thereby exploring the potential of scleral imaging as a non-invasive auxiliary marker for CHD risk assessment.

CHD patients frequently exhibit microcirculatory disorders13,14, oxidative stress15,16, and autonomic nervous system dysfunction1719, leading to scleral capillary tortuosity, lipid deposition, or abnormal vascular tone. Observing scleral signs for syndrome differentiation facilitates the early identification of CHD symptoms, including qi stagnation, blood stasis, and phlegm-turbidity, which holds significant clinical value.

Convolutional Neural Networks (CNNs) have become a fundamental component of modern image processing, particularly for tasks involving image recognition and classification20. CNNs are highly effective in feature extraction, as they learn hierarchical representations of data. Lower layers capture simple features such as edges and textures, while higher layers represent increasingly abstract and complex patterns. Additionally, techniques such as transfer learning and fine-tuning enhance CNN performance across diverse tasks, even in scenarios with limited training data. CNNs have been widely applied in medical imaging, achieving remarkable success across diverse fields. For instance, they have been utilized in lung cancer detection using CT scans21,22, skin lesion classification23,24, and diabetic retinopathy screening with fundus images25. These applications highlight the potential of CNNs to match or even surpass human-level performance in disease diagnosis based on medical images26,27. Given the proven success of CNNs in medical imaging, extending this technology to scleral image analysis as an auxiliary tool for coronary heart disease diagnosis represents both a logical and promising direction.

Building on these insights2832, we propose a deep-learning system that leverages CNNs to analyze scleral image features for the auxiliary diagnosis of CHD. The scleral image analysis process comprises three key stages: image segmentation, deep feature extraction and image classification. We employ the U-Net + + algorithm to segment the sclera from eye images, utilize the DenseNet121 model to extract deep features, and implement a multiple instance learning (MIL) model with an attention mechanism for classification. The proposed algorithm attained an average classification accuracy of 0.891 and an AUC of 0.942. The training dataset was internally curated and comprises scleral images from 500 participants, including 240 CHD patients and 260 healthy individuals. Furthermore, by visualizing the extracted deep features, we showed that the trained deep learning system primarily focused on scleral blood vessels, which aligns with known microcirculatory abnormalities in CHD patients, and in certain cases, it also highlighted pigmentation spots as auxiliary cues. While vascular features were consistently emphasized across most visualizations, pigmentation spots were less frequently observed but remain physiologically plausible given their association with oxidative stress and lipid deposition. This interpretability analysis suggests that the model captures physiologically relevant patterns rather than relying on spurious correlations, thereby supporting the potential of scleral imaging as an auxiliary diagnostic tool for CHD. Nevertheless, we acknowledge that pigmentation-related findings require further validation in larger datasets and with expert-annotated reference standards.

Methods and materials

Utilizing CNNs for robust deep feature extraction, we propose a CHD risk prediction system based on scleral images. As illustrated in Fig. 1, the system consists of three main components: feature extraction, classification, and prediction. The feature extraction module employs the DenseNet121 network to extract deep features from scleral images. The classification module applies MIL combined with a loss attention mechanism to determine image bag labels using extracted deep features, followed by classification. Finally, the prediction module evaluates CHD risk for subjects based on classified image bags.

Fig. 1.

Fig. 1

Overview of scleral image processing pipeline. (A) Acquisition equipment for scleral images taking. (B) Image preprocessing for data preprocessing. (C) Sclera segmentation to obtain enhanced images of sclera. (D) Feature extraction to take scleral characteristics associated with CHD. (E) Classify healthy people and CHD by multiple instance learning (MIL) and full connect neural network (FC). (F) CHD prediction according to the above steps.

Instrument and data collection

In order to collect a usable scleral image dataset for CHD, we designed and developed a specialized device28,29,33 for scleral image acquisition, as illustrated in Fig. 2A. This device is based on the shadow-free scleral imaging technology we developed. Common methods for capturing eye images usually use forward lighting and take pictures from the same direction, as shown in Fig. 2B. They usually light directly on the sclera, causing severe reflection and loss of scleral detail information. To avoid this problem, we developed a shadow-free imaging method for the sclera, as depicted in Fig. 2C. In the new method, Illumination light from the point light source is directed against the observation direction of the eyeball, we combined artificial intelligence optimization to improve the illumination angle and imaging light paths, making all reflected shadows from different eye structures (such as Cornea, Sclera, Iris, Ciliary Body, Choroid, etc.) can be focused into a point and coincides with the pupil, then achieve clear imaging of the sclera and iris without reflected light shadow interference .

Fig. 2.

Fig. 2

Schematic diagram of the sclera imaging acquisition device. (A) Actual drawing of instrument. (B) Schematic of common eye imaging. (C) Schematic of shadow-free scleral imaging. (D) Eye movement dynamics during imaging. (E) Representative acquired ocular image.

To capture comprehensive scleral information, images are captured in the frontal gaze as well as in four additional directions—up, down, left, and right (Fig. 2D). Both eyes are imaged, resulting in 8 images per participant (Fig. 2E is 4 of the images for the left eye, there are also similar 4 images for the right eye). During the acquisition process, participants place their chin on a chin rest, keep one eye close to the frame and gently hold their upper and lower eyelids with their fingers to fully expose the sclera. They are instructed to gaze toward the indicator light, and the entire procedure is completed within one minute. Sample images acquired through this process are shown in Fig. 2E. In this study, we focused on the characteristics of CHD on the sclera. Therefore, only 8 fully exposed images of the sclera were selected for the left and right eyes of each subject.

A total of 621 adult participants were recruited as volunteers for this study. The flow chart of the exclusion process is shown in Fig. 3. More than 5000 raw images were collected After data cleaning, which entailed the removal of blurry or incomplete images, data from 500 participants were retained, including 240 patients diagnosed with CHD by professional cardiovascular specialists. 4000 high-quality scleral images (8 per subject) from 500 participants were finally used for model training and evaluation. The data were collected at Dongzhimen Hospital of Beijing University of Chinese Medicine (Beijing, China) from April 2024 to November 2024. To reduce variability associated with different collection times and operators, the imaging device was equipped with an automatic color correction feature. During the hospital deployment phase, a standard color chart was captured under the hospital’s lighting conditions to serve as a calibration reference. After image acquisition, histogram equalization was applied to all images to mitigate system-induced data biases. Eight images were captured per participant, representing the sclera of both eyes in four gaze directions (up, down, left, and right), to form a single dataset entry. Participants with a CHD diagnosis were classified as positive cases, while others were classified as negative. This classification was based solely on the presence of CHD and did not account for the presence of other medical conditions.

Fig. 3.

Fig. 3

Flow chart of the exclusion process.

To evaluate classification performance, 80 out of 500 subjects were allocated to the validation set, while the remaining 420 subjects constituted the training set for classifier network optimization, maintaining a 5:1 ratio. Both the training and validation sets maintained the same ratio of healthy individuals to CHD patients. Subject-wise data partitioning was implemented by grouping all scleral images from each participant with corresponding labels, ensuring that all images from a given subject were assigned exclusively to either the training or validation set.

In total, more than 5000 raw ocular images were captured from 621 participants. After removing blurred or incomplete images, 500 participants were retained. For each participant, eight high-quality scleral images (left and right eyes in four gaze directions) were selected, yielding a final dataset of 4000 images for model development. The segmentation dataset (1607 training images and 321 testing images) described in “Sclera segmentation” section was collected independently for U-Net + + training and is not part of these 4000 images.

Image preprocessing

To enhance the accuracy and generalizability of classification models, we designed a image preprocessing pipeline consists of two steps.

The first step is to remove low-definition images for data cleaning. A subset of the collected scleral images was found to be blurry, which could negatively impact classification model performance. Therefore, it was necessary to filter and correct these images. Detecting blur is a fundamental challenge in computer vision. Commonly used blur detection algorithms include edge sharpness analysis, Bayesian discrimination functions, low depth-of-field (DOF) image segmentation, minimum directional high-frequency energy (for motion blur detection), and wavelet-based support vector machine (SVM) histogram methods. These algorithms extract blur-related features from input images based on established blur models for quantifying image blur. For instance, Chung et al.34 analyzed edge sharpness to identify blurry regions, determining that objects with lower edge sharpness appear blurred. Rugna et al.35 demonstrated that blurry regions exhibit greater invariance to low-pass filtering compared to non-blurry areas. Similarly, Ko et al.36 used statistical measures such as mean and standard deviation, finding that blurry regions tend to have lower values for both compared to clear regions. We adopted the blur detection method based on the Fast Fourier Transform (FFT). In an image, clear regions exhibit pronounced intensity variations, resulting in higher high-frequency components after Fourier transformation. In contrast, blurry regions exhibit smoother intensity transitions and are dominated by low-frequency components. The amplitude spectrum slope of blurry regions is steeper than that of non-blurry regions, enabling the distinction between blurry and clear images.

The second step involves the enhancement of scleral images. A review of relevant research and medical literature indicates that CHD exhibits distinct disease-related features in the eye37,38, such as radiating vessels at the lateral canthus39,40, haze-like patterns41, and colored spots42,43. These features are primarily concentrated in the sclera and are potentially significant for CHD classification using scleral images. However, these scleral features are often subtle, making them challenging to detect. To improve classification algorithm performance and highlight disease-related features, it is essential to enhance these features while mitigating the influence of varying illumination conditions. The enhancement process comprises two primary steps. First, a Gaussian blur is applied to the image to filter out high-frequency components, thereby reducing noise and suppressing fine details. Second, a weighted blending process is performed. Once Gaussian blurring suppresses the high-frequency components, the blurred image is subtracted from the original image, resulting in a residual “mask” that isolates the filtered high-frequency components. The mask is then blended with the original image using a weighted combination, enhancing high-frequency details for improved feature visibility. The enhanced scleral images reveal clearer structural details, such as pale pink blood vessel patterns and brown spots. Additionally, yellow-brown plaques, often imperceptible to the naked eye, become distinctly visible. For scleral images without prominent color features, the processed images do not exhibit noticeable artifacts or color distortions, maintaining enhancement robustness while preventing artificial distortions.

Sclera segmentation

As illustrated in Fig. 2, the images captured by our custom-designed imaging device include the sclera, iris, and periocular skin. Since our objective is to analyze pathological features in the sclera, isolating the scleral region is essential to remove irrelevant features. In this study, we employ U-Net++, an advanced medical image segmentation model, for scleral segmentation. U-Net + + is an enhanced version of the original U-Net architecture, originally introduced by Ronneberger et al.44. The U-Net model is particularly advantageous in medical image segmentation as it effectively integrates both deep and shallow features, facilitating precise segmentation with minimal labeled data. Zhou et al. introduced U-Net + + 45, which enhances feature fusion through dense skip connections and nested architectures, enabling adaptive depth adjustment. This improvement enhances accuracy and adaptability during segmentation, particularly in scenarios where varying levels of detail are required across different datasets.

U-Net + + comprises an encoder subnetwork (backbone) followed by a decoder subnetwork. U-Net + + differs from U-Net primarily in its redesigned skip connections (highlighted in green and blue) and the incorporation of deep supervision (indicated in red). To achieve superior scleral segmentation using the U-Net + + model, we employed an independent scleral segmentation dataset consisting of a demographically diverse cohort, including both genders and a wide age range. This dataset was collected and annotated specifically for segmentation purposes and was not derived from the 420 subjects described in “Instrument and data collection” section for CHD classification. A randomly selected subset of the dataset served as the test set to evaluate the segmentation network’s performance. The segmentation dataset (1607 training images and 321 testing images) was collected independently from the previous work for segmentation task training, and has no concern with these 4000 images mentioned in “Instrument and data collection” section. The dataset was split into 1607 images for training and 321 images for testing. The pixel-level annotations for the 1,607 training scleral images were generated by three medical master’s students from Beijing University of Chinese Medicine using the Any Labeling software. The annotation process involved three rounds of quality control. First, two annotators independently delineated the scleral region for each image. Second, a third annotator systematically reviewed their outputs: annotations with complete agreement were directly accepted, while discrepancies were flagged for further review. In cases of disagreement, the third annotator provided an assessment, and the first two annotators re-annotated the image accordingly until consensus was reached. This multi-round independent annotation and cross-review procedure ensured high accuracy and consistency of the pixel-level labels, thereby providing a reliable foundation for training the U-Net + + segmentation model. The model achieved an intersection over union (IoU) score of 0.907, highlighting its strong performance in precise scleral segmentation. Following these three preprocessing steps, we produced segmented scleral images with improved detail clarity.

Feature extraction

Traditional scleral feature extraction methods primarily focus on visually discernible features such as blood vessels, haze, and spots on the sclera. These features typically require manual annotation and are then detected using conventional image processing algorithms or deep learning networks. However, the relationship between these manually annotated features and CHD is often complex, frequently exhibiting one-to-many or many-to-one correlations. In contrast, deep features learned by CNNs exhibit a stronger and more direct correlation with CHD, leading to superior classification performance. In this study, we employ the DenseNet121 network for deep feature extraction. Proposed by Huang et al. in 201746, DenseNet121 incorporates dense connections between layers, allowing each layer to receive input from all preceding layers. This architecture facilitates efficient propagation of features and gradients, alleviates gradient vanishing issues, and enhances the learning of hierarchical features. Moreover, DenseNet minimizes model parameters and error rates, eliminates redundant feature learning, lowers model complexity, and improves classification accuracy. In this study, we choose DenseNet as the feature extraction model for depth information for its highest accuracy and F1 score, which was discussed in “Results” section. U-Net + + achieves high-precision segmentation of scleral boundaries by virtue of nested skip connections, ensuring the complete capture of tiny blood vessels and pigment structures. Through feature reuse and dense connections, DenseNet enhances its representational capability under the condition of limited samples, improves the efficiency of transmission and reuse of deep-level features, and thereby optimizes the classification performance.

In this study, DenseNet121 was first fine-tuned on the scleral dataset to extract robust deep feature representations. The extracted features were then fixed and used as inputs for the MIL classification stage.

Classification

Through this process, we collected ten ocular images from each subject, as illustrated in Fig. 2E. From these, we selected eight images that provided sufficient scleral exposure for segmentation and deep feature extraction, thereby minimizing the risk of information loss. Instead of assessing a subject’s health status based on the deep features of a single scleral image, we integrated information from all eight scleral images from each subject. This approach mitigates the effects of variability and potential confounders from individual images, enhancing the robustness and reliability of our analysis. To facilitate this, we reconstructed the dataset by aggregating the deep features from the eight scleral images of each subject. Each data package was labeled according to the subject’s health status: 0 for CHD cases and 1 for healthy individuals. This formulation represents a multi-instance binary classification problem, which we tackled through a MIL framework incorporating an attention mechanism.

For each subject, the deep features extracted from individual scleral images are fed into the MIL model to obtain the aggregated features of the entire image bag. The MIL model consolidates the feature vectors of all images within a given bag using an aggregation function. Conventional MIL approaches typically employ max-pooling or average-pooling to process deep features from multiple instances within a bag, ultimately determining the bag’s label. However, in this study, we employ the attention-based pooling mechanism proposed by Ilse et al.47 for aggregating bag-level features, which introduces an instance weighting mechanism where a neural network dynamically assigns weights. Let Inline graphicrepresent the set of instance embeddings within a bag. The MIL embedding is defined as follows:

graphic file with name d33e619.gif
graphic file with name d33e622.gif

W and V are network parameters.

The input to the MIL model comprises deep feature vectors extracted from individual scleral images, each with a dimension of 1024. The output is a single aggregated feature vector representing the entire image bag for a subject. This aggregated feature vector is then processed by a fully connected neural network to generate the final classification result for the bag. In this classification scheme, a label of 1 indicates a healthy individual, while a label of 0 corresponds to a CHD patient, as illustrated in Fig. 1. The MIL classifier with attention pooling was trained separately on the fixed feature representations obtained from DenseNet121, rather than in an end-to-end fashion.

Ethical approval and informed consent

We confirm that all methods were performed in accordance with the relevant guidelines and regulations. For more information about ethical approval and informed consent, please refer to ethics statement and the attachment.

Results

We applied the proposed algorithm to a dataset of ocular images acquired from patients with CHD and healthy individuals. To identify the optimal deep feature extraction network with respect to accuracy and computational efficiency, we conducted a comparative analysis of multiple CNNs. The evaluated models comprised the ResNet and VGG series, GoogleNet, and EfficientNet, all assessed as potential feature extractors for CHD detection.

We employed commonly used evaluation metrics, including accuracy, precision, recall, and the F1 score, to evaluate the performance of the proposed algorithm. Accuracy indicates the overall correctness of classification, precision represents the proportion of correctly predicted positive cases, recall measures the model’s capability to identify actual positive samples, and the F1 score balances precision and recall, mitigating class imbalance effects. To further analyze the model’s performance, we generated the Receiver Operating Characteristic (ROC) curve, a commonly utilized tool for assessing binary classification models. The ROC curve depicts the trade-off between the true positive rate (TPR) and the false positive rate (FPR) across varying threshold values. The area under the ROC curve (AUC) provides a quantitative assessment of model performance, with a higher AUC denoting superior classification capability. Using the ROC curve, we identified the optimal decision threshold via Youden’s J statistic and subsequently calculated the four evaluation metrics. The mathematical formulations of Youden’s J statistic, accuracy, precision, recall, and the F1 score are provided below, where TPR denotes the true positive rate, FPR denotes the false positive rate, TP signifies true positives, TN represents true negatives, FP corresponds to false positives, and FN indicates false negatives.

graphic file with name d33e642.gif
graphic file with name d33e645.gif
graphic file with name d33e648.gif
graphic file with name d33e651.gif
graphic file with name d33e654.gif

Momentum Stochastic Gradient Descent (MSGD) was chosen as the optimizer, with the momentum term set to 0.9 and an initial learning rate of 0.003. Across different experiments, the optimal learning rate was selected based on the minimum error. Additionally, cross-entropy was utilized as the loss function, and the batch size was set to 16. To balance convergence and training efficiency, each experiment was conducted for 100 epochs, providing adequate time for model stabilization. All deep learning experiments were conducted in PyTorch and executed on a single NVIDIA 4060 Ti GPU with 16GB of memory.

The test results of various models trained and evaluated on the same dataset are presented in Table 1; Fig. 4. As indicated in Table 1, substantial variations exist in performance metrics, such as AUC, among convolutional networks for sclera-based coronary heart disease classification. Among the evaluated models, DenseNet121 demonstrated the highest overall performance, achieving an average accuracy of 0.891, precision of 0.896, recall of 0.891, F1 score of 0.888, and AUC of 0.942, outperforming all other compared networks. In comparison, the highest average accuracy recorded for the VGG series was 0.873, for the ResNet series was 0.864, for AlexNet was 0.724, and for GoogleNet was 0.828.

Table 1.

Comparison results of different feature extraction networks.

Model Accuracy Precision Recall F1-score AUC
VGG13bn 0.873 0.890 0.873 0.866 0.912
VGG16bn 0.873 0.875 0.873 0.870 0.922
VGG19bn 0.855 0.862 0.855 0.848 0.867
ResNet18 0.842 0.841 0.842 0.838 0.864
ResNet34 0.828 0.828 0.828 0.828 0.851
ResNet50 0.864 0.864 0.864 0.861 0.898
AlexNet 0.724 0.715 0.724 0.716 0.691
GoogleNet 0.828 0.832 0.828 0.820 0.794
DenseNet121 0.891 0.896 0.891 0.888 0.942

Fig. 4.

Fig. 4

Comparison of ROC curves of different feature extraction networks.

Figure 4 depicts the ROC curves for all experiments, where a curve closer to the top-left corner and an AUC value near 1 signify superior model performance. A comparison of the nine models reveals that DenseNet121 exhibits the ROC curve nearest to the top-left corner, underscoring its superior sensitivity and specificity.

Discussion and conclusion

Traditional scleral imaging has been predominantly utilized for identity recognition, with limited exploration in disease detection. This study introduces a comprehensive processing and analysis algorithm to facilitate CHD detection using scleral images. The first step involves image clarity analysis, where the quality of collected eye images is assessed to eliminate blurred or unsuitable images, ensuring data cleanliness. The second step focuses on feature enhancement, applying Gaussian blur and low-frequency component removal to emphasize key features, particularly in the scleral region. The third step entails sclera segmentation, employing a pre-trained U-net + + model specifically designed for this task. The fourth step involves deep feature extraction, utilizing the DenseNet121 network, trained for sclera-based CHD classification, to extract deep features from individual scleral images. The fifth step encompasses feature aggregation and classification, where deep features extracted from scleral images of the same subject are aggregated into a bag and input into a MIL model with an attention mechanism. This model assigns weights to each instance and combines them into a single vector, which is then fed into a Fully Connected Classification Network (FCN) to derive the probability distribution of bag categories. The sixth step involves CHD risk prediction, where the subject’s CHD risk is estimated based on bag category probabilities. The proposed CHD risk prediction algorithm demonstrates strong performance across key metrics, including accuracy, recall, and AUC. Experimental results indicate that, compared to convolutional neural networks such as ResNet18, ResNet34, VGG13bn, and VGG16bn, DenseNet121 yields superior performance. Employing DenseNet121 as the feature extraction network markedly enhances model accuracy.

Grad-CAM (Gradient-weighted Class Activation Mapping)48, a widely adopted visualization technique for convolutional neural networks, was employed to analyze the deep features of the sclera in the context of CHD. This technique identifies the most influential regions of the input image for the model’s predictions by leveraging gradients of the target concept propagated to the final convolutional layer. Specifically, Grad-CAM produces a heatmap that highlights the relative importance of each spatial location in the image for the model’s decision-making process. This visualization enhances model interpretability by identifying the scleral image features that most significantly contribute to CHD classification. Representative samples from CHD patients are depicted in Fig. 5. The scleral features identified by the model are primarily localized to the scleral blood vessels, with secondary attention directed toward spots. These features exhibit a strong correlation with the pathophysiological mechanisms underlying CHD. For instance, systemic microcirculation disorders in CHD patients, including endothelial dysfunction and elevated blood viscosity, can induce dilation and stagnation of scleral capillaries, consistent with the TCM concept of “ocular collateral stasis.” Furthermore, oxidative stress linked to atherosclerosis may contribute to lipid deposition or pigmentation in the scleral connective tissue, aligning with the TCM concept of “phlegm turbidity clouding the clear orifices.” The strong concordance between the scleral features identified by the neural network and the clinical manifestations of CHD enhances the interpretability of this approach. This finding further highlights the potential of scleral imaging as a valuable auxiliary tool for disease diagnosis.

Fig. 5.

Fig. 5

Original sclera images of one CHD patient and corresponding visualized deep features via Grad-CAM.

Although the deep learning algorithm achieved high accuracy on the CHD scleral dataset collected in this study, several limitations remain and warrant further investigation. First, the dataset requires expansion to improve its representativeness and robustness. CHD and its closely associated condition, systemic atherosclerosis, involve complex pathological processes that impact multiple organ systems. Physiological changes in the sclera may be influenced by multiple internal organs, including the liver and kidneys. Consequently, expanding the dataset to include scleral data from a more diverse and extensive population is imperative. More critically, the current dataset defines positive cases exclusively based on the presence of CHD, without considering potential confounding factors such as comorbidities, specific medical conditions, regional geographic influences, or other relevant variables. Future research should prioritize the development of more comprehensive case classifications within larger, more diverse cohorts to improve generalizability. Furthermore, as an exploratory study, this work necessitates the integration of additional clinical data, such as detailed case histories, physiological indicators, and clinical manifestations. A more in-depth investigation into the pathological relationship between CHD and scleral changes is crucial for advancing our understanding and enhancing diagnostic accuracy.

This study aims to develop a non-invasive auxiliary diagnostic method for CHD and highlights two key contributions. First, the proposed method demonstrates exceptional classification performance, achieving an average AUC of 0.942. Second, it reveals that vascular abnormalities are the predominant interpretable features associated with the pathophysiological mechanisms of CHD, while pigmentation spots were identified in a subset of cases as auxiliary cues. These findings enhance the interpretability of the method and underscore the clinical relevance of monitoring scleral vascular changes in CHD patients, with pigmentation requiring further validation in larger and more diverse datasets. By incorporating scleral imaging into CHD auxiliary diagnosis, this study expands the scope of early disease detection and monitoring through scleral analysis, highlighting the potential of scleral imaging for clinical diagnostic applications. The integration of scleral image acquisition with artificial intelligence algorithms provides a promising non-invasive diagnostic approach with substantial research potential. One limitation of this study is that we did not perform ablation experiments on direct classification of raw scleral images using DenseNet alone, without segmentation or MIL. Such a baseline could provide additional insights into the specific contributions of the segmentation and MIL components. Future work will therefore include ablation studies to systematically evaluate the incremental value of each module in our proposed framework.

Author contributions

Yixuan Shi undertook the main work of coding, analyzing data and writing the manuscript. Jianing Wang reviewed the manuscript. Zhaoxuan Ding had key contribution in data collection. Chuxiang Gao and Machao Li did the main work on the device’s software. provided some suggestions, Li Ma and Xianbo Luo designed and built the hardware of the device. Dongsheng Wei, Jianing Wang and Hao Ma assisted in analyzing data and proved some suggestions on the algorithm. Yixuan Shi and Zhaoxuan Ding contributed equally to this work. Jiang Zhu, Xiaoqing Zhang, Guoliang Huang are the corresponding authors, guided the writing and research work of the thesis. All authors contributed to the article and approved the submitted version.

Funding

This study was funded by the National Key Research and Development Program of China [2022YFC3502301, 2022YFC3502300, 2023YFF0721501, 2018YFA0704000], National Natural Science Foundation of China [62375148, 61927819, 81827808, 62105177], Natural Science Foundation of Beijing Municipality [L246037], Tsinghua University.

Data availability

The original image data of this paper are available in part if you request it from the author. For more information, please contact Yixuan Shi by email (shiyx23@mails.tsinghua.edu.cn).

Declarations

Competing interests

The authors declare no competing interests.

Ethical approval

All procedures performed in studies involving human participants were in accordance with the ethical standards of the institutional and/or national research committee and with the 1964 Helsinki Declaration and its later amendments or comparable ethical standards. The study was reviewed and approved by the institution review board of Tsinghua University (NO.20170012). Written informed consent was obtained from all subjects and/or their legal guardian(s) and the informed consent form is attached.

Footnotes

Publisher’s note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Yixuan Shi and Zhaoxuan Ding have contributed equally to this work.

Contributor Information

Jiang Zhu, Email: jiangzhu@bistu.edu.cn.

Xiaoqing Zhang, Email: 202001003@bucm.edu.cn.

Guoliang Huang, Email: tshgl@mail.tsinghua.edu.cn.

References

  • 1.Tsao, C. W. et al. Heart disease and stroke statistics-2023 update: a report from the American Heart Association. Circulation147, E93–E621 (2023). 10.1161/cir.0000000000001123 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 2.Figal, D. P. et al. Comments on the 2021 ESC guidelines for the diagnosis and treatment of acute and chronic heart failure. Rev. Esp. Cardiol.75, 458–465. 10.1016/j.rec.2022.05.005 (2022). [DOI] [PubMed] [Google Scholar]
  • 3.McDonagh, T. A. et al. Focused Update of the 2021 ESC Guidelines for the diagnosis and treatment of acute and chronic heart failure: Developed by the task force for the diagnosis and treatment of acute and chronic heart failure of the European Society of Cardiology (ESC) With the special contribution of the Heart Failure Association (HFA) of the ESC. Eur. Heart J.44, 3627–3639 (2023). 10.1093/eurheartj/ehad195 [DOI] [PubMed] [Google Scholar]
  • 4.Heusch, G. Coronary blood flow in heart failure: Cause, consequence and bystander. Basic. Res. Cardiol.10.1007/s00395-022-00909-8 (2022). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5.Mahendiran, T., Collet, C. & De Bruyne, B. Coronary-artery autoregulation with increasing stenosis. N. Engl. J. Med.390, 2030–2032. 10.1056/NEJMc2402216 (2024). [DOI] [PubMed] [Google Scholar]
  • 6.Olthuis, S. G. H. et al. Endovascular treatment versus no endovascular treatment after 6–24 h in patients with ischaemic stroke and collateral flow on CT angiography (MR CLEAN-LATE) in the netherlands: a multicentre, open-label, blinded-endpoint, randomised, controlled, phase 3 trial. Lancet401, 1371–1380. 10.1016/s0140-6736(23)00575-5 (2023). [DOI] [PubMed] [Google Scholar]
  • 7.Yu, B. et al. The Eye-image features of patients with coronary heart disease assessed by the combination of traditional Chinese and Western medicine. (2023).
  • 8.Sideri, A. M. et al. Correlation between coronary and retinal microangiopathy in patients with STEMI. Transl. Vis. Sci. Technol.12, 8–8 (2023). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9.Huang, L. et al. Exploring associations between cardiac structure and retinal vascular geometry. J. Am. Heart Assoc. 9, e014654 (2020). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10.Rusu, A. C. et al. Retinal structural and vascular changes in patients with coronary artery disease: A systematic review and meta-analysis. Life14, 448 (2024). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11.Schuster, A. K. et al. Choroidal thickness is associated with cardiovascular risk factors and cardiac health: The Gutenberg health study. Clin. Res. Cardiol.109, 172–182 (2020). [DOI] [PubMed] [Google Scholar]
  • 12.Stone, P. H., Libby, P. & Boden, W. E. Fundamental pathobiology of coronary atherosclerosis and clinical implications for chronic ischemic heart disease management—the plaque hypothesis: A narrative review. JAMA Cardiol.8, 192–201 (2023). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.Lee, J. M. et al. Coronary flow reserve and microcirculatory resistance in patients with intermediate coronary stenosis. J. Am. Coll. Cardiol.67, 1158–1169. 10.1016/j.jacc.2015.12.053 (2016). [DOI] [PubMed] [Google Scholar]
  • 14.Van Gaal, L. F., Mertens, I. L. & De Block, C. E. Mechanisms linking obesity with cardiovascular disease. Nature444, 875–880. 10.1038/nature05487 (2006). [DOI] [PubMed] [Google Scholar]
  • 15.Kotova, J. A. et al. Markers of endothelial Damage, Inflammation, oxidative and cellular stress in patients with coronary artery disease and type 2 diabetes. Int. J. Biomed.10, 104–107. 10.21103/Article10(2)_OA2 (2020). [Google Scholar]
  • 16.Li, X. et al. Interplay of TNF-α, soluble TNF receptors and oxidative stress in coronary chronic total occlusion of the oldest patients with coronary heart disease. Cytokine (2020). 10.1016/j.cyto.2019.154836 (2020). [DOI] [PubMed] [Google Scholar]
  • 17.Yuan, R., Wang, J. & Guo, L. The effect of sleep deprivation on coronary heart disease. Chin. Med. Sci. J.31, 247–253 (2016). [DOI] [PubMed] [Google Scholar]
  • 18.Feng, J. et al. Altered heart rate variability depend on the characteristics of coronary lesions in stable angina pectoris. Anatol. J. Cardiol.15, 496–501. 10.5152/akd.2014.5642 (2015). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19.Carney, R. M., Freedland, K. E. & Veith, R. C. Depression, the autonomic nervous system, and coronary heart disease. Psychosom. Med.67, 29–S33. 10.1097/01.psy.0000162254.61556.d5 (2005). [DOI] [PubMed] [Google Scholar]
  • 20.Ji, S., Xu, W., Yang, M. & Yu, K. 3D convolutional neural networks for human action recognition. IEEE Trans. Pattern Anal. Mach. Intell.35, 221–231. 10.1109/tpami.2012.59 (2013). [DOI] [PubMed] [Google Scholar]
  • 21.de Margerie-Mellon, C. & Chassagnon, G. Artificial intelligence: A critical review of applications for lung nodule and lung cancer. Diagn. Interv. Imaging. 104, 11–17. 10.1016/j.diii.2022.11.007 (2023). [DOI] [PubMed] [Google Scholar]
  • 22.Javed, R. et al. Deep learning for lungs cancer detection: A review. Artif. Intell. Rev.10.1007/s10462-024-10807-1 (2024). [Google Scholar]
  • 23.Esteva, A. et al. Dermatologist-level classification of skin cancer with deep neural networks. Nature542, 115. 10.1038/nature21056 (2017). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 24.Maqsood, S. & Damasevicius, R. Multiclass skin lesion localization and classification using deep learning based features fusion and selection framework for smart healthcare. Neural Netw.160, 238–258. 10.1016/j.neunet.2023.01.022 (2023). [DOI] [PubMed] [Google Scholar]
  • 25.Nahiduzzaman, M. et al. Diabetic retinopathy identification using parallel convolutional neural network based feature extractor and ELM classifier. Expert Syst. Appl.10.1016/j.eswa.2023.119557 (2023).37274610 [Google Scholar]
  • 26.Poplin, R. et al. Prediction of cardiovascular risk factors from retinal fundus photographs via deep learning. Nat. Biomed. Eng.2, 158–164. 10.1038/s41551-018-0195-0 (2018). [DOI] [PubMed] [Google Scholar]
  • 27.Zhang, K. et al. Deep-learning models for the detection and incidence prediction of chronic kidney disease and type 2 diabetes from retinal fundus images. Nat. Biomed. Eng.5, 533. 10.1038/s41551-021-00745-6 (2021). [DOI] [PubMed] [Google Scholar]
  • 28.Huang, Q. et al. Machine learning system for lung neoplasms distinguished based on scleral data. Diagnostics13, 648 10.3390/diagnostics13040648 (2023). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 29.Zhang, Z. et al. Noninvasive and affordable type 2 diabetes screening by deep learning-based risk assessment and detection using ophthalmic images inspired by traditional Chinese medicine. Med-X1, 2 (2023). [Google Scholar]
  • 30.Babenko, B. et al. Detection of signs of disease in external photographs of the eyes via deep learning. Nat. Biomed. Eng.6, 1370. 10.1038/s41551-022-00867-5 (2022). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 31.Cheung, C. Y. et al. A deep-learning system for the assessment of cardiovascular disease risk via the measurement of retinal-vessel calibre. Nat. Biomed. Eng.5, 498. 10.1038/s41551-020-00626-4 (2021). [DOI] [PubMed] [Google Scholar]
  • 32.Dai, L. et al. A deep learning system for detecting diabetic retinopathy across the disease spectrum. Nat. Commun.12, 3242 10.1038/s41467-021-23458-5 (2021). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 33.Lv, W. Q. et al. Deep learning algorithm for automated detection of polycystic ovary syndrome using scleral images. Front. Endocrinol.10.3389/fendo.2021.789878 (2022). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 34.Chung, Y. C. et al. In IEEE Conference on Cybernetics and Intelligent Systems. 356–360 (2004).
  • 35.Da Rugna, J. & Konik, H. Automatic blur detection for meta-data extraction in content-based retrieval context. Proc. SPIE Int. Soc. Opt. Eng.5304, 285–294. 10.1117/12.526949 (2003). [Google Scholar]
  • 36.Ko, J. & Kim, C. & Etri. In 11th International Conference on Advanced Communication Technology. 1605–1610 (2009).
  • 37.McGeechan, K. et al. Meta-analysis: Retinal vessel caliber and risk for coronary heart disease. Ann. Intern. Med.151, 404–131. 10.7326/0003-4819-151-6-200909150-00005 (2009). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 38.Rusu, A. C. et al. Retinal structural and vascular changes in patients with coronary artery disease: A systematic review and meta-analysis. Life-Basel (2024). 10.3390/life14040448 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 39.Wu, R., Wang, J. J., Tai, E. S. & Wong, T. Y. Cardiovascular risk factors, inflammation, and corneal arcus: The Singapore Malay eye study. Am. J. Ophthalmol.150, 581–587. 10.1016/j.ajo.2010.04.028 (2010). [DOI] [PubMed] [Google Scholar]
  • 40.Matuleviciute, I. et al. Coronary artery disease, its associations with ocular, genetic and blood lipid parameters. Eye10.1038/s41433-023-02703-9 (2023). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 41.Chambless, L. E. et al. Corneal arcus with coronary heart-disease and cardiovascular-disease mortality. Am. J. Public. Health. 80, 1200–1204 10.2105/ajph.80.10.1200 (1990). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 42.Seidelmann, S. B. et al. Retinal vessel calibers in predicting long-term cardiovascular outcomes the atherosclerosis risk in communities study. Circulation134, 1328–. 10.1161/circulationaha.116.023425 (2016). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 43.Dong, X. Y. et al. Novel 2D/3D vascular biomarkers reveal association between fundus changes and coronary heart disease. Microvasc. Res.159, 104793 10.1016/j.mvr.2025.104793 (2025). [DOI] [PubMed] [Google Scholar]
  • 44.Ronneberger, O., Fischer, P. & Brox, T. In 18th International Conference on Medical Image Computing and Computer-Assisted Intervention (MICCAI). 234–241 (2015).
  • 45.Zhou, Z., Siddiquee, M. M. R., Tajbakhsh, N. & Liang, J. UNet++: A Nested U-Net Architecture for Medical Image Segmentation. Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support: 4th International Workshop, DLMIA 2018, and 8th International Workshop, ML-CDS 2018, Held in Conjunction with MICCAI 2018, Granada, Spain, S… 11045, 3–11 (2018). 10.1007/978-3-030-00889-5_1 [DOI] [PMC free article] [PubMed]
  • 46.Huang, G., Liu, Z., van der Maaten, L. & Weinberger, K. Q. & IEEE. In 30th IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). 2261–2269 (2017).
  • 47.Ilse, M., Tomczak, J. M. & Welling, M. In 35th International Conference on Machine Learning (ICML). (2018).
  • 48.Selvaraju, R. R. et al. Grad-CAM: Visual explanations from deep networks via gradient-based localization. Int. J. Comput. Vis. 128, 336–359. 10.1007/s11263-019-01228-7 (2020). [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Data Availability Statement

The original image data of this paper are available in part if you request it from the author. For more information, please contact Yixuan Shi by email (shiyx23@mails.tsinghua.edu.cn).


Articles from Scientific Reports are provided here courtesy of Nature Publishing Group

RESOURCES