Skip to main content
Patterns logoLink to Patterns
. 2025 Feb 4;6(2):101175. doi: 10.1016/j.patter.2025.101175

AI-assisted facial analysis in healthcare: From disease detection to comprehensive management

Chaoyu Lei 1,2,3,14, Kang Dang 4,5,14, Sifan Song 4,5,14, Zilong Wang 6,14, Sien Ping Chew 1,2,3, Ruitong Bian 1,2,3, Xichen Yang 1,2,3, Zhouyu Guan 7, Claudia Isabel Marques de Abreu Lopes 8, Mini Hang Wang 9, Richard Wai Chak Choy 9, Xiaoyan Hu 9, Kenneth Ka Hei Lai 9,10, Kelvin Kam Lung Chong 9,10,11,12, Chi Pui Pang 9,10,11, Xuefei Song 1,2,3,, Jionglong Su 4,∗∗, Xiaowei Ding 13,∗∗∗, Huifang Zhou 1,2,3,∗∗∗∗
PMCID: PMC11873005  PMID: 40041850

Summary

Medical conditions and systemic diseases often manifest as distinct facial characteristics, making identification of these unique features crucial for disease screening. However, detecting diseases using facial photography remains challenging because of the wide variability in human facial features and disease conditions. The integration of artificial intelligence (AI) into facial analysis represents a promising frontier offering a user-friendly, non-invasive, and cost-effective screening approach. This review explores the potential of AI-assisted facial analysis for identifying subtle facial phenotypes indicative of health disorders. First, we outline the technological framework essential for effective implementation in healthcare settings. Subsequently, we focus on the role of AI-assisted facial analysis in disease screening. We further expand our examination to include applications in health monitoring, support of treatment decision-making, and disease follow-up, thereby contributing to comprehensive disease management. Despite its promise, the adoption of this technology faces several challenges, including privacy concerns, model accuracy, issues with model interpretability, biases in AI algorithms, and adherence to regulatory standards. Addressing these challenges is crucial to ensure fair and ethical use. By overcoming these hurdles, AI-assisted facial analysis can empower healthcare providers, improve patient care outcomes, and enhance global health.

Keywords: artificial intelligence, facial analysis, healthcare, disease screening, global health

The bigger picture

Imagine a world in which a simple photograph of one’s face could offer insights that could assist doctors in detecting and managing certain diseases earlier. This possibility has emerged through the integration of artificial intelligence (AI) into facial analysis. Several health conditions leave subtle clues on our faces, indicating that AI can learn to recognize with remarkable accuracy. AI can help with early disease detection, guide treatment decisions, and track health over time by analyzing facial features. This approach has the potential to significantly improve patient care and public health worldwide. However, for this technology to be truly beneficial, we must address important issues, including data privacy, model interpretability, accuracy and fairness of AI algorithms, and compliance with digital health regulations. If these obstacles can be effectively resolved, then AI-driven facial analysis could emerge as a valuable tool in healthcare, offering advanced medical insights that are both accessible and impactful.


AI is changing healthcare by enabling facial analysis for health monitoring and disease screening and management. This review highlights the potential of AI-assisted facial analysis to detect subtle facial phenotypes linked to health disorders. While promising, adoption requires addressing key challenges, including privacy, interpretability, accuracy, bias, and regulation. By overcoming these hurdles, this technology can benefit patient care, offering a non-invasive, cost-effective, and user-friendly approach to improving global health outcomes.

Introduction

Human faces encompass a wealth of information about age, ethnicity, emotions, and health status.1 Distinct facial features can serve as indicators of various medical conditions and systemic diseases.2 For instance, Down syndrome can be identified by distinct features such as small eye fissures, a wide distance between the eyes, upward-slanting eyes, a flat facial profile, and a protruding tongue.3 However, the early detection of these phenotypic markers poses a significant challenge owing to the considerable variability in human facial features and the subtlety of some disease manifestations. This variability often necessitates the expertise and experience of healthcare professionals to make accurate assessments.4 Moreover, there is a pressing need for an effective, convenient, and cost-effective screening tool that can facilitate patient screening outside of hospital settings.5

In the contemporary healthcare era, marked by the integration of advanced digital technologies, artificial intelligence (AI) has catalyzed significant innovations.6,7,8 AI-assisted facial analysis is a promising new method for disease screening. This technology demonstrates remarkable potential in recognizing and interpreting facial attributes associated with various health conditions, offering a convenient and accessible preliminary screening method.9 While AI models offer valuable insights, it is essential to understand that they yield probabilistic assessments rather than definitive diagnoses. By providing a non-invasive, user-friendly, and cost-effective screening approach, AI-assisted facial analysis can contribute to early detection and prompt medical attention, particularly in low- and middle-income countries (LMICs).10 These regions often face a limited healthcare infrastructure and a shortage of medical professionals, leading to higher rates of medical complications and fatalities.11

AI-assisted facial analysis could bridge this gap by enabling earlier disease detection and management due to three key attributes. First, it is user-friendly and convenient, simplifying health assessments by utilizing facial images for analysis.4 Second, it is non-invasive and safe, prioritizes patient comfort, and minimizes risks by avoiding invasive procedures.12 Thirdly, it is cost-effective and affordable, requiring no expensive equipment or complex procedures, making it especially attractive for healthcare in resource-limited settings.13 These attributes establish AI-assisted facial analysis as a valuable tool in healthcare, facilitating comprehensive and continuous disease management. Notably, AI-assisted facial analysis is intended to complement professional medical evaluations, and its findings should be interpreted within the context of comprehensive clinical assessments. Moreover, ethical concerns regarding bias, fairness, and profiling remain critical.14,15 There is also a need for transparency, regulation, and alignment with societal expectations.16,17

Thus, we begin by outlining four distinct stages of a prototypical AI-assisted facial analysis technology pipeline (workflow of AI-assisted facial analysis technology). This is followed by a comprehensive examination of its main applications in disease screening (AI-assisted facial analysis for disease detection). Subsequently, we explore its potential to contribute to healthcare across other domains, including health monitoring, aiding treatment decision-making, and post-treatment follow-up, thereby facilitating comprehensive disease management (AI-assisted comprehensive management). However, these advances present technical and ethical challenges that demand careful consideration. We thoroughly and critically investigate these related issues and suggest potential future research directions (challenges and future directions). By providing a balanced perspective, we aim to highlight the potential challenges of AI-assisted facial analysis in healthcare.

Workflow of AI-assisted facial analysis technology

The convergence of computer vision and healthcare is changing the medical sector, particularly disease identification through AI-assisted facial analysis. This multifaceted process begins with face detection and alignment, potentially progressing to face reconstruction and ultimately culminating in face recognition (Figure 1). Each stage enhances the accuracy and efficiency of disease detection, thus shaping the future of healthcare.

Figure 1.

Figure 1

The workflow of AI-assisted facial analysis in disease detection

There are four distinct stages of a prototypical AI-assisted facial analysis technology pipeline: face detection, face alignment, face reconstruction, and face recognition.

Face detection

Face detection is a critical component of facial processing applications. This involves locating and identifying faces in images or videos by filtering background noise and prioritizing faces as the main focus. Face detection techniques can be categorized into traditional methods and deep learning-based strategies. Traditional methods, such as those found in the OpenCV and Dlib libraries, utilize algorithms such as Haar cascades18 and Dlib’s histogram of oriented gradients with linear support vector machines.19

Deep learning methods significantly outperform traditional techniques. Convolutional neural network (CNN) cascades,20 multitask cascaded CNNs,21 and two-stage detectors, such as region-based CNN,22 represent some of the advances in this field. Additionally, anchor-free techniques like CenterFace23 and one-stage detectors like Single Shot MultiBox Detector24 have been developed for real-time detection and robust performance.

Face alignment

Face alignment or facial landmark localization is essential for highlighting anatomical landmarks on the face. Techniques for facial landmark localization are primarily divided into heatmap regression and coordinate regression. Heatmap regression networks like Hour-Glass and High-Resolution Network25 predict likely landmark locations, while coordinate regression uses streamlined CNNs to precisely pinpoint facial landmarks.26 Enhanced methods such as HRNetv2,27 Practical Fast Landmark Detector,28 and Deep Adaptive Graph Network29 have further improved the accuracy of facial landmark localization, opening new opportunities.

Face reconstruction

The depth and contours of facial structures are critically important. Disease screening based on facial features can be limited to a two-dimensional (2D) perspective. Therefore, three-dimensional (3D) face reconstruction is utilized to generate 3D facial models from 2D images. Recent breakthroughs in single-image 3D face reconstruction have shown promising results.30 For instance, direct volumetric CNN regression involves a straightforward CNN that transforms 2D images into volumetric depth maps.31 Multi-image 3D face reconstruction leverages multiple 2D images to aggregate shape information, handling occlusions and significant pose variations better than single-image reconstruction techniques.32,33

Face recognition

The final stage in the AI-assisted facial analysis process is face recognition, which confirms an individual’s identity using data obtained from the previous stages. Beyond verification and identification, face recognition is also useful for attribute recognition, making it valuable in detecting a wide range of diseases.34 Advancements in network architectures have transitioned from early models such as AlexNet35 to lighter, more recent networks such as SENet.36 Improvements in loss function design, which are crucial for training deep neural networks, have also been noted. Techniques like CurricularFace37 enhance the performance of face recognition systems.

Face attribute recognition plays a key role in disease detection by identifying specific facial attributes. Multitask networks such as ANet38 and Partially Shared Multi-task Convolutional Neural Network with Local Constraint39 have been proposed for simultaneous attribute prediction, enabling the identification of diseases. Future improvements in AI-assisted facial analysis may enable earlier and more precise screening and advance healthcare outcomes.

AI-assisted facial analysis for disease detection

AI-assisted facial analysis has been used to screen a broad spectrum of diseases, as illustrated in Figure 2 and Table 1. For diseases with noticeable facial alterations, where diagnosis relies heavily on facial images, such as skin diseases, and where AI facial analysis achieves high accuracy, this tool provides valuable diagnostic support. In cases where facial changes are superficial indicators of underlying conditions such as cardiovascular disease, AI-assisted facial analysis serves as an effective screening tool to facilitate early detection. Our aim is to explore the potential of this technology to aid early disease detection, which could greatly improve patient health outcomes. This is especially beneficial for LMICs because (1) it is widely accessible, therefore enhancing global health equity; (2) it promotes early detection and improves chances of recovery; and (3) it can be used for self-monitoring and for increasing healthcare access.

Figure 2.

Figure 2

Diseases detectable by AI-assisted facial analysis

Diseases detectable by AI-assisted facial analysis include skin diseases, neuropsychiatric diseases, ophthalmic diseases, genetic diseases, endocrine diseases, cardiovascular diseases, hematological diseases, and digestive diseases.

Table 1.

AI facial analysis for detecting diverse diseases

Study and year Disease Countries Data type No. of images Compared with clinicians Performance
Genetic diseases

Porras et al.40 genetic syndromes Argentina, Australia, Belgium, and 26 other countries 2D image (frontal) 2,800 no accuracy: 0.88; sensitivity: 0.90; specificity: 0.86
Pan et al.41 Turner syndrome China 2D image (frontal) 1,281 no AUC: 0.95, 0.97, 0.96 (three scenarios); sensitivity: 0.97; specificity: 0.97
Yang et al.42 Noonan syndrome China 2D image (frontal) 420 yes AUC: 0.98; accuracy: 0.92
Latorre-Pellicer et al.43 Cornelia de Lange syndrome Spain 2D image (frontal) 49 no accuracy: 0.98
Mishima et al.44 congenital dysmorphic syndromes Japan 2D image 108 no accuracy: 0.86

Skin diseases

Han et al.45 skin cancer South Korea 2D image (frontal, side, and diagonal lines) 182,348 yes AUC: 0.91; specificity: 0.77; sensitivity: 0.91
Wen et al.46 acne China 2D image (half-face) 1,222 no mean average precision: 0.54
Binol et al.47 rosacea lesions US 2D image (frontal, half-side, looking up) 41 no Dice coefficients: 0.90, 0.88 (two models)
Peng et al.48 non-neoplastic facial pigmentation disorders Singapore 2D image (site of the facial lesion) 150 no accuracy: 0.92

Ophthalmic diseases

Karlin et al.49 TED US 2D image (frontal) 2,166 yes accuracy: 0.89; sensitivity: 0.93; specificity: 0.87; precision: 0.80; F1 score: 0.86
Huang et al.50 thyroid-associated ophthalmopathy China 2D image (front, left, right, and nine eye positions) 21,840 no AUC: 0.85; sensitivity: 0.80; specificity: 0.79; F1 score: 0.80
Lei et al.51 Graves’ orbitopathy China 2D image (frontal) 943 no AUC: 0.85
Hung et al.52 ptosis US, China (Taiwan) 2D image (eye images) 832 yes AUC: 0.986; sensitivity: 0.92; specificity: 0.88
Lou et al.53 ptosis China 2D image (eye images) 135 yes the ICCs between manual and automated measurements ranged from 0.934 to 0.971 (p < 0.01).
Tabuchi et al.54 ptosis Japan 2D image (eye images) 1,276 yes AUC: 0.90; accuracy: 0.83; sensitivity: 0.83; specificity: 0.83
Hui et al.55 benign and malignant eyelid cancer China 2D image (frontal) 345 yes AUC: 0.97; accuracy: 0.89; sensitivity: 0.93; specificity: 0.86
Chen et al.56 16 ophthalmic disorders China videos 3,652 no AUC: 0.843, 0.859 (two settings)

Endocrine diseases

Kong et al.57 acromegaly China 2D image (frontal) 1,365 no sensitivity: 0.96; specificity: 0.96
Kong et al.58 acromegaly China 2D image and 3D video frames 2,148 yes accuracy: 0.91
Wei et al.59 Cushing’s syndrome, acromegaly China 2D image 14,544 yes AUC: 0.96 (for both diseases)

Neuropsychiatric diseases

Ali et al.60 PD US video (10–12 s) 1,812 no AUC, 0.94; accuracy: 0.96; precision: 0.96; sensitivity: 0.94; F1 score: 0.95
Alam et al.61 ASD 2D image 2,840 no accuracy: 0.95
Zhou et al.62 MDD video (25 min) 150 no mean absolute error/root-mean-square error: 6.20/8.28, 6.21/8.39 (for two datasets)

Other diseases

Xiao et al.63 hepatobiliary diseases China 2D image (ocular image) 3,550 yes AUC: 0.74; sensitivity: 0.64; specificity: 0.73; F1 score: 0.68
Lin et al.64 CAD China 2D image (frontal, profile, and top views) 6,229 yes AUC: 0.73; sensitivity: 0.80; specificity: 0.54
Zhang et al.65 anemia China video (15 s) 316 yes AUC: 0.84; accuracy: 0.82

ICC, intraclass correlation coefficient.

Genetic disease

Advances in AI have enabled the recognition of facial features associated with genetic diseases. One study performed an analysis of 2,800 facial photographs from 1,400 children encompassing 128 different genetic diseases and 1,400 matched controls.40 Their AI model achieved an average accuracy of 88% for detecting these genetic diseases. This efficient technique offers a more accessible and affordable option for the early risk assessment of specialist referrals. Turner syndrome often presents with distinct facial deformities, such as a highly arched palate, low posterior hairline, and micrognathia. Researchers have developed a face recognition system specifically for detecting Turner syndrome.41 Similarly, Noonan syndrome, characterized by rapid eye movement and ptosis, can also benefit from AI-based screening.42 Furthermore, a smartphone app called Face2Gene was evaluated for screening Cornelia de Lange syndrome, and the correct detection was listed as the first prediction for 83.7%.43 Additionally, a study in Japan demonstrated the efficacy of Face2Gene in aiding clinical specialists in the identification of congenital dysmorphic syndromes.44 These tools offer accessible and affordable options for preliminary screening, potentially leading to earlier specialist referrals and interventions.

Skin disease

Many skin conditions require visual detection, which is subjective and prone to error.66 Screening skin diseases in resource-limited areas is challenged by a lack of dermatologists and screening tools. The application of AI to the identification of skin diseases on the face has seen remarkable advancements. One study implemented region-based CNNs to determine the presence of skin tumor lesions in images and classified them as benign or malignant.45 Their research employed a dataset comprising 182,348 training images, 2,844 validation images, and 325 test images. The accuracy of the algorithm was comparable with that of the dermatologists (F1 score: 0.831 vs. 0.835; Youden index score: 0.675 vs. 0.671). Acne vulgaris is a prevalent skin disorder that primarily affects adolescents and is characterized by papules, nodules, and facial cysts. Researchers have devised CNNs that were subsequently integrated into a free WeChat mini program to screen and classify the severity of acne.46 Rosacea is a treatable but chronic condition characterized by facial flushing, erythema, telangiectasia, and edema. A computer-aided screening system known as Ros-NET was developed to effectively identify rosacea lesions.47 Additionally, the detection of non-neoplastic skin pigmentation disorders was addressed through deep learning models.48 Therefore, AI-assisted facial analysis can improve the accuracy of detecting several skin diseases, which is expected to facilitate early intervention.

Ophthalmic disease

Ophthalmic disease management in low-resource settings is constrained by insufficient ophthalmic services and screening equipment.67 Many eye conditions, such as orbital and eyelid diseases, exhibit changes in the ocular and facial regions, providing a foundation for the application of AI-assisted facial analysis.68,69 Thyroid eye disease (TED) is the most common orbital disease and is characterized by exophthalmos, strabismus, and eyelid retraction.49 An ensemble neural network model was developed, using a dataset of 1944 photographs for training, 344 for performance metric calculations, and 222 for assessment.49 The model demonstrated an accuracy of 0.892. Additionally, the ResNet-50 model was employed to detect thyroid-associated ophthalmopathy based on external ocular images,50 and facial expressions were also explored to detect Graves’ orbitopathy.51 Ptosis can affect vision and is a symptom of ocular myasthenia. AI-based systems capable of automatically screening ptosis and measuring eyelid morphology using external ocular images were developed.52,53 The iOS app was devised to detect ptosis via smartphone-captured facial images.54 Eight CNNs were employed to automatically detect benign and malignant eyelid tumors from facial images, achieving an area under the curve (AUC) of 0.966.55 Furthermore, a smartphone-based deep learning system was developed to detect visual impairment in young children.56 In summary, AI-assisted facial analysis offers a promising solution for the early detection of eye diseases, especially orbital and eyelid diseases.

Endocrine disease

Endocrine diseases often go undiagnosed due to the lack of routine screening and awareness, with delayed diagnosis leading to severe complications.70 Patients with Cushing’s syndrome and acromegaly exhibit distinct facial characteristics. One study focused on the detection of acromegaly with facial features, including an enlarged nose, prominent mandibular and zygomatic arches, thickened lips, and facial soft tissue swelling. Various machine learning techniques were harnessed to identify acromegaly using facial images,57 achieving a sensitivity of 96% and a specificity of 96%, thereby facilitating automated acromegaly detection. In subsequent research, a novel deep learning-based model was developed to not only screen acromegaly but also classify its severity, expanding the scope of AI facial analysis.58 AI’s impact extends to other endocrine disorders, including Cushing’s syndrome, with an estimated incidence of hypercortisolism being 3 per million.59 A deep learning-based model was proposed for identifying Cushing’s syndrome and acromegaly from facial images, demonstrating an AUC of 0.96 for both diseases.59 The results were comparable in sensitivity and specificity to those of a physician. Hence, AI-assisted facial analysis can play a crucial role in identifying the signs of endocrine disorders, enabling prompt management and improving long-term health outcomes.

Neuropsychiatric disease

The shortage of mental health professionals and prevailing stigma significantly hinder the early diagnosis of neuropsychiatric disorders in resource-constrained settings.71 Conditions such as autism spectrum disorder (ASD), major depressive disorder (MDD), and Parkinson’s disease (PD) often display symptoms of reduced or abnormal facial expressions, rendering them suitable for detection through AI-assisted facial analysis. One study used a dataset of 1,812 videos from 604 individuals to develop a model for PD detection, achieving an accuracy of 0.956, precision of 0.958, and AUC of 0.940.60 Similarly, the efficacy of AI-assisted facial analysis in accurately detecting ASD and MDD was also demonstrated.61,62 These investigations underscore the potential of AI-assisted facial analysis for enhancing and expediting the early identification of neuropsychiatric disorders.

Other diseases

AI-assisted facial analysis is promising for the detection of other conditions. One study utilized external ocular images for the automatic detection of hepatobiliary diseases, achieving an AUC of 0.74 (95% confidence interval 0.71–0.76).63 Notably, among the included hepatobiliary diseases, the model demonstrated the best performance for liver cancer and liver cirrhosis. Moreover, the feasibility of employing facial images to detect coronary artery disease (CAD) was confirmed through deep learning models.64 The algorithm for detecting CAD exhibited a sensitivity of 0.80 and a specificity of 0.54 in the test group. Additionally, multiple models capable of predicting anemia were developed.65 The accuracy of the image level was 82.37%, and the AUC of the image level was 0.84. In summary, the integration of AI-assisted facial analysis presents a promising avenue for simplified and efficient detection of diverse medical conditions.

AI-assisted comprehensive management

AI-assisted comprehensive management extends beyond disease screening to health monitoring, treatment decision-making, and follow-up. It aims to provide a holistic approach to disease management (Figure 3). Although AI has yet to be applied to every stage of disease management for a single disease, there are applications for each stage across various diseases.

Figure 3.

Figure 3

Comprehensive AI-assisted disease management

This includes five aspects: health monitoring, early screening, disease diagnosis, treatment decision-making, and follow-up.

Empowering health monitoring

AI-assisted facial analysis can also be used for health monitoring. Personal attributes such as age, sex, and physique can be identified from facial images. For instance, the GRA Net utilizes facial images to predict age and sex, aiding in detecting premature facial aging.72 Xia et al. developed CNNs trained on non-invasive 3D facial images of 5,000 Han ethnic Chinese individuals, achieving an average difference between predicted and chronological/perceived age of ±2.8 and 2.9 years, respectively.73 ConstitutionNet was proposed to classify different constitutions in traditional Chinese medicine based on facial images, assisting in determining the most effective personalized conditioning program for individuals.74 Tay et al. developed models for real-time monitoring of nutritional status, enabling clinicians to remotely monitor patients and deliver necessary interventions.75 These applications provide individuals and healthcare providers with valuable information for maintaining health and preventing disease.

Aiding treatment decision-making

Efforts have been made in preoperative planning and postoperative outcome assessments in treatment decision-making. For preoperative planning, Van Brummen et al. facilitated the measurement of the marginal reflex distance (MRD), which is used to evaluate eyelid and soft tissue position.76 Additionally, a specialized smartphone application for MRD and levator muscle function measurement was developed to enhance user convenience. Song et al. introduced a clinical decision model to aid in making surgical decisions based on multidimensional facial imaging data.77 For pain assessment, ResNet-18 was trained to assess pain intensity based on facial expressions, assisting vulnerable groups in objectively assessing pain intensity and helping doctors tailor treatments by gauging pain responses.78

AI-assisted facial analysis is valuable for evaluating postoperative outcomes. Lou et al. devised a method to automate the measurement of eyelid morphology before and after surgery, minimizing human error.53 Bahçeci Şimşek and Şirolu utilized the DLIB-ML toolkit, which contains various image processing and machine learning power detectors, to create an automated approach for evaluating the outcome of Muller’s muscle-conjunctival resection surgery.79 Hidaka et al. designed an esthetic evaluation tool for mandibular reconstruction based on facial images.80 Similarly, Zhai et al. developed the Beautynet system to preoperatively predict facial beauty outcomes, enabling personalized surgery plans.81 Thus, AI-assisted facial analysis provides more objective and efficient evaluations compared to manual assessments, producing compelling evidence to aid therapeutic implementation.

Facilitating disease follow-up

AI-assisted facial analysis has the potential to greatly enhance disease follow-up by providing detailed and objective information regarding a patient’s progress and response to treatment. AI systems track changes in a patient’s facial features and expressions by comparing facial images captured at different intervals. For example, Hossain and Muhammad proposed an e-healthcare framework in which mobile devices such as smartphones capture images and medical data, which are then sent to the cloud for processing.82 The results are shared with healthcare professionals, and in cases of negative emotions such as pain, caregivers can be alerted for patient support. In dermatology, AI facial technology can monitor the progress of skin conditions, such as acne, eczema, or psoriasis.83 This enhances patient engagement and can improve adherence to treatment plans.

To conclude, AI-assisted facial analysis enables effective disease management. Individuals can use the AI-based software on their smartphones to monitor their health. If potential health issues are detected, the software identifies patients who may require specialist care, thereby facilitating early intervention. Doctors can also use this technology after diagnosis, and AI models can aid in creating treatment plans and evaluating surgical results using patient photographs. After treatment, patients can upload their facial images to the software for follow-up, which enables doctors to collect data and track recovery. Therefore, AI-assisted facial analysis provides an integrated solution that combines technology, healthcare, and patient well-being.

Challenges and future directions

Although AI-assisted facial analysis offers significant opportunities, challenges must be addressed to ensure its effective and ethical implementation. These include privacy concerns, unverified accuracy, weak interpretability, model unfairness, and regulatory adherence (Figure 4). In this section, we delve into each challenge, propose viable strategies to address them, and highlight the importance of future research. Such measures are crucial not only for the integration and acceptance of technology in healthcare environments but also for building trust among patients and practitioners. Collaborative efforts among technologists, clinicians, ethicists, and policymakers are vital for addressing these challenges and harnessing the full potential of AI-assisted facial analysis in healthcare.

Figure 4.

Figure 4

Challenges and future directions

They include privacy concerns, unverified accuracy, weak interpretability, model unfairness, and regulation adherence.

Privacy concerns

Privacy concerns are critical challenges when deploying AI-assisted facial analysis for disease detection. The collection, storage, and analysis of facial data raises significant concerns about the confidentiality and security of individuals’ personal information.84 Without stringent data protection measures, there is a risk of unauthorized access to sensitive health information, which can lead to misuse, discrimination, and a breach of trust between patients and healthcare providers.85 Malicious individuals may exploit user profiles on social media platforms to develop applications capable of identifying potential health risks without user consent. This could result in the unwarranted disclosure of private health information and expose users to targeted advertising for medical treatment or services, infringing upon their privacy and autonomy. Such activities may also lead to psychological distress or social stigma among affected individuals. This issue is compounded in areas where legal frameworks for data protection are weak or nonexistent, thereby increasing the vulnerability of individuals to privacy violations.

It is important to balance these concerns with the urgent need for AI-assisted facial analysis and the substantial value it brings to healthcare. Privacy concerns must be considered when deploying AI-assisted facial analysis. First, robust informed consent is critical.86 Individuals must have a clear understanding of what consent entails, the associated risks of sharing their facial data, and their rights regarding the use and disclosure of disease detection. Second, federated learning enables AI models to train on decentralized medical data without requiring the data to be centralized, thereby mitigating the risk of data breaches.87,88 Building upon this, swarm learning extends this concept by combining edge computing and blockchain-based peer-to-peer networks, enabling global collaboration on medical data while preserving privacy and eliminating the need for a central coordinator.89 Third, leveraging privacy-enhancing technologies throughout the AI life cycle is recommended. During training, differential privacy is employed to prevent the model from memorizing individual patient data while still allowing it to learn generalizable patterns.90 During inference, auxiliary patient data are encrypted before transmission to the model and decrypted and processed within a secure, isolated environment. Fourth, homomorphic encryption can be utilized to allow data to be processed in encrypted form, ensuring that data privacy is maintained during analysis.91 Furthermore, adopting face anonymization techniques can help maintain the utility of AI-assisted facial analysis for disease detection while safeguarding individual identities.92 It is important to note that, while these techniques protect data during training, they may not address all privacy concerns in deployment, particularly when making individual health inferences. In addition, raising public awareness about the potential misuse of facial data can empower users to take precautions when sharing images online. Collaborative efforts among governments, technology companies, and international organizations are essential for developing and enforcing standards that prevent the exploitation of AI technologies for malicious purposes.

Unverified accuracy

The unverified accuracy of AI-assisted facial analysis for disease detection in low-resource settings presents a complex challenge closely tied to the complexities of datasets and the shortage of comprehensive and diverse data crucial for developing effective AI models. These settings often lack the advanced imaging technology and infrastructure needed to gather, store, and process the extensive, high-quality data required.93 Existing datasets are usually small and not representative of the wide range of racial backgrounds or medical contexts needed for AI to be truly effective.94 Additionally, the variability in real-world data capture conditions, including differing camera specifications, poor lighting conditions, blurred facial images, and deviations in facial angles, further complicates the dataset and impacts the precision and reliability of AI models.67 Thus, the accuracy rates reported in the literature may not translate to real-world settings. This variability renders it particularly challenging to deploy these technologies in low-resource areas to broaden access to healthcare facilities through mobile AI-assisted facial analysis applications.52

To overcome the challenge of low model accuracy owing to dataset complexities and data shortages, a multifaceted strategy is required. Enhancing dataset diversity and volume through international collaboration and inclusive data collection efforts is fundamental.95 Engaging a broad spectrum of participants from various ethnic backgrounds and medical conditions can enrich datasets, ensuring that AI models are trained and validated on comprehensive and representative data. Embracing advanced learning methodologies, such as transfer learning,96 few-shot learning,97 and self-supervised learning,98 can mitigate the impact of limited data by enabling models to learn effectively from small datasets. Other algorithmic solutions, such as domain adaptation and robustness enhancement strategies, can further refine model performance under diverse conditions.34,99 However, these methods may not fully compensate for the fundamental limitations in data representativeness and may introduce new uncertainties. Standardization of data acquisition processes is another critical measure that involves establishing uniform protocols and quality control mechanisms. For images with suboptimal quality, sequential enhancement techniques can be employed to improve image fidelity. Techniques such as facial image relighting adjust lighting conditions while preserving facial features, mitigating the effects of poor illumination.100 Deblurring algorithms can reduce or eliminate motion blur by learning from large datasets of clear and blurred images, thereby restoring image clarity.101 Facial image registration, also known as face normalization, calibrates images with angular deviations to align faces directly with the camera, providing a more comprehensive view of facial features.102 By implementing these solutions, the dual challenges of dataset complexity and data shortage can be addressed. Finally, AI research incorporating images captured by patients or users using smartphones as external test datasets is highly recommended. Including such real-world data can better assess the robustness and generalizability of models in practical applications, ensuring that they perform effectively when deployed outside of controlled environments.

Weak interpretability

Weak interpretability in AI models, especially in facial image analysis for disease detection in low-resource settings, poses a significant challenge to the acceptance and trust of AI-driven screening.103 Advanced techniques, such as class activation mapping (CAM) and attention-based mechanisms, despite providing visual insights, often fail to offer medically comprehensible explanations, especially without clear causal links between facial features and diseases. Medicine clearly understands the facial changes that many skin and genetic diseases cause. However, for conditions such as coronary heart disease, better explanations from medical mechanisms are needed; otherwise, relying solely on detection accuracy makes it difficult for healthcare professionals and patients to fully understand and trust AI decisions.104,105 The lack of algorithmic transparency not only hinders clinical integration but also limits the potential of AI to contribute to medical research by obscuring insights into disease pathophysiology.106,107 Consequently, enhancing interpretability is essential for building confidence in AI within healthcare, ensuring that AI tools are both reliable and comprehensible, particularly in settings where the need for accessible and understandable screening solutions is critical.

To overcome the challenge of weak interpretability in AI models, it is crucial to develop explainable AI algorithms.108 Enhancements in CAM and attention-based mechanisms should aim not only to provide visual evidence of AI’s conclusions but also to translate these insights into explanations that are clear and meaningful for healthcare professionals and patients.109 This involves collaborative efforts between AI developers, data scientists, and medical experts to refine these techniques, ensuring that they yield transparent and clinically relevant outcomes. Additionally, investing in educational programs that equip healthcare professionals with the knowledge to understand and interpret AI-driven disease detection can foster trust and acceptance in clinical settings.110 This includes exploring why diseases exhibit characteristic manifestations on the face by investigating at the anatomical, pathophysiological, and other mechanistic levels. Not only can this better explain AI facial analysis in disease detection from a medical perspective, but it may also serve as a window to discover new pathogeneses of diseases. Such initiatives could significantly improve the usability of AI in healthcare, particularly in low-resource environments, by providing tools that are reliable and easily integrated into clinical workflows.

Model unfairness

Model unfairness in AI-assisted facial analysis for disease detection arises when AI models perform differently across demographic groups due to biases in their training datasets, a situation exacerbated by a lack of diversity in ethnicity, sex, and age.111 This can result in reduced accuracy for underrepresented populations, exacerbating health disparities and leading to potentially mistaken or missed detection.104 This issue is highlighted by the remarkable differences in facial characteristics among ethnic groups, suggesting the need for more inclusive datasets and the creation of ethnicity-specific datasets to improve screening accuracy.112,113 Furthermore, AI products, often designed in East Asia and Western contexts, may not effectively translate to low-resource environments without adjustments to align with local healthcare needs.114 This reliance on data from high-resource areas overlooks the diverse disease spectrum and genetic diversities in less affluent regions, potentially exacerbating health inequities and limiting the global applicability of AI in healthcare.

To address these issues, developing AI systems with built-in fairness mechanisms is crucial. Despite the evolution of advanced AI models, such as large language models,115 generative pre-trainedtransformers,116,117 and Segment Anything Model,118 which have benefitted from extensive training on big data, persistent biases underscore the necessity of international collaboration among healthcare entities. This collaboration aims to cultivate diverse datasets with significant representation from low-resource settings to address the critical challenge of bias and ensure equitable, unbiased, and effective global healthcare solutions. To address this, leveraging larger models alongside varied datasets has emerged as a promising approach to mitigate biases and foster fairness within AI-based disease screening.119 Furthermore, contemporary research is delving into strategies such as unbalanced training,120 attribute obfuscation,121 and domain adaptation.118,122 Nevertheless, eliminating bias remains challenging, and ongoing efforts are required to understand and address the underlying causes of model disparities.

Regulation adherence

The challenges of regulatory adherence in deploying AI-assisted facial analysis for disease detection highlight this complex landscape. The swift evolution of AI technologies often surpasses the pace at which regulatory frameworks are established, creating a gap in ensuring the ethical and fair use of the technology.123 In addition, regulations vary across jurisdictions, creating complexities in ensuring compliance, especially in international applications. The unique concerns of AI-assisted facial analysis, such as privacy and consent, also require stringent data governance frameworks that are challenging to implement, especially in low-resource environments.124 Furthermore, the necessity for AI models to continuously learn and adapt to new data complicates adherence, as models may deviate from their original validation and require ongoing oversight.125

Solving the challenges of regulatory adherence involves several key steps. First, creating dynamic regulatory frameworks that can quickly adapt to new technological advancements ensures that guidelines remain relevant and enforceable as AI evolves.126 This requires close collaboration between technology developers, regulators, and ethicists to anticipate future trends and address them proactively. Second, establishing international standards for AI use in healthcare can help manage differences across jurisdictions, ensuring a more uniform approach to ethical considerations and privacy protection.127 Recently, Annex III of the EU AI Act specified high-risk AI systems, including those for biometric identification.128 While facial data-based disease detection AI is not explicitly listed, such systems must still rigorously protect data privacy and security. Title II, Article 5, section (dc) limits emotion recognition systems, although medical use is exempted. Therefore, facial data, which are rich in biological and emotional information, require careful consideration in all applications.

Additionally, enhancing education and awareness of data privacy and consent, especially in low-resource settings, is crucial. This could involve training programs for healthcare workers and public awareness campaigns to inform patients about the rights and implications of AI in their care. It is encouraging to see regulatory frameworks being initiated, including the GDPR (2018) and AI Act (2021) in the EU as well as China’s Data Security Law and Personal Information Protection Law (2021).129 However, the underlying principles remain subject to ongoing debate and are influenced by cultural contexts. Therefore, establishing robust data governance, along with continuous monitoring and reevaluation of AI systems, will be essential for maintaining compliance over time as models evolve and adapt.130 This includes setting up mechanisms to regularly review and adjust AI models to ensure they meet regulatory standards and continue to operate ethically.131 Finally, it is crucial to differentiate medical uses from surveillance applications, as the context of use significantly shapes the ethical and practical implications. While public surveillance and law enforcement applications of facial recognition raise unique concerns, these challenges are distinct from those encountered in medical contexts, where the focus is on patient care and health outcomes.132 By tackling these areas, we can work toward a solution that ensures that AI is used responsibly and ethically in healthcare, benefitting patients globally while adhering to regulatory standards.

Discussion

AI-assisted facial analysis has significant potential in healthcare. This technology, with its non-invasive, user-friendly, and cost-effective attributes, is poised not only to advance disease screening but also to seamlessly integrate with telemedicine frameworks, enhancing disease management.133 The performance of the AI model has proven to be comparable to that of board-certified specialists and surpassed that of physicians in certain studies.45 This evidence supports the potential reliability of AI-assisted facial analysis in settings lacking specialist expertise. Smartphone applications have been developed for screening genetic disorders,43 detecting ptosis,54 and identifying visual impairment in young children.56 These applications have indeed improved screening efficiency, helping patients understand their health status more conveniently and achieving early detection of certain diseases. By improving the accessibility of health assessments, these applications conveniently help individuals understand their health status, which is crucial for preventing disease progression and improving health outcomes at the population level. The widespread implementation of these technologies holds promise for enhancing public health by facilitating early diagnosis, reducing healthcare disparities, and strengthening disease management strategies.

The ethical implications of using this technology should be fully considered before deployment, with particular attention paid to its application in specific contexts.134 These include disparities related to digital access that reflect health and social inequalities across populations based on sex, education, income, and rural/urban location. Individuals facing barriers to healthcare also tend to be digitally excluded. Therefore, AI-assisted facial analysis should be designed and implemented with an equal lens to benefit those who need it the most. This necessitates algorithms with higher sensitivity to detect as many patients with the disease as possible. Moreover, it requires the patients themselves, their families, and even neighbors and friends to have access to smartphones to conveniently utilize this technology. Challenges include instituting stringent privacy protections, expanding and diversifying datasets to accurately represent all patient demographics, enhancing the interpretability of AI models, eliminating potential biases within AI applications, and meeting regulatory standards.40 Collaborative efforts involving governments, non-governmental organizations, technology companies, educational services, and local communities will be essential in tailoring these technologies to the unique needs of different regions, including addressing linguistic and cultural considerations.135,136,137 These partnerships are crucial not only for ensuring the technology’s acceptance and effectiveness but also for building local capacities for its maintenance and advancement, ensuring its sustainable integration into healthcare systems. Future research should focus on refining AI algorithms and increasing the range of diseases detectable by AI to ensure the broad applicability and effectiveness of the technology in diverse clinical and cultural contexts.52,75

Overcoming these challenges will pave the way for AI-assisted facial analysis to make a profound impact on healthcare.83 It promises to improve the accessibility and efficiency of healthcare, particularly benefitting those in LMICs, where such advancements could dramatically change healthcare delivery. In these regions, the technology’s ability to provide high-quality care with minimal resources could improve medical practices, making it a critical component in the global effort to enhance healthcare outcomes and achieve health equity.138 Continued innovation, guided by ethical considerations and collaborative efforts, will enable this technology to contribute meaningfully to global health goals, ultimately improving access to quality healthcare for all.

Acknowledgments

We thank all of the investigators and participants in this study. This study was supported by the National Natural Science Foundation of China (82071003, 82271122, and 82388101); the Science and Technology Commission of Shanghai (20DZ2270800 and 23DZ2302200); the Shanghai Key Clinical Specialty, Shanghai Eye Disease Research Center (2022ZZ01003); and the Clinical Acceleration Program of Shanghai Ninth People’s Hospital, Shanghai Jiao Tong University School of Medicine (JYLJ202202). We also thank VoxelCloud Inc. for their support.

Author contributions

C.L., K.D., S.S., Z.W., R.B., and X.Y. drafted the manuscript. All authors contributed to the conceptualization, analysis, and editing of the manuscript.

Declaration of interests

The authors declare no competing interests.

Contributor Information

Xuefei Song, Email: songxuefei@shsmu.edu.cn.

Jionglong Su, Email: jionglong.su@xjtlu.edu.cn.

Xiaowei Ding, Email: dingxiaowei@sjtu.edu.cn.

Huifang Zhou, Email: fangzzfang@sjtu.edu.cn.

References

  • 1.Ebner N.C., He Y., Johnson M.K. Age and emotion affect how we look at a face: Visual scan patterns differ for own-age versus other-age emotional faces. Cognit. Emot. 2011;25:983–997. doi: 10.1080/02699931.2010.540817. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 2.Nelson M.M., Heffernan M.P. Facial signs of systemic disease. Facial Plast. Surg. Clin. 2003;11:175–195. doi: 10.1016/S1064-7406(02)00048-2. [DOI] [PubMed] [Google Scholar]
  • 3.Antonarakis S.E., Lyle R., Dermitzakis E.T., Reymond A., Deutsch S. Chromosome 21 and Down syndrome: from genomics to pathophysiology. Nat. Rev. Genet. 2004;5:725–738. doi: 10.1038/nrg1448. [DOI] [PubMed] [Google Scholar]
  • 4.Gurovich Y., Hanani Y., Bar O., Nadav G., Fleischer N., Gelbman D., Basel-Salmon L., Krawitz P.M., Kamphausen S.B., Zenker M., et al. Identifying facial phenotypes of genetic disorders using deep learning. Nat. Med. 2019;25:60–64. doi: 10.1038/s41591-018-0279-0. [DOI] [PubMed] [Google Scholar]
  • 5.Qiang J., Wu D., Du H., Zhu H., Chen S., Pan H. Review on facial-recognition-based applications in disease diagnosis. Bioengineering. 2022;9:273. doi: 10.3390/bioengineering9070273. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6.Guan Z., Li H., Liu R., Cai C., Liu Y., Li J., Wang X., Huang S., Wu L., Liu D., et al. Artificial intelligence in diabetes management: Advancements, opportunities, and challenges. Cell Rep. Med. 2023;4 doi: 10.1016/j.xcrm.2023.101213. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7.Kermany D.S., Goldbaum M., Cai W., Valentim C.C.S., Liang H., Baxter S.L., McKeown A., Yang G., Wu X., Yan F., et al. Identifying medical diagnoses and treatable diseases by image-based deep learning. Cell. 2018;172:1122–1131.e9. doi: 10.1016/j.cell.2018.02.010. [DOI] [PubMed] [Google Scholar]
  • 8.Okolo C.T. Optimizing human-centered AI for healthcare in the Global South. Patterns. 2022;3 doi: 10.1016/j.patter.2021.100421. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9.Kortli Y., Jridi M., Al Falou A., Atri M. Face Recognition Systems: A Survey. Sensors. 2020;20 doi: 10.3390/s20020342. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10.Schwalbe N., Wahl B. Artificial intelligence and the future of global health. Lancet. 2020;395:1579–1586. doi: 10.1016/S0140-6736(20)30226-9. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11.Beaglehole R., Epping-Jordan J., Patel V., Chopra M., Ebrahim S., Kidd M., Haines A. Improving the prevention and management of chronic disease in low-income and middle-income countries: a priority for primary health care. Lancet. 2008;372:940–949. doi: 10.1016/S0140-6736(08)61404-X. [DOI] [PubMed] [Google Scholar]
  • 12.Leo M., Carcagnì P., Mazzeo P.L., Spagnolo P., Cazzato D., Distante C. Analysis of facial information for healthcare applications: A survey on computer vision-based approaches. Information. 2020;11:128. [Google Scholar]
  • 13.Su Z., Liang B., Shi F., Gelfond J., Šegalo S., Wang J., Jia P., Hao X. Deep learning-based facial image analysis in medical research: a systematic review protocol. BMJ Open. 2021;11 doi: 10.1136/bmjopen-2020-047549. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14.Buolamwini J., Gebru T. Conference on Fairness, Accountability and Transparency (PMLR) 2018. Gender shades: Intersectional accuracy disparities in commercial gender classification; pp. 77–91. [Google Scholar]
  • 15.Goldenfein J. Proceedings of the Conference on Fairness, Accountability, and Transparency (ACM) 2019. The Profiling Potential of Computer Vision and the Challenge of Computational Empiricism; pp. 110–119. [DOI] [Google Scholar]
  • 16.Stark L., Hoey J. Proceedings of the 2021 ACM Conference on Fairness, Accountability, and Transparency (ACM) 2021. The Ethics of Emotion in Artificial Intelligence Systems; pp. 782–793. [DOI] [Google Scholar]
  • 17.Engelmann S., Ullstein C., Papakyriakopoulos O., Grossklags J. 2022 ACM Conference on Fairness, Accountability, and Transparency (ACM) 2022. What People Think AI Should Infer From Faces; pp. 128–141. [DOI] [Google Scholar]
  • 18.Viola P., Jones M.J. Robust real-time face detection. Int. J. Comput. Vis. 2004;57:137–154. [Google Scholar]
  • 19.Dalal N., Triggs B. 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’05) 2005. Histograms of oriented gradients for human detection; pp. 886–893. [Google Scholar]
  • 20.Li H., Lin Z., Shen X., Brandt J., Hua G. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 2015. A convolutional neural network cascade for face detection; pp. 5325–5334. [Google Scholar]
  • 21.Jiang B., Ren Q., Dai F., Xiong J., Yang J., Gui G. Communications, Signal Processing, and Systems: Proceedings of the 2018 CSPS Volume III: Systems 7th. Springer; 2020. Multi-task cascaded convolutional neural networks for real-time dynamic face recognition method; pp. 59–66. [Google Scholar]
  • 22.Girshick R., Donahue J., Darrell T., Malik J. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 2014. Rich feature hierarchies for accurate object detection and semantic segmentation; pp. 580–587. [Google Scholar]
  • 23.Xu Y., Yan W., Yang G., Luo J., Li T., He J. CenterFace: joint face detection and alignment using face as point. Sci. Program. 2020;2020:1–8. [Google Scholar]
  • 24.Liu W., Anguelov D., Erhan D., Szegedy C., Reed S., Fu C.-Y., Berg A.C. Computer Vision–ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, October 11–14, 2016, Proceedings, Part I. Springer; 2016. Ssd: Single shot multibox detector; pp. 21–37. [Google Scholar]
  • 25.Wang J., Sun K., Cheng T., Jiang B., Deng C., Zhao Y., Liu D., Mu Y., Tan M., Wang X., et al. Deep high-resolution representation learning for visual recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2021;43:3349–3364. doi: 10.1109/TPAMI.2020.2983686. [DOI] [PubMed] [Google Scholar]
  • 26.Deng J., Guo J., Ververas E., Kotsia I., Zafeiriou S. Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. 2020. Retinaface: Single-shot multi-level face localisation in the wild; pp. 5203–5212. [Google Scholar]
  • 27.Sun K., Zhao Y., Jiang B., Cheng T., Xiao B., Liu D., Mu Y., Wang X., Liu W., Wang J. High-resolution representations for labeling pixels and regions. arXiv. 2019 doi: 10.48550/arXiv.1904.04514. Preprint at. [DOI] [Google Scholar]
  • 28.Guo X., Li S., Yu J., Zhang J., Ma J., Ma L., Liu W., Ling H. PFLD: A practical facial landmark detector. arXiv. 2019 doi: 10.48550/arXiv.1902.10859. Preprint at. [DOI] [Google Scholar]
  • 29.Li W., Lu Y., Zheng K., Liao H., Lin C., Luo J., Cheng C.-T., Xiao J., Lu L., Kuo C.-F., et al. Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, August 23–28, 2020, Proceedings, Part IX. Springer; 2020. Structured landmark detection via topology-adapting deep graph learning; pp. 266–283. [Google Scholar]
  • 30.Lattas A., Moschoglou S., Gecer B., Ploumpis S., Triantafyllou V., Ghosh A., Zafeiriou S. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2020. AvatarMe: Realistically Renderable 3D Facial Reconstruction" in-the-wild; pp. 760–769. [Google Scholar]
  • 31.Jackson A.S., Bulat A., Argyriou V., Tzimiropoulos G. Proceedings of the IEEE International Conference on Computer Vision. 2017. Large pose 3D face reconstruction from a single image via direct volumetric CNN regression; pp. 1031–1039. [Google Scholar]
  • 32.Wu F., Bao L., Chen Y., Ling Y., Song Y., Li S., Ngan K.N., Liu W. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2019. Mvf-net: Multi-view 3D face morphable model regression; pp. 959–968. [Google Scholar]
  • 33.Ning X., Nan F., Xu S., Yu L., Zhang L. Multi-view frontal face image generation: a survey. Concurr. Comput. 2023;35 [Google Scholar]
  • 34.Zeng D., Veldhuis R., Spreeuwers L. A survey of face recognition techniques under occlusion. IET Biom. 2021;10:581–606. [Google Scholar]
  • 35.Krizhevsky A., Sutskever I., Hinton G.E. Imagenet classification with deep convolutional neural networks. Commun. ACM. 2017;60:84–90. [Google Scholar]
  • 36.Hu J., Shen L., Sun G. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 2018. Squeeze-and-excitation networks; pp. 7132–7141. [Google Scholar]
  • 37.Huang Y., Wang Y., Tai Y., Liu X., Shen P., Li S., Li J., Huang F. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2020. Curricularface: adaptive curriculum learning loss for deep face recognition; pp. 5901–5910. [Google Scholar]
  • 38.Liu Z., Luo P., Wang X., Tang X. Proceedings of the IEEE International Conference on Computer Vision. 2015. Deep learning face attributes in the wild; pp. 3730–3738. [Google Scholar]
  • 39.Cao J., Li Y., Zhang Z. Proceedings of the IEEE Conference on Conference on Computer Vision and Pattern Recognition. 2018. Partially shared multi-task convolutional neural network with local constraint for face attribute learning; pp. 4290–4299. [Google Scholar]
  • 40.Porras A.R., Rosenbaum K., Tor-Diez C., Summar M., Linguraru M.G. Development and evaluation of a machine learning-based point-of-care screening tool for genetic syndromes in children: a multinational retrospective study. Lancet Digit. Health. 2021;3:e635–e643. doi: 10.1016/S2589-7500(21)00137-0. [DOI] [PubMed] [Google Scholar]
  • 41.Pan Z., Shen Z., Zhu H., Bao Y., Liang S., Wang S., Li X., Niu L., Dong X., Shang X., et al. Clinical application of an automatic facial recognition system based on deep learning for diagnosis of Turner syndrome. Endocrine. 2021;72:865–873. doi: 10.1007/s12020-020-02539-3. [DOI] [PubMed] [Google Scholar]
  • 42.Yang H., Hu X.R., Sun L., Hong D., Zheng Y.Y., Xin Y., Liu H., Lin M.Y., Wen L., Liang D.P., Wang S.S. Automated Facial Recognition for Noonan Syndrome Using Novel Deep Convolutional Neural Network With Additive Angular Margin Loss. Front. Genet. 2021;12 doi: 10.3389/fgene.2021.669841. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 43.Latorre-Pellicer A., Ascaso Á., Trujillano L., Gil-Salvador M., Arnedo M., Lucia-Campos C., Antoñanzas-Pérez R., Marcos-Alcalde I., Parenti I., Bueno-Lozano G., Musio A. Evaluating Face2Gene as a Tool to Identify Cornelia de Lange Syndrome by Facial Phenotypes. Int. J. Mol. Sci. 2020;21:1042. doi: 10.3390/ijms21031042. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 44.Mishima H., Suzuki H., Doi M., Miyazaki M., Watanabe S., Matsumoto T., Morifuji K., Moriuchi H., Yoshiura K.I., Kondoh T., Kosaki K. Evaluation of Face2Gene using facial images of patients with congenital dysmorphic syndromes recruited in Japan. J. Hum. Genet. 2019;64:789–794. doi: 10.1038/s10038-019-0619-z. [DOI] [PubMed] [Google Scholar]
  • 45.Han S.S., Moon I.J., Lim W., Suh I.S., Lee S.Y., Na J.-I., Kim S.H., Chang S.E. Keratinocytic skin cancer detection on the face using region-based convolutional neural network. JAMA Dermatol. 2020;156:29–37. doi: 10.1001/jamadermatol.2019.3807. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 46.Wen H., Yu W., Wu Y., Zhao J., Liu X., Kuang Z., Fan R. Acne Detection and Severity Evaluation with Interpretable Convolutional Neural Network Models. Technol. Health Care. 2022;30:143–153. doi: 10.3233/THC-228014. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 47.Binol H., Plotner A., Sopkovich J., Kaffenberger B., Niazi M.K.K., Gurcan M.N. Ros-NET: A deep convolutional neural network for automatic identification of rosacea lesions. Skin Res. Technol. 2020;26:413–421. doi: 10.1111/srt.12817. [DOI] [PubMed] [Google Scholar]
  • 48.Peng J., Gao R., Thng S., Huang W., Lin Z. 2021 43rd Annual International Conference of the IEEE Engineering in Medicine & Biology Society (EMBC) 2021. Classification of Non-tumorous Facial Pigmentation Disorders Using Generative Adversarial Networks and Improved SMOTE; pp. 3770–3773. [DOI] [PubMed] [Google Scholar]
  • 49.Karlin J., Gai L., LaPierre N., Danesh K., Farajzadeh J., Palileo B., Taraszka K., Zheng J., Wang W., Eskin E., Rootman D. Ensemble neural network model for detecting thyroid eye disease using external photographs. Br. J. Ophthalmol. 2023;107:1722–1729. doi: 10.1136/bjo-2022-321833. [DOI] [PubMed] [Google Scholar]
  • 50.Huang X., Ju L., Li J., He L., Tong F., Liu S., Li P., Zhang Y., Wang X., Yang Z., et al. An Intelligent Diagnostic System for Thyroid-Associated Ophthalmopathy Based on Facial Images. Front. Med. 2022;9 doi: 10.3389/fmed.2022.920716. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 51.Lei C., Qu M., Sun H., Huang J., Huang J., Song X., Zhai G., Zhou H. Facial expression of patients with Graves’ orbitopathy. J. Endocrinol. Invest. 2023;46:2055–2066. doi: 10.1007/s40618-023-02054-y. [DOI] [PubMed] [Google Scholar]
  • 52.Hung J.Y., Chen K.W., Perera C., Chiu H.K., Hsu C.R., Myung D., Luo A.C., Fuh C.S., Liao S.L., Kossler A.L. An Outperforming Artificial Intelligence Model to Identify Referable Blepharoptosis for General Practitioners. J. Personalized Med. 2022;12 doi: 10.3390/jpm12020283. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 53.Lou L., Cao J., Wang Y., Gao Z., Jin K., Xu Z., Zhang Q., Huang X., Ye J. Deep learning-based image analysis for automated measurement of eyelid morphology before and after blepharoptosis surgery. Ann. Med. 2021;53:2278–2285. doi: 10.1080/07853890.2021.2009127. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 54.Tabuchi H., Nagasato D., Masumoto H., Tanabe M., Ishitobi N., Ochi H., Shimizu Y., Kiuchi Y. Developing an iOS application that uses machine learning for the automated diagnosis of blepharoptosis. Graefes Arch. Clin. Exp. Ophthalmol. 2022;260:1329–1335. doi: 10.1007/s00417-021-05475-8. [DOI] [PubMed] [Google Scholar]
  • 55.Hui S., Dong L., Zhang K., Nie Z., Jiang X., Li H., Hou Z., Ding J., Wang Y., Li D. Noninvasive identification of Benign and malignant eyelid tumors using clinical images via deep learning system. J. Big Data. 2022;9:84. [Google Scholar]
  • 56.Chen W., Li R., Yu Q., Xu A., Feng Y., Wang R., Zhao L., Lin Z., Yang Y., Lin D., et al. Early detection of visual impairment in young children using a smartphone-based deep learning system. Nat. Med. 2023;29:493–503. doi: 10.1038/s41591-022-02180-9. [DOI] [PubMed] [Google Scholar]
  • 57.Kong X., Gong S., Su L., Howard N., Kong Y. Automatic Detection of Acromegaly From Facial Photographs Using Machine Learning Methods. EBioMedicine. 2018;27:94–102. doi: 10.1016/j.ebiom.2017.12.015. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 58.Kong Y., Kong X., He C., Liu C., Wang L., Su L., Gao J., Guo Q., Cheng R. Constructing an automatic diagnosis and severity-classification model for acromegaly using facial photographs by deep learning. J. Hematol. Oncol. 2020;13 doi: 10.1186/s13045-020-00925-y. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 59.Wei R., Jiang C., Gao J., Xu P., Zhang D., Sun Z., Liu X., Deng K., Bao X., Sun G., et al. Deep-Learning Approach to Automatic Identification of Facial Anomalies in Endocrine Disorders. Neuroendocrinology. 2020;110:328–337. doi: 10.1159/000502211. [DOI] [PubMed] [Google Scholar]
  • 60.Ali M.R., Myers T., Wagner E., Ratnu H., Dorsey E.R., Hoque E. Facial expressions can detect Parkinson’s disease: preliminary evidence from videos collected online. NPJ Digit. Med. 2021;4:129. doi: 10.1038/s41746-021-00502-8. [DOI] [PMC free article] [PubMed] [Google Scholar] [Retracted]
  • 61.Zhou X., Jin K., Shang Y., Guo G. Visually interpretable representation learning for depression recognition from facial images. IEEE Trans. Affect. Comput. 2020;11:542–552. [Google Scholar]
  • 62.Alam M.S., Rashid M.M., Roy R., Faizabadi A.R., Gupta K.D., Ahsan M.M. Empirical study of autism spectrum disorder diagnosis using facial images by improved transfer learning approach. Bioengineering. 2022;9:710. doi: 10.3390/bioengineering9110710. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 63.Xiao W., Huang X., Wang J.H., Lin D.R., Zhu Y., Chen C., Yang Y.H., Xiao J., Zhao L.Q., Li J.-P.O., et al. Screening and identifying hepatobiliary diseases through deep learning using ocular images: a prospective, multicentre study. Lancet. Digit. Health. 2021;3:e88–e97. doi: 10.1016/S2589-7500(20)30288-0. [DOI] [PubMed] [Google Scholar]
  • 64.Lin S., Li Z., Fu B., Chen S., Li X., Wang Y., Wang X., Lv B., Xu B., Song X., et al. Feasibility of using deep learning to detect coronary artery disease based on facial photo. Eur. Heart J. 2020;41:4400–4411. doi: 10.1093/eurheartj/ehaa640. [DOI] [PubMed] [Google Scholar]
  • 65.Zhang A., Lou J., Pan Z., Luo J., Zhang X., Zhang H., Li J., Wang L., Cui X., Ji B., Chen L. Prediction of anemia using facial images and deep learning technology in the emergency department. Front. Public Health. 2022;10:964385. doi: 10.3389/fpubh.2022.964385. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 66.Groh M., Badri O., Daneshjou R., Koochek A., Harris C., Soenksen L.R., Doraiswamy P.M., Picard R. Deep learning-aided decision support for diagnosis of skin disease across skin tones. Nat. Med. 2024;30:573–583. doi: 10.1038/s41591-023-02728-3. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 67.George G., Murphy D.C., Hogg H.D.J., Boniface J.B., Urasa S., Rwiza J., Uwemeye L., Bristow C., Hillsmith G., Rainey E., et al. Evaluation of a low-resource screening strategy for ophthalmic pathologies and associated neurological morbidity in an older Tanzanian HIV-positive population. Sci. Rep. 2022;12:1434. doi: 10.1038/s41598-022-04989-3. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 68.Bao X.-L., Sun Y.-J., Zhan X., Li G.-Y. Orbital and eyelid diseases: The next breakthrough in artificial intelligence? Front. Cell Dev. Biol. 2022;10 doi: 10.3389/fcell.2022.1069248. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 69.Ting D.S.W., Peng L., Varadarajan A.V., Keane P.A., Burlina P.M., Chiang M.F., Schmetterer L., Pasquale L.R., Bressler N.M., Webster D.R., et al. Deep learning in ophthalmology: the technical and clinical considerations. Prog. Retin. Eye Res. 2019;72 doi: 10.1016/j.preteyeres.2019.04.003. [DOI] [PubMed] [Google Scholar]
  • 70.Aikaeli F., Njim T., Gissing S., Moyo F., Alam U., Mfinanga S.G., Okebe J., Ramaiya K., Webb E.L., Jaffar S., Garrib A. Prevalence of microvascular and macrovascular complications of diabetes in newly diagnosed type 2 diabetes in low-and-middle-income countries: A systematic review and meta-analysis. PLOS Glob. Public Health. 2022;2 doi: 10.1371/journal.pgph.0000599. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 71.Birbeck G.L., Meyer A.-C., Ogunniyi A. Nervous system disorders across the life course in resource-limited settings. Nature. 2015;527:S167–S171. doi: 10.1038/nature16031. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 72.Garain A., Ray B., Singh P.K., Ahmadian A., Senu N., Sarkar R. GRA_Net: A deep learning model for classification of age and gender from facial images. IEEE Access. 2021;9:85672–85689. [Google Scholar]
  • 73.Xia X., Chen X., Wu G., Li F., Wang Y., Chen Y., Chen M., Wang X., Chen W., Xian B., et al. Three-dimensional facial-image analysis to predict heterogeneity of the human ageing rate and the impact of lifestyle. Nat. Metab. 2020;2:946–957. doi: 10.1038/s42255-020-00270-x. [DOI] [PubMed] [Google Scholar]
  • 74.Huan E.-Y., Wen G.-H. Transfer learning with deep convolutional neural network for constitution classification with face image. Multimed. Tool. Appl. 2020;79:11905–11919. [Google Scholar]
  • 75.Tay W., Quek R., Kaur B., Lim J., Henry C.J. Use of Facial Morphology to Determine Nutritional Status in Older Adults: Opportunities and Challenges. JMIR Public Health Surveill. 2022;8 doi: 10.2196/33478. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 76.Van Brummen A., Owen J.P., Spaide T., Froines C., Lu R., Lacy M., Blazes M., Li E., Lee C.S., Lee A.Y., Zhang M. PeriorbitAI: artificial intelligence automation of eyelid and periorbital measurements. Am. J. Ophthalmol. 2021;230:285–296. doi: 10.1016/j.ajo.2021.05.007. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 77.Song X., Tong W., Lei C., Huang J., Fan X., Zhai G., Zhou H. A clinical decision model based on machine learning for ptosis. BMC Ophthalmol. 2021;21 doi: 10.1186/s12886-021-01923-5. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 78.Fontaine D., Vielzeuf V., Genestier P., Limeux P., Santucci-Sivilotto S., Mory E., Darmon N., Lanteri-Minet M., Mokhtar M., Laine M., et al. Artificial intelligence to evaluate postoperative pain based on facial expression recognition. Eur. J. Pain. 2022;26:1282–1291. doi: 10.1002/ejp.1948. [DOI] [PubMed] [Google Scholar]
  • 79.Bahçeci Şimşek İ., Şirolu C. Analysis of surgical outcome after upper eyelid surgery by computer vision algorithm using face and facial landmark detection. Graefes Arch. Clin. Exp. Ophthalmol. 2021;259:3119–3125. doi: 10.1007/s00417-021-05219-8. [DOI] [PubMed] [Google Scholar]
  • 80.Hidaka T., Tanaka K., Mori H. An Artificial Intelligence-Based Cosmesis Evaluation for Temporomandibular Joint Reconstruction. Laryngoscope. 2023;133:841–848. doi: 10.1002/lary.30239. [DOI] [PubMed] [Google Scholar]
  • 81.Zhai Y., Cao H., Deng W., Gan J., Piuri V., Zeng J. BeautyNet: Joint multiscale CNN and transfer learning method for unconstrained facial beauty prediction. Comput. Intell. Neurosci. 2019;2019 doi: 10.1155/2019/1910624. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 82.Hossain M.S., Muhammad G. Emotion-aware connected healthcare big data towards 5G. IEEE Internet Things J. 2018;5:2399–2406. [Google Scholar]
  • 83.Wang J., Luo Y., Wang Z., Hounye A.H., Cao C., Hou M., Zhang J. A cell phone app for facial acne severity assessment. Appl. Intell. 2023;53:7614–7633. doi: 10.1007/s10489-022-03774-z. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 84.Dyda A., Purcell M., Curtis S., Field E., Pillai P., Ricardo K., Weng H., Moore J.C., Hewett M., Williams G., Lau C.L. Differential privacy for public health data: An innovative tool to optimize information sharing while protecting data confidentiality. Patterns. 2021;2 doi: 10.1016/j.patter.2021.100366. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 85.Mohapatra S. Use of facial recognition technology for medical purposes: balancing privacy with innovation. Pepp. L. Rev. 2015;43:1017. [Google Scholar]
  • 86.Kassam I., Ilkina D., Kemp J., Roble H., Carter-Langford A., Shen N. Patient Perspectives and Preferences for Consent in the Digital Health Context: State-of-the-art Literature Review. J. Med. Internet Res. 2023;25 doi: 10.2196/42507. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 87.Fang C., Dziedzic A., Zhang L., Oliva L., Verma A., Razak F., Papernot N., Wang B. Decentralised, collaborative, and privacy-preserving machine learning for multi-hospital data. EBioMedicine. 2024;101 doi: 10.1016/j.ebiom.2024.105006. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 88.Niu Y., Deng W. Proceedings of the AAAI Conference on Artificial Intelligence. 2022. Federated learning for face recognition with gradient correction; pp. 1999–2007. [Google Scholar]
  • 89.Warnat-Herresthal S., Schultze H., Shastry K.L., Manamohan S., Mukherjee S., Garg V., Sarveswara R., Händler K., Pickkers P., Aziz N.A., et al. Swarm Learning for decentralized and confidential clinical machine learning. Nature. 2021;594:265–270. doi: 10.1038/s41586-021-03583-3. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 90.Margetts H., Dorobantu C. Rethink government with AI. Nature. 2019;568:163–165. doi: 10.1038/d41586-019-01099-5. [DOI] [PubMed] [Google Scholar]
  • 91.Falcetta A., Roveri M. Privacy-preserving deep lea/rning with homomorphic encryption: An introduction. IEEE Comput. Intell. Mag. 2022;17:14–25. [Google Scholar]
  • 92.Wang R., Lin H. Anonymizing facial images to improve patient privacy. Nat. Med. 2022;28:1767–1768. doi: 10.1038/s41591-022-01967-0. [DOI] [PubMed] [Google Scholar]
  • 93.Uthoff R.D., Song B., Sunny S., Patrick S., Suresh A., Kolur T., Keerthi G., Spires O., Anbarani A., Wilder-Smith P., et al. Point-of-care, smartphone-based, dual-modality, dual-view, oral cancer screening device with neural network classification for low-resource communities. PLoS One. 2018;13 doi: 10.1371/journal.pone.0207493. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 94.He J., Baxter S.L., Xu J., Xu J., Zhou X., Zhang K. The practical implementation of artificial intelligence technologies in medicine. Nat. Med. 2019;25:30–36. doi: 10.1038/s41591-018-0307-0. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 95.Sharma B., Gao Y., Miller T., Churpek M.M., Afshar M., Dligach D. Multi-Task Training with In-Domain Language Models for Diagnostic Reasoning. arXiv. 2023 doi: 10.48550/arXiv.2306.04551. Preprint at. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 96.Akter T., Ali M.H., Khan M.I., Satu M.S., Uddin M.J., Alyami S.A., Ali S., Azad A., Moni M.A. Improved transfer-learning-based facial recognition framework to detect autistic children at an early stage. Brain Sci. 2021;11:734. doi: 10.3390/brainsci11060734. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 97.Zhang N., Ruan M., Wang S., Paul L., Li X. Discriminative few shot learning of facial dynamics in interview videos for autism trait classification. IEEE Trans. Affect. Comput. 2022;14:1110–1124. doi: 10.1109/TAFFC.2022.3178946. [DOI] [Google Scholar]
  • 98.Krishnan R., Rajpurkar P., Topol E.J. Self-supervised learning in medicine and healthcare. Nat. Biomed. Eng. 2022;6:1346–1352. doi: 10.1038/s41551-022-00914-1. [DOI] [PubMed] [Google Scholar]
  • 99.Wang H., Chi L., Su C., Zhao Z. Proceedings of the 31st ACM International Conference on Information & Knowledge Management. 2022. ASDFace: Face-based Autism Diagnosis via Heterogeneous Domain Adaptation; pp. 4999–5003. [Google Scholar]
  • 100.Li Y., Liu M.-Y., Li X., Yang M.-H., Kautz J. Proceedings of the European Conference on Computer Vision (ECCV) 2018. A closed-form solution to photorealistic image stylization; pp. 453–468. [Google Scholar]
  • 101.Shen Z., Lai W.-S., Xu T., Kautz J., Yang M.-H. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 2018. Deep semantic face deblurring; pp. 8260–8269. [Google Scholar]
  • 102.Yim J., Jung H., Yoo B., Choi C., Park D., Kim J. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 2015. Rotating your face using multi-task deep neural network; pp. 676–684. [Google Scholar]
  • 103.Zhu Y., Salowe R., Chow C., Li S., Bastani O., O’Brien J.M. Advancing Glaucoma Care: Integrating Artificial Intelligence in Diagnosis, Management, and Progression Detection. Bioengineering. 2024;11:122. doi: 10.3390/bioengineering11020122. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 104.Andrews M., Smart A., Birhane A. The reanimation of pseudoscience in machine learning and its ethical repercussions. Patterns. 2024;5 doi: 10.1016/j.patter.2024.101027. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 105.Zhou B., Khosla A., Lapedriza A., Oliva A., Torralba A. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 2016. Learning deep features for discriminative localization; pp. 2921–2929. [Google Scholar]
  • 106.Kaur D. Purdue University; 2023. Trustworthy AI: Ensuring Explainability & Acceptance. PhD thesis. [Google Scholar]
  • 107.Fordyce A.J. Georgetown University; 2021. Toward Ethical Applications of Artificial Intelligence: Understanding Current Uses of Facial Recognition Technology and Advancing Bias Mitigation Strategies. PhD thesis. [Google Scholar]
  • 108.Mohanty S.D., Lekan D., McCoy T.P., Jenkins M., Manda P. Machine learning for predicting readmission risk among the frail: Explainable AI for healthcare. Patterns. 2022;3 doi: 10.1016/j.patter.2021.100395. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 109.Raghavan K., Sivasilan B., Kamakoti V. Attention guided grad-CAM: an improved explainable artificial intelligence model for infrared breast cancer detection. Multimed. Tool. Appl. 2023;83:57551–57578. [Google Scholar]
  • 110.Alowais S.A., Alghamdi S.S., Alsuhebany N., Alqahtani T., Alshaya A.I., Almohareb S.N., Aldairem A., Alrashed M., Bin Saleh K., Badreldin H.A., et al. Revolutionizing healthcare: the role of artificial intelligence in clinical practice. BMC Med. Educ. 2023;23:689. doi: 10.1186/s12909-023-04698-z. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 111.Virdi S.S., Wertheim D., Naini F.B. Normative anthropometry and proportions of the Kenyan-African face and comparative anthropometry in relation to African Americans and North American Whites. Maxillofac. Plast. Reconstr. Surg. 2019;41:9–14. doi: 10.1186/s40902-019-0191-7. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 112.Lu A., Perkowski M. Deep learning approach for screening autism spectrum disorder in children with facial images and analysis of ethnoracial factors in model development and application. Brain Sci. 2021;11:1446. doi: 10.3390/brainsci11111446. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 113.Chen R.J., Wang J.J., Williamson D.F.K., Chen T.Y., Lipkova J., Lu M.Y., Sahai S., Mahmood F. Algorithmic fairness in artificial intelligence for medicine and healthcare. Nat. Biomed. Eng. 2023;7:719–742. doi: 10.1038/s41551-023-01056-8. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 114.Were M.C. Challenges in digital medicine applications in under-resourced settings. Nat. Commun. 2022;2022:3020. doi: 10.1038/s41467-022-30728-3. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 115.Sheng B., Guan Z., Lim L.-L., Jiang Z., Mathioudakis N., Li J., Liu R., Bao Y., Bee Y.M., Wang Y.-X., et al. Large language models for diabetes care: Potentials and prospects. Sci. Bull. 2024:S2095–S9273. doi: 10.1016/j.scib.2024.01.004. [DOI] [PubMed] [Google Scholar]
  • 116.Radford, A., Narasimhan, K., Salimans, T., Sutskever, I. (2018). Improving language understanding by generative pre-training.
  • 117.Brown T., Mann B., Ryder N., Subbiah M., Kaplan J.D., Dhariwal P., Neelakantan A., Shyam P., Sastry G., Askell A., et al. Language models are few-shot learners. Adv. Neural Inf. Process. Syst. 2020;33:1877–1901. [Google Scholar]
  • 118.Kirillov A., Mintun E., Ravi N., Mao H., Rolland C., Gustafson L., Xiao T., Whitehead S., Berg A.C., Lo W.Y., Dollár P. Proceedings of the IEEE/CVF International Conference on Computer Vision. 2023. Segment anything; pp. 4015–4026. [Google Scholar]
  • 119.Norori N., Hu Q., Aellen F.M., Faraci F.D., Tzovara A. Addressing bias in big data and AI for health care: A call for open science. Patterns. 2021;2 doi: 10.1016/j.patter.2021.100347. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 120.Wang M., Deng W. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2020. Mitigating bias in face recognition using skewness-aware reinforcement learning; pp. 9322–9331. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 121.Alvi M., Zisserman A., Nellåker C. Proceedings of the European Conference on Computer Vision (ECCV) Workshops. 2018. Turning a blind eye: Explicit removal of biases and variation from deep neural network embeddings. [Google Scholar]
  • 122.Zhang, Y., and Chen, D.Z. (2023). GPT4MIA: Utilizing Generative Pre-trained Transformer (GPT-3) as a Plug-and-Play Transductive Model for Medical Image Analysis. In Medical Image Computing and Computer Assisted Intervention – MICCAI 2023 Workshops Lecture Notes in Computer Science., M. E. Celebi, M. S. Salekin, H. Kim, S. Albarqouni, C. Barata, A. Halpern, P. Tschandl, M. Combalia, Y. Liu, G. Zamzmi, et al., eds. (Springer Nature Switzerland), pp. 151–160. 10.1007/978-3-031-47401-9_15. [DOI]
  • 123.Parsheera, S. (2019). Adoption and Regulation of Facial Recognition Technologies in India: Why and Why Not?. 10.2139/ssrn.3525324. [DOI]
  • 124.Raposo V.L., Du L. Facial recognition technology: is it ready to be used in public health surveillance? Int. Data Priv. Law. 2023;14:66–86. [Google Scholar]
  • 125.Fihn S., Saria S., Mendonça E., Hain E., Matheny M., Shah N., Liu H., Auerbach A. Artificial Intelligence in Health Care: The Hope, the Hype, the Promise, the Peril. National Academies Press; 2019. Deploying AI in clinical settings. [DOI] [PubMed] [Google Scholar]
  • 126.Wirtz B.W., Weyerer J.C., Kehl I. Governance of artificial intelligence: A risk and guideline-based integrative framework. Govern. Inf. Q. 2022;39 [Google Scholar]
  • 127.Walz A., Firth-Butterfield K. Implementing ethics into artificial intelligence: A contribution, from a legal perspective, to the development of an AI governance regime. Duke Tech Rev. 2019;18:176. [Google Scholar]
  • 128.European Union (2024). Artificial Intelligence Act: Text of the provisional agreement.
  • 129.Kao Y.-H., Sapp S.G. The effect of cultural values and institutional trust on public perceptions of government use of network surveillance. Technol. Soc. 2022;70 [Google Scholar]
  • 130.Fontes C., Hohma E., Corrigan C.C., Lütge C. AI-powered public surveillance systems: why we (might) need them and how we want them. Technol. Soc. 2022;71 [Google Scholar]
  • 131.Janssen M., Brous P., Estevez E., Barbosa L.S., Janowski T. Data governance: Organizing data for trustworthy Artificial Intelligence. Govern. Inf. Q. 2020;37 [Google Scholar]
  • 132.Lee L.M., Thacker S.B. Public health surveillance and knowing about health in the context of growing sources of health data. Am. J. Prev. Med. 2011;41:636–640. doi: 10.1016/j.amepre.2011.08.015. [DOI] [PubMed] [Google Scholar]
  • 133.Dorsey E.R., Topol E.J. Telemedicine 2020 and the next decade. Lancet. 2020;395:859. doi: 10.1016/S0140-6736(20)30424-4. [DOI] [PubMed] [Google Scholar]
  • 134.Richardson S., Lawrence K., Schoenthaler A.M., Mann D. A framework for digital health equity. NPJ Digit. Med. 2022;5:119. doi: 10.1038/s41746-022-00663-0. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 135.Ampamya S., Kitayimbwa J.M., Were M.C. Performance of an open source facial recognition system for unique patient matching in a resource-limited setting. Int. J. Med. Inf. 2020;141 doi: 10.1016/j.ijmedinf.2020.104180. [DOI] [PubMed] [Google Scholar]
  • 136.McMahon A., Buyx A., Prainsack B. Big Data Governance Needs More Collective Responsibility: The Role of Harm Mitigation in the Governance of Data Use in Medicine and Beyond. Med. Law Rev. 2020;28:155–182. doi: 10.1093/medlaw/fwz016. [DOI] [PubMed] [Google Scholar]
  • 137.Wilson S., Tolley C., Mc Ardle R., Lawson L., Beswick E., Hassan N., Slight R., Slight S. Recommendations to advance digital health equity: a systematic review of qualitative studies. NPJ Digit. Med. 2024;7:173. doi: 10.1038/s41746-024-01177-7. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 138.Amirav I., Masumbuko C.K., Hawkes M.T., Solomon I., Aldar Y., Margalit G., Zvirin A., Honen Y., Sivasivugha E.S., Kimmel R. 3D analysis of child facial dimensions for design of medical devices in low-middle income countries (LMIC) PLoS One. 2019;14 doi: 10.1371/journal.pone.0216548. [DOI] [PMC free article] [PubMed] [Google Scholar]

Articles from Patterns are provided here courtesy of Elsevier

RESOURCES