Skip to main content
Springer logoLink to Springer
. 2022 Sep 24;26(12):6871–6879. doi: 10.1007/s00784-022-04724-2

Artificial intelligence in medico-dental diagnostics of the face: a narrative review of opportunities and challenges

Raphael Patcas 1,, Michael M Bornstein 2, Marc A Schätzle 1, Radu Timofte 3,4
PMCID: PMC9708749  PMID: 36153437

Abstract

Objectives

This review aims to share the current developments of artificial intelligence (AI) solutions in the field of medico-dental diagnostics of the face. The primary focus of this review is to present the applicability of artificial neural networks (ANN) to interpret medical images, together with the associated opportunities, obstacles, and ethico-legal concerns.

Material and methods

Narrative literature review.

Results

Narrative literature review.

Conclusion

Curated facial images are widely available and easily accessible and are as such particularly suitable big data for ANN training. New AI solutions have the potential to change contemporary dentistry by optimizing existing processes and enriching dental care with the introduction of new tools for assessment or treatment planning. The analyses of health-related big data may also contribute to revolutionize personalized medicine through the detection of previously unknown associations. In regard to facial images, advances in medico-dental AI-based diagnostics include software solutions for the detection and classification of pathologies, for rating attractiveness and for the prediction of age or gender. In order for an ANN to be suitable for medical diagnostics of the face, the arising challenges regarding computation and management of the software are discussed, with special emphasis on the use of non-medical big data for ANN training. The legal and ethical ramifications of feeding patients’ facial images to a neural network for diagnostic purposes are related to patient consent, data privacy, data security, liability, and intellectual property. Current ethico-legal regulation practices seem incapable of addressing all concerns and ensuring accountability.

Clinical significance

While this review confirms the many benefits derived from AI solutions used for the diagnosis of medical images, it highlights the evident lack of regulatory oversight, the urgent need to establish licensing protocols, and the imperative to investigate the moral quality of new norms set with the implementation of AI applications in medico-dental diagnostics.

Keywords: Artificial intelligence, Photography, Face, Neural networks, Government regulation and oversight

Introduction

The face as an object of diagnostic interest

Interest in the analysis of the face is apparent in the earliest records of medical art. The E. Smith Papyrus, dating c. 1600 before the Common Era and as such the most ancient known medical text, opens with a diagnostic description of 15 injuries of the face. Since the preserved Papyrus only illustrates a total of 48 case histories, emphasis on the examination of the face, which in ancient Egypt included visual, tactile, and also olfactory clues, is evident [1, 2].

In the context of dental medicine in general, and orthodontic or maxillofacial surgery in particular, the analysis of the face has enjoyed great interest, both for diagnostic purposes and for outcome assessments. The information contained within a facial image is manifold and includes evidence on possible pathologies, malformations, or deviations of the norm, be it of the skin or of underlying musculoskeletal structures, that may all be instrumental in the diagnostic analysis of a case. The face mirrors, on the other hand, the aesthetic success of an orthodontic or maxillofacial intervention, and judgment is made on whether a balanced and harmonious facial appearance is restored or established.

Yet, the analysis of the face has remained predominantly a clinical task that depends largely on the clinician’s expertise and experience, and the assessment of the aesthetic outcome suffers of inconsistency owing to subjectivity. Medical sciences have thus since long attempted to infer information contained in medical images of the human face in a structured and dependable approach, in order to overcome the shortcomings related to subjectivity or the dependency of expertise and experience.

Applying artificial intelligence (AI) to medical diagnostics and outcome evaluation would seem to satisfy these demands. With the intention to reproduce the human cognitive process, AI has the potential to entirely modify dental clinical care [3] and research [4]. Presenting the potentials of AI application in medico-dental diagnostics of the face, this review ventures to share—from the vantage point of a practitioner—current developments of AI in this field, present possible novel opportunities, and discuss prevalent obstacles.

AI and mimicking natural intelligence

Natural intelligence is based on a learning process where a stimulus is first perceived, then its attributes interpreted. The interpretation of the stimulus leads, in an intelligent living being, to a response in terms of an adaptation when exposed to the same stimulus again. AI can be understood in a similar vein: data serve as stimulus, and patterns within the data are being recognized, extracted, and interpreted. Deep learning is the part of machine learning that aims to emulate the natural learning process by use of artificial neural networks (ANN) which are based on multiple layers trained to progressively extract higher-level features (Fig. 1). These convolutional neural networks are patterned loosely on the mammalian visual cortex [5] and will, this time for an intelligent agent, likewise lead as part of a learning process to an adaptation when exposed to the same stimulus again. In medical imaging, the triplet terminology of stimulus/pattern recognition/interpretation would thus be reiterated as data/findings/diagnostics.

Fig. 1.

Fig. 1

Typically, for (1) an input image, (2) the face is first detected and fed to (3–6) a deep learning model that (3) extracts facial features and (4) assesses the relationships between them and annotated data to (5) regress to a range of values that it is finally (6) mapped to a predicted label output

In dental medicine, the stimuli are mostly data originating from images or patient files, yet some research have also used machine learning to find associations to olfactory patterns such as oral malodour [6].

Data acquisition and curation

Photographic images, which are being used as data for facial diagnostics, either are produced a priori as dedicated medical facial images or consist of re-utilized non-medical images. While customized medical facial images are generated under tailored and standardized conditions, the use of non-medical data allows to access an immense amount of commercially collected data. Obviously, the origin of the facial images will have considerable implications on research design and ethical considerations. Some of them will be discussed below.

Before data can be fed to a neural network, they must be curated. Data curation involves the organization and annotation of the information contained within the data. More specifically, data curation in medicine focuses on extracting relevant information from patient history or medical images. One of the prime factors of driving the applicability of AI and deep learning in medical sciences at such incredible pace is surely the fact that digitization of health-related records has become the norm, which facilitates data curation immensely. Patient history and imaging (radiographic, histologic, and others) are usually not only produced in a digital format, but often also even labeled or categorized as they are being created [5]. This is especially true for the face, for which established standardized high-resolution 2D-photography has been complemented by modern 3D acquisition techniques such as stereophotogrammetry, laser scanning, structured light scanning, or 3D image reconstructions of CT or CBCT data [7, 8]. (A comparative analysis on the diagnostic performance of the different image acquisition types would indeed be valuable, but the seeming lack of scientific literature on this subject precludes any detailed comparison.)

Acquisition and storage of health-related digital data were immensely facilitated in the last two decades, as computational devices became ubiquitous, interlinked, and more potent [5]. These winning conditions led the foundation for a remarkable growth in the volume of information. Such “big data” are, however, too large or too complex to be dealt with by traditional data-processing application software [3], and only the evolution of powerful capabilities in machine learning as enabled AI to process and analyze such vast and multidimensional data [9].

With the introduction of AI in medicine, many possibilities and challenges have emerged. It is not the aim of the review to debate on all possible opportunities and eventualities, but rather focus on the predominant benefits and acute concerns related to the use of deep learning in facial image analysis in a medico-dental setting.

AI applications in medico-dental diagnostics of the face: opportunities

Deep learning perfectly suited to diagnose facial images

Not all machine learning approaches are suitable for medical diagnostics, but deep learning models based on artificial neural networks generated an intuitive drawing power to health-related, image-based applications, given their apparent strengths in processing big data, pattern recognition, and predictive model building from large high-dimensional data sets [5], and have shown from their very onset outstanding performance in image processing [10]. Particularly for facial recognition, artificial neural networks (ANN) have proven to perform computation-intensive operations based on massive data sets with significant accuracy and performance [11].

Over the last few years, researchers have advanced the implementation of ANN-based predictions in medico-dental diagnostics of the face. The applications in facial diagnostics include predictions for classification of skin pathologies [1214] or dysmorphic features [1519], attractiveness [4, 2022], perceived age [4, 20, 23], or gender [24]. ANN have also been trained to detect health-related patterns and traits in patients, such as pain [25, 26] or stress [27]. These and many more have the potential to alter the healing arts by providing the following benefits to medico-dental care:

Optimized care

Deep learning techniques applied to identify or classify patterns in a facial image as a diagnostic task have often performed on par or better than professionals [18]. Using neural networks, for instance to classify pathologies or assess attractiveness, is expected to streamline care and relieve the dental workforce from laborious routine tasks [28]. Optimizing diagnostic processes will surely result in faster and more reliable health care, and may hopefully help reduce costs and make inexpensive diagnostics accessible to less privileged parts of the globe [28]. Lastly, AI-based diagnostics holds the potential to increase the impartiality in decision-making processes relating to funding bodies (i.e., insurances, health services).

Enriched care

The historical approaches to scoring facial attractiveness of patients have frequently—and rightfully—been disputed [20]. Professional appraisal of attractiveness performed by dentists or surgeon rely on taught rules of an ideal beauty, which often do not concur with the laymen’s or patient’s impression [20]. Moreover, many studies present panel-based assessments, which remain unavailable for the individual. Neural networks based on big data appear to be a promising tool to render additional and novel assessment methods of facial features, on an individual level, available. Sometimes, the processing pipeline will enable to produce a visualization of an outcome. Being able to score the degree of feminization after surgery, or to quantify the changes in age appearance or attractiveness, on an individual basis for each patient—independently of panels and not subject to the clinician’s expertise and experience—will indeed make health care safer, more personalized, and consolidate the participatory paradigm [5].

Revolutionized care

Neural networks permit the integration of diverse and complex data. Facial images can be annotated with various information from the patients’ dental or medical history, and then be linked to sociodemographic, social network, or bio-molecular data. New predictions can be conjectured on such vast and varied multi-level data, with the objective to discover new associations [28]. As an example, when ANN based on big data analyzed the outcome of orthognathic surgery, it became apparent that the therapy had a juvenescent effect, and that especially patients undergoing mandibular osteotomies were found to appear younger after the treatment [20]. Neural networks can not only be used to interconnect or link existing variables, but can also be computed to generate a visualization of an inexistent, projected treatment outcome [29, 30]. In what has been coined deep dreaming, trained deep neural networks are run in reverse, being commanded to adjust the original image slightly according to a determined output feature (Fig. 2). A network trained to predict the age can thus be used to modify the input image with exaggerated features related to age that the model learned during training (Fig. 3) [31]. It is therefore expected that AI will eventually facilitate predictive and preventive dental medicine [28], and ultimately be a key factor in the advancement of personalized medicine.

Fig. 2.

Fig. 2

An image generated through deep dreaming: a neural network for facial attractiveness is run in reverse, and activated features related to facial attractiveness are visualized, without actual input image

Fig. 3.

Fig. 3

Given any input image of a face, aging predictions can be made using deep learning models (https://arxiv.org/abs/1702.01983). Here, the aging effects of a person are being simulated with aligned and misaligned teeth (the latter using mock-up teeth). Each image shift to the right represents an aging factor of 10 years. Simulating the long-term effect of an intervention may be beneficial to the decision process on whether to undergo a certain treatment or not

Whatever the degree of change in dental care that may be pursued, the success of AI-assisted diagnostics and treatment planning evidently hinges on its integration in the digital dentistry workflow. It is only with this amalgamation that the desired advancements and improvements in dental medicine can be achieved.

AI applications in medico-dental diagnostics of the face: challenges

Transforming commercial AI software of the face to medical AI software

Over the last years, many commercial AI solutions for the analysis of the face have emerged, which allow the recognition of gender, age, attractiveness, mood and emotions, or race and ethnicity, all based on a 2D image of the face. Their widespread popularity has led some authors to advocate their use in a medical setting [4, 23, 24, 32]. In contrast, few AI solutions for facial images exist that were designed de novo exclusively for medical diagnostics, which were trained entirely and solely on medical images of the face (e.g., for facial palsy [33] or clefts [15] patients). These new solutions are usually less robust, owing to the smaller amount of data available, and often prone to selection bias (e.g., purely based on hospital data) [34]. Most consider therefore the benefit of commercial AI systems based on big data too important to be ignored. AI is the perfect tool to utilize the abundance of available facial images generated from various sources including social networks for predictive analytics. Many times, these facial images are not just available as raw data, but include an annotation for age, gender, or attractiveness, which makes their use even more appropriate. However, the applicability of commercial AI solutions, which are not designed for a medical propose, is on many levels very problematic. The architectural design of commercially available software may not be compatible or ideal for medical diagnostics and therapy assessments [33].

Addressing distributional shift

It is evident that the most desirable solution is offered when a network is trained on big data, but then re-trained and refined on medical images of the face. More specifically, when altering AI solutions to obtain a prediction of diagnostic value, the processing pipeline of the network must include training to eliminate or significantly reduce distributional shift, i.e., a mismatch between the data or environment the system is trained on and the data or environment used in operation [35]. AI solutions for attractiveness scoring based on facial images exemplify this issue very well. A commercial AI software trained on social media data originating from a dating platform offers the advantage of using big data, namely millions of individual ratings on a plethora of facial images. This may deliver a reliable portrait of what society currently considers attractive, but it does not necessarily reflect the opinion and judgment of experts. Substantial differences in scoring attractiveness indeed appear to exist between the AI-based prediction and the scores of medical professionals. In fact, even when laypeople (for whom no mismatch would be expected) are given the task to evaluate the attractiveness of faces in a medical environment, their score might differ from the AI processed score, too [21].

Addressing confounders

Accounting for confounding effects (or biases) is a serious challenge when using deep learning in medical imaging studies [36]. Confounders distort the feature extraction and thus the prediction. Certain distractors can easily be removed by training the network to exclude clues outside the region of interest. Others can be more difficult to filter. For facial images, the issue of emotions illustrates this problem nicely. Facial expressions in general, and emotions in particular, affect facial attractiveness: happiness increases facial attractiveness, and stress will reduce it. While there can be no objection if happiness increases attractiveness on an image featured on social media, in a diagnostic setting, such a distortion of face is considered a transient aspect that must be excluded in order not to impair the accuracy of the evaluation. Re-training a neural network on dedicated medical images taken under a controlled setting supports the network in recognizing emotions as distractors and such a re-training can reduce the bias caused by facial expressions [37].

Unsafe failure mode

An AI solution is able to calculate a prediction even in a case where the prediction will not be performed with high accuracy. Insufficient training or missing data can provoke insensible outcomes, and it is therefore of primordial importance that the prediction is supplemented with a “degree of confidence.” Best practice design would entail procedural precautions that would failsafe the system and make a prediction impossible when the system’s confidence is low on a prediction [38]. Unlike human performance, which is prone to significant inter- and intra-individual inconsistency, neural network will generate predictions with little variance and high repeatability, but this should not erroneously be interpreted as high confidence in a result.

Management of medical data used in AI software for facial imaging

With the use of deep learning in medical diagnostics, dynamic autonomous AI solutions and big data are being introduced to medicine, and conventional ethical and legal principles may therefore not be capable to ensure that all facets of accountability are regulated. In order to uphold the quality of the software and warrant its safeness as a medical tool, it will be necessary to establish new standards and stipulate them in quality management and patient safety plans. These plans must cover a wide variety of issues, and most of them are very acute when working with images of the face. As an example, the feeding of patients’ facial image in a neural network will raise a multitude of ethical and juridical concerns [39], among them:

  1. informed consent: should medical professionals disclose that the diagnostic outcome of multi-layer convolutional networks are not fully explainable?

  2. data privacy: should the medical professional disclose that the image of the patient and the obtained prediction will be used to further train the network?

  3. data security and cybersecurity: facial images are legally considered “particularly sensitive personal data” [40]. Thus, how can the required higher standards of privacy protection be met?

  4. liability: would a detrimental treatment based on an incorrect prediction of an AI-software be charged for medical malpractice (i.e., the dentist) or product liability (i.e., the company) [3]?

  5. intellectual proprietary right: who owns the data of a continuously evolving AI software? And may the owner dispose of the data freely?

While some have voiced that the demands for health care privacy may no longer be attainable with the enormous scale of data sharing [3], there is no question that the wider public is uncomfortable with the prospect of companies or even the government selling patient data for profit, sometimes for billions [41]. These are not hypothetical considerations. For example, the Royal Free NHS Foundation trust was found to have breached laws of data privacy when it shared personal data of 1.6 mio. patients to Google Deep Mind for the development of an app to diagnose acute kidney injury [39].

Necessary improvements

Interpretability and explainable AI

In machine learning, there is usually a trade-off between accuracy and interpretability: Rule-based systems are highly interpretable but not very precise. Using deep models, interpretable rules are traded for complex algorithms that achieve superior performance through greater abstraction (more layers) and tighter integration (end-to-end training). Thus, recently introduced deep networks are built sometimes on an uninterpretable amount of 200 layers in order to achieve a state-of-the-art performance in a variety of challenging tasks [42]. As such, the system is built to produce an inscrutable prediction that is not open to inspection or interpretation. To mitigate this so-called black box decision-making, explainable AI (XAI), or interpretable AI, aims to establish more transparency by making the results of the solution comprehensible to humans. For predictions based on image analysis, a broadly used approach in XAI is class activation mapping, i.e., generating a sort of heat map. The idea is to identify exactly which regions of an image are being used for discrimination [43]. Such heat maps have been advocated for medico-dental imaging [28] and have indeed already been applied to the analysis of cephalometric [44] or panoramic radiography [45]. The use of heat maps in image analyses of the face, be it for the diagnostic evaluation of attractiveness, gender, or age, would—at least in theory—allow to increase the interpretability of the system and confidence in the prediction [17]. But even with newest techniques of visual explanation, it is unrealistic to assume that the heat map is able to point to such fine-grained details that built the prediction for attractiveness, age, or gender (Fig. 4) [42].

Fig. 4.

Fig. 4

Data visualization method (“heat map”) based on Grad-CAM (https://arxiv.org/abs/1610.02391) to identify discriminative regions by magnitude, used by the convolutional neural network (CNN) for attractiveness of the face. As can be seen, it is unrealistic to assume that the heat map would be able to point to such fine-grained details and disclose on which the network focuses

Regulatory oversight

Expanding on the concerns outlined above related to the translation of commercial software to medical application, it seems crucial to establish regulation processes to overcome the obstacles associated to lack of transparency or to bias inherent in the data. Moreover, the algorithms should be monitored for their performance, robustness, and resilience to withstand changing clinical inputs and conditions [46].

Both the European Medicine Agency (EMA) and the US Food and Drugs Administration (FDA) have addressed the necessity to establish legal and ethical regulation processes in licensing AI-based medical devices. In a bold step, the FDA has acknowledged that the current regulations are not up to par for the modern techniques, and while the administration pro-actively presents in a published action plan concrete steps it intends to undertake [47], it also reaches out and requests feedback from the community [48]. In its action plan, the FDA proposes among other changes probably most importantly the introduction of a good machine learning practice scheme, the establishment of a public repository for approved AI-based software, and the revision of the regulatory oversight. The EU and the FDA stipulate that regulation and approval of AI-based software as medical devices should be effected according to the risk categorization described in the Medical Device Regulation (MDR), which came into effect in 2021 [39]. This regulation considers even a software that is solely used for the medical purpose of prediction or prognosis of disease as a medical device. The recommended categorization by MDR differentiates between the state of health care situation (critical, serious, non-serious), the intended use of the software-based information, and the level of impact (treat or diagnose, drive clinical management, mitigating public health). The use of most if not all AI-based algorithms for the face presented in this review would probably be considered to diagnose or drive clinical management of serious but non-critical cases.

Yet, the most serious impediment in regulatory oversight is the very essence of what makes AI so exceptional. The forte of neural networks is their ability to constantly improve by means of real-world training. As much as this perpetual learning process based on highly iterative and adaptive algorithms has made these technologies uniquely situated among software and has helped to open countless new avenues, it is exactly this autonomy that has created a novel state of affair that current regulatory frameworks were not designed for and are struggling to handle. In the past, the FDA has therefore mostly approved algorithms that were “locked” prior to marketing (meaning the algorithm is tampered and provides the identical outcome each time the same input is applied to it and does not change when exposed to real-world feedback) [48]. This wing-clipping is however absurd, as it literally removes the “intelligence” of AI, and it seriously curtails the inherent autonomy and the desired capability of AI-based software to learn and adapt in a real-world environment. Currently, the FDA is working on a framework described as “Predetermined Change Control Plan.” This plan intends to pre-define already in the conception phase the various anticipated modifications together with the associated methodology to be used, to allow these changes in a controlled manner [47]. Such a plan would enable to monitor an adaptive software product from its premarket development through post-market performance. This strategy would indeed provide a certain degree of assurance of safety and effectiveness while still acquiescing to the autonomy of artificial intelligence in medical devices. Time will tell whether the FDA will be able to have its cake and eat it.

The moral imperative

Any rise of novel technological possibilities in medicine is accompanied by new challenges regarding moral responsibilities that go far beyond the discussed technical and judicial requirements. If a moral compass is lacking, misuse cannot be prevented, as any regulations will be viewed as just an obstacle course. So, while thorough regulations are indisputably of great importance, and while the application of AI in medicine can be promoted to advance science or even the profit of shareholders, it should be guided and dictated by moral accountability. Everyone involved in the development of AI solutions in medicine is called upon this moral imperative.

As a case in point, current law largely fails to address discrimination when extracting information from big data [35, 39]. Trained to perform certain diagnostic or interventional tasks according to a predetermined goal, the AI solution will depend immensely on the input. Yet, the input is not value free. Let us consider the use of attractiveness—instead of beauty—to score medical images. Attractiveness is defined as the quality of being pleasing, and unlike beauty, attractiveness does not describe the face per se, but rather how the face is being perceived. This facilitates the task, because attractiveness can easily be rated and ranked by observers (i.e., perceiver), and these many observations can be fed to CNN. But it is vital to appreciate the fact that the ratings are not value free, as a specific combination of people being scored and people scoring the images is actually setting a norm. The use of a multitude of subjective opinions with big data will not make the score any more objective or free of bias. The bias of the individual will be replaced by the bias of the group, and a whole group might be biased toward a certain ethnicity, skin color, or lip volume.

Therefore, applying AI toward personalized medicine could also be considered a step in the wrong direction. Personalized orthognathic surgery, for example, would in many instances make faces less deviating, but not ideal. Striving for the highest attractiveness score might be incompatible with an individual’s perceived idea of beauty and might cause negative psychological effects owing to estrangement.

Research groups, software developers, shareholders, and practitioners all have a moral obligation to recognize that feeding big data to AI solution is a balancing act, and that prior to any endorsement, all involved people are called upon to question the moral quality of new norms that will inevitably be set with advancement of AI applications in medico-dental diagnostics.

Funding

Open access funding provided by University of Zurich

Declarations

Ethics approval

Not applicable.

Consent to participate

The person depicted in Figs. 1, 3, and 4 gave written informed consent for the published images.

Conflict of interest

The authors declare no competing interests.

Footnotes

Publisher's note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

References

  • 1.Redford DB. The Oxford Encyclopedia of Ancient Egypt. Oxford: Oxford University Press; 2001. [Google Scholar]
  • 2.Gysel C. Histoire de l’orthodontie: ses origines, son archéologie et ses précurseurs. Bruxelles: Societe Belge d’Orthodontie; 1997. [Google Scholar]
  • 3.Shan T, Tay FR, Gu L. Application of artificial intelligence in dentistry. J Dent Res. 2021;100:232–244. doi: 10.1177/0022034520969115. [DOI] [PubMed] [Google Scholar]
  • 4.Peck CJ, Patel VK, Parsaei Y, Pourtaheri N, Allam O, Lopez J, Steinbacher D. Commercial artificial intelligence software as a tool for assessing facial attractiveness: a proof-of-concept study in an orthognathic surgery cohort. Aesthetic Plast Surg. 2021 doi: 10.1007/s00266-021-02537-4. [DOI] [PubMed] [Google Scholar]
  • 5.Naylor CD. On the prospects for a (deep) learning health care system. JAMA. 2018;320:1099–1100. doi: 10.1001/jama.2018.11103. [DOI] [PubMed] [Google Scholar]
  • 6.Nakano Y, Suzuki N, Kuwata F. Predicting oral malodour based on the microbiota in saliva samples using a deep learning approach. BMC Oral Health. 2018;18:128. doi: 10.1186/s12903-018-0591-6. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7.Rasteau S, Sigaux N, Louvrier A, Bouletreau P. Three-dimensional acquisition technologies for facial soft tissues – applications and prospects in orthognathic surgery. J Stomatol Oral Maxillofac Surg. 2020;121:721–728. doi: 10.1016/j.jormas.2020.05.013. [DOI] [PubMed] [Google Scholar]
  • 8.Hung K, Montalvao C, Tanaka R, Kawai T, Bornstein MM. The use and performance of artificial intelligence applications in dental and maxillofacial radiology: a systematic review. Dentomaxillofac Rad. 2020;49:20190107. doi: 10.1259/dmfr.20190107. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9.Joda T, Yeung AWK, Hung K, Zitzmann NU, Bornstein MM. Disruptive innovation in dentistry: what it is and what could be next. J Dent Res. 2021;100:448–453. doi: 10.1177/0022034520978774. [DOI] [PubMed] [Google Scholar]
  • 10.Cireşan D, Meier U, Masci J, Schmidhuber J. Multi-column deep neural network for traffic sign classification. Neural Netw. 2012;32:333–338. doi: 10.1016/j.neunet.2012.02.023. [DOI] [PubMed] [Google Scholar]
  • 11.Zhu Y, Jiang Y. Optimization of face recognition algorithm based on deep learning multi feature fusion driven by big data. Image Vision Comput. 2020;104:104023. doi: 10.1016/j.imavis.2020.104023. [DOI] [Google Scholar]
  • 12.Zhao Z, Wu C-M, Zhang S, He F, Liu F, et al. A novel convolutional neural network for the diagnosis and classification of rosacea: usability study. JMIR Med Inform. 2021;9:e23415–e23415. doi: 10.2196/23415. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.Han SS, Moon IJ, Lim W, Suh IS, Lee SY, Na J-I, Kim SH, Chang SE. Keratinocytic skin cancer detection on the face using region-based convolutional neural network. JAMA Dermatol. 2020;156:29–37. doi: 10.1001/jamadermatol.2019.3807. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14.Haggenmüller S, Maron RC, Hekler A, Utikal JS, Barata C, et al. Skin cancer classification via convolutional neural networks: systematic review of studies involving human experts. Eur J Cancer. 2021;156:202–216. doi: 10.1016/j.ejca.2021.06.049. [DOI] [PubMed] [Google Scholar]
  • 15.McCullough M, Ly S, Auslander A, Yao C, Campbell A, Scherer S, Magee WP., 3rd Convolutional neural network models for automatic preoperative severity assessment in unilateral cleft lip. Plast Reconstr Surg. 2021;148:162–169. doi: 10.1097/PRS.0000000000008063. [DOI] [PubMed] [Google Scholar]
  • 16.Özdemir ME, Telatar Z, Eroğul O, Tunca Y. Classifying dysmorphic syndromes by using artificial neural network based hierarchical decision tree. Australas Phys Eng S. 2018;41:451–461. doi: 10.1007/s13246-018-0643-x. [DOI] [PubMed] [Google Scholar]
  • 17.Gurovich Y, Hanani Y, Bar O, Nadav G, Fleischer N, et al. Identifying facial phenotypes of genetic disorders using deep learning. Nat Med. 2019;25:60–64. doi: 10.1038/s41591-018-0279-0. [DOI] [PubMed] [Google Scholar]
  • 18.Su Z, Liang B, Shi F, Gelfond J, Šegalo S, Wang J, Jia P, Hao X. Deep learning-based facial image analysis in medical research: a systematic review protocol. BMJ Open. 2021;11:e047549. doi: 10.1136/bmjopen-2020-047549. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19.Jeong SH, Yun JP, Yeom H-G, Lim HJ, Lee J, Kim BC. Deep learning based discrimination of soft tissue profiles requiring orthognathic surgery by facial photographs. Sci Rep. 2020;10:16235–16235. doi: 10.1038/s41598-020-73287-7. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 20.Patcas R, Bernini DAJ, Volokitin A, Agustsson E, Rothe R, Timofte R. Applying artificial intelligence to assess the impact of orthognathic treatment on facial attractiveness and estimated age. Int J Oral Maxillofac Surg. 2019;48:77–83. doi: 10.1016/j.ijom.2018.07.010. [DOI] [PubMed] [Google Scholar]
  • 21.Patcas R, Timofte R, Volokitin A, Agustsson E, Eliades T, Eichenberger M, Bornstein MM. Facial attractiveness of cleft patients: a direct comparison between artificial-intelligence-based scoring and conventional rater groups. Eur J Orthod. 2019;41:428–433. doi: 10.1093/ejo/cjz007. [DOI] [PubMed] [Google Scholar]
  • 22.Obwegeser D, Timofte R, Mayer C, Eliades T, Bornstein MM, Schätzle MA, Patcas R (2022) Using artificial intelligence to determine the influence of dental aesthetics on facial attractiveness in comparison to other facial modifications. Eur J Orthod:[epub ahead of print]. 10.1093/ejo/cjac016 [DOI] [PubMed]
  • 23.Dorfman R, Chang I, Saadat S, Roostaeian J. Making the subjective objective: machine learning and rhinoplasty. Aesthet Surg J. 2019;40:493–498. doi: 10.1093/asj/sjz259. [DOI] [PubMed] [Google Scholar]
  • 24.Chen K, Lu SM, Cheng R, Fisher M, Zhang BH, Di Maggio M, Bradley JP. Facial recognition neural networks confirm success of facial feminization surgery. Plast Reconstr Surg. 2020;145:203–209. doi: 10.1097/prs.0000000000006342. [DOI] [PubMed] [Google Scholar]
  • 25.Bargshady G, Zhou X, Deo RC, Soar J, Whittaker F, Wang H. Ensemble neural network approach detecting pain intensity from facial expressions. Artif Intell Med. 2020;109:101954. doi: 10.1016/j.artmed.2020.101954. [DOI] [PubMed] [Google Scholar]
  • 26.Baumgartl H, Flathau D, Bayerlein S, Sauter D, Timm IJ, Buettner R (2021) Pain level assessment for infants using facial expression scores. 2021 IEEE 45th Annual Computers, Software, and Applications Conference (COMPSAC):590–598. 10.1109/compsac51774.2021.00087
  • 27.Jeon T, Bae HB, Lee Y, Jang S, Lee S. Deep-learning-based stress recognition with spatial-temporal facial information. Sensors. 2021;21:7498. doi: 10.3390/s21227498. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 28.Schwendicke F, Samek W, Krois J. Artificial intelligence in dentistry: chances and challenges. J Dent Res. 2020;99:769–774. doi: 10.1177/0022034520915714. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 29.Tanikawa C, Yamashiro T. Development of novel artificial intelligence systems to predict facial morphology after orthognathic surgery and orthodontic treatment in Japanese patients. Sci Rep. 2021;11:15853–15853. doi: 10.1038/s41598-021-95002-w. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 30.ter Horst R, van Weert H, Loonen T, Bergé S, Vinayahalingam S, Baan F, Maal T, de Jong G, Xi T. Three-dimensional virtual planning in mandibular advancement surgery: soft tissue prediction based on deep learning. J Craniomaxillofac Surg. 2021;49:775–782. doi: 10.1016/j.jcms.2021.04.001. [DOI] [PubMed] [Google Scholar]
  • 31.Palsson S, Agustsson E, Timofte R, Gool LV (2018) Generative adversarial style transfer networks for face aging. 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW):2165–21658. 10.1109/cvprw.2018.00282
  • 32.Rezende Machado AL, Dezem TU, Bruni AT, Alves da Silva RH. Age estimation by facial analysis based on applications available for smartphones. J Forensic Odontostomatol. 2017;35:55–65. [PMC free article] [PubMed] [Google Scholar]
  • 33.Guarin DL, Yunusova Y, Taati B, Dusseldorp JR, Mohan S, et al. Toward an automatic system for computer-aided assessment in facial palsy. Facial Plast Surg Aesthet Med. 2020;22:42–49. doi: 10.1089/fpsam.2019.29000.gua. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 34.Gianfrancesco MA, Tamang S, Yazdany J, Schmajuk G. Potential biases in machine learning algorithms using electronic health record data. JAMA Intern Med. 2018;178:1544–1547. doi: 10.1001/jamainternmed.2018.3763. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 35.Challen R, Denny J, Pitt M, Gompels L, Edwards T, Tsaneva-Atanasova K. Artificial intelligence, bias and clinical safety. BMJ Qual Saf. 2019;28:231–237. doi: 10.1136/bmjqs-2018-008370. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 36.Zhao Q, Adeli E, Pohl KM. Training confounder-free deep learning models for medical applications. Nat commun. 2020;11:6010. doi: 10.1038/s41467-020-19784-9. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 37.Obwegeser D (2021) How facial attractiveness is perceived differently when deep convolutional neural networks are fine-tune to assess medical images. Master thesis No. 15-738-313, University of Zurich, Zurich
  • 38.Varshney KR (2016) Engineering safety in machine learning. 2016 Information Theory and Applications Workshop (ITA):1–5. 10.48550/arXiv.1601.04126
  • 39.Gerke S, Minssen T, Cohen G (2020) Ethical and legal challenges of artificial intelligence-driven healthcare. In: Artificial Intelligence in Healthcare, 295–336. 10.1016/b978-0-12-818438-7.00012-5
  • 40.Lopes IM, Guarda T, Oliveira P. General data protection regulation in health clinics. J Med Syst. 2020;44:53. doi: 10.1007/s10916-020-1521-0. [DOI] [PubMed] [Google Scholar]
  • 41.Moberly T. Should we be worried about the NHS selling patient data? BMJ. 2020;368:m113. doi: 10.1136/bmj.m113. [DOI] [PubMed] [Google Scholar]
  • 42.Selvaraju RR, Cogswell M, Das A, Vedantam R, Parikh D, Batra D. Grad-CAM: visual explanations from deep networks via gradient-based localization. Int J Comput Vision. 2020;128:336–359. doi: 10.1007/s11263-019-01228-7. [DOI] [Google Scholar]
  • 43.Zhou B, Khosla A, Lapedriza À, Oliva A, Torralba A (2016) Learning deep features for discriminative localization. 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR):2921–2929. 10.48550/arXiv.1512.04150
  • 44.Yu HJ, Cho SR, Kim MJ, Kim WH, Kim JW, Choi J. Automated skeletal classification with lateral cephalometry based on artificial intelligence. J Dent Res. 2020;99:249–256. doi: 10.1177/0022034520901715. [DOI] [PubMed] [Google Scholar]
  • 45.Kim S, Lee Y-H, Noh Y-K, Park FC, Auh QS. Age-group determination of living individuals using first molar images based on artificial intelligence. Sci Rep. 2021;11:1073–1073. doi: 10.1038/s41598-020-80182-8. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 46.Benjamens S, Dhunnoo P, Meskó B. The state of artificial intelligence-based FDA-approved medical devices and algorithms: an online database. npj Digit Med. 2020;3:118. doi: 10.1038/s41746-020-00324-0. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 47.U.S. Food and Drug Administration (2021) Artificial intelligence/machine learning (AI/ML)-based software as a medical device (SaMD) action plan. https://www.fda.gov/medical-devices/software-medical-device-samd/artificial-intelligence-and-machine-learning-software-medical-device#regulation. Accessed on: 27.02.2022
  • 48.U.S. Food and Drug Administration (2019) Proposed regulatory framework for modifications to artificial intelligence/machine learning (AI/ML)-based software as a medical device (SaMD) - Discussion Paper and Request for Feedback. https://www.fda.gov/medical-devices/software-medical-device-samd/artificial-intelligence-and-machine-learning-software-medical-device#regulation. Accessed on: 27.02.2022

Articles from Clinical Oral Investigations are provided here courtesy of Springer

RESOURCES