Skip to main content
Dentistry Journal logoLink to Dentistry Journal
. 2023 Feb 8;11(2):43. doi: 10.3390/dj11020043

Artificial Intelligence in Periodontology: A Scoping Review

James Scott 1,*, Alberto M Biancardi 2, Oliver Jones 1, David Andrew 1
Editors: Christos Rahiotis, Nikolaos Gkantidis
PMCID: PMC9955396  PMID: 36826188

Abstract

Artificial intelligence (AI) is the development of computer systems whereby machines can mimic human actions. This is increasingly used as an assistive tool to help clinicians diagnose and treat diseases. Periodontitis is one of the most common diseases worldwide, causing the destruction and loss of the supporting tissues of the teeth. This study aims to assess current literature describing the effect AI has on the diagnosis and epidemiology of this disease. Extensive searches were performed in April 2022, including studies where AI was employed as the independent variable in the assessment, diagnosis, or treatment of patients with periodontitis. A total of 401 articles were identified for abstract screening after duplicates were removed. In total, 293 texts were excluded, leaving 108 for full-text assessment with 50 included for final synthesis. A broad selection of articles was included, with the majority using visual imaging as the input data field, where the mean number of utilised images was 1666 (median 499). There has been a marked increase in the number of studies published in this field over the last decade. However, reporting outcomes remains heterogeneous because of the variety of statistical tests available for analysis. Efforts should be made to standardise methodologies and reporting in order to ensure that meaningful comparisons can be drawn.

Keywords: periodontology, artificial intelligence, convolutional neural networks, radiography

1. Introduction

Artificial intelligence (AI) aims to develop computer systems that can mimic human behaviour using machines. Within medicine and dentistry, commentators predicted as early as the 1970s that AI would bring clinical careers to an end [1]; however, this has not been the case. Science fiction will present AI as a comprehensive overarching intelligence [2], but this is far from the truth. Thus far, AI development has proved successful in solving problems in specific areas by learning distinct thinking mechanisms and perceptions.

Periodontitis is the sixth most prevalent disease worldwide. It is characterised by microbially associated, host-mediated inflammation that results in loss of alveolar bone and periodontal attachment, which can lead to tooth loss [3]. This disease has a well-reported but complex relationship with a number of other physiological systems leading to detrimental effects on quality of life and general health [4]. Further to this, a bi-directional relationship between systemic conditions, including chronic inflammatory disease such as diabetes [5,6] and atherosclerosis [7], has been shown.

Periodontitis is also challenging for clinicians to accurately recognise and diagnose [8]. Best practice currently focuses on measuring soft tissues with a graduated probe [9] and assessing hard tissues with radiographic imaging [10]. However, these methods have poor inter- and intra-operator reliability due to variations in probing pressure and radiographic angulation [8].

As such, the study of periodontitis presents a diagnostic challenge linked to a disease process of complex relationships between predisposing factors that are difficult for clinicians and scientific processes to fully comprehend. These complex factors lend the study of this disease to the application of AI to best comprehend how these factors affect the diagnostics or understanding of its aetiology.

It is important to differentiate AI from traditional software development. In the traditional approach to software development, the researchers identify a series of processing steps and, optionally, a data-dependent strategy to reach the results. This is best described as an input ‘A’ is received, it is computed through the pre-defined strategy of sub-tasks, and an output ‘B’ is returned. As such, whilst this performs incredibly useful tasks for humankind, it requires vast amounts of effort to perform complex tasks and risks providing only limited availability for adaptation to unseen scenarios. Artificial intelligence, in contrast, has a different mode of work. When developing an AI-based tool, both the input ‘A’ and the required output ‘B’ are provided; the AI approach will then tune the tool to leverage the link between inputs and outputs, which can then be used on new (unseen) data sets, typically with remarkable performance [11].

AI is ever-increasing in medicine and dentistry as an assistive tool, becoming a central tenet in providing safe and effective healthcare. More recently, deep learning (DL) has been the mainstay of this endeavour, mainly through its applications stemming from the use of artificial neural networks (ANN) that exhibit a very high degree of complexity [12], where large numbers of artificial neurons (or nodes) are connected into layers and several hundreds, or thousands, of layers are assembled into specific structures called architectures. DL networks can assess large volumes of data to perform specific tasks, among which electronic health records, imaging data, wearable-device sensor collections, and DNA sequencing play a prominent role. Within medical fields, these are classically used for computer-aided diagnosis, personalised treatments, genomic analysis, treatment response assessment.

When images are used as input data, we as humans perceive digital images as analogues or as a continual flow of information. Still, a digital (planar) image is nothing more than a collection of millions of tiny points of colour or pixels, each with their 2D location. Thus, these pixel series can be viewed as strings of values with additional information about their neighbouring locations that software can process efficiently.

A subset of ANNs, a convolutional neural network (CNN), is specifically designed for handling imaging data. The CNN concept was developed to replicate the visual cortex and differentiate patterns in an image [13]. Classic neural networks typically need to consider each pixel individually to process an image and therefore are heavily constrained in the size of the images that can be analysed. CNNs, on the other hand, are capable of working with the image data in their spatial layout; their output is a new set of data replicating the original layout of the image while increasing or condensing the information stored at each location. This process is similar to applying several digital filters to an image to ‘highlight’ key features that will collectively help perform the task at hand, e.g., select distinct aspects of an object to identify its presence inside the image.

CNN-based architectures will often have multiple layers, or multiple levels at which these transformations are applied. Early layers will focus on picking up gross content such as edges, gradient orientation, and colour, with later layers focusing on higher-level (more task specific) features. This kind of approach is usually called an encoder because the iconic information inside the image is transformed into a more abstract, symbolic representation, and this is achieved by juxtaposing CNN-based blocks that progressively reduce the size of the image being processed while, concurrently, increasing the number of the channels, i.e., the number of values associated with each image pixel. The complementary approach to an encoder is a decoder where the abstract information is transformed into an iconic representation by successively increasing the image size while reducing the channel number. A common pattern in ANN architectures based on CNN layers is to have an encoder section, a decoder section, or both; for instance, the U-Net architecture [14], which is one of the most used approaches when segmenting imaging data, is structured as an encoder section followed by a decoder section to achieve a transformation of the image information from iconic to abstract and then back to iconic while performing the task at hand.

Therefore, CNNs can be taught to recognise specific collections of pixel/location pairs and subsequently find similar patterns in new image datasets. For CNNs to perform this function, they require a so called ‘training’ stage. Training is a process whereby humans identify target subsets and show a CNN what to look for. In image analysis, this often relies on experts labelling image sections of interest so that CNN can find similar regions in the future. For example, radiologists would draw around these nodules for a CNN to recognise nodules on CT images of a lung to show their correct extent. As the CNN is shown more nodules, it will become more capable of identifying similar regions. This process, leading to the software’s ability to carry out the nodule localisation independently, is called supervised learning [14].

Whilst in some cases CNNs can be definitive in their image recognition, they are more commonly used as assistive tools, whereby AI can highlight the areas that are likely to contain the sought pathology or image type. This has been demonstrated in radiology (for detecting abnormalities within chest X-rays) [15], dermatology (to detect lesions of oncological potential) [16], and ophthalmology (to detect specific types of retinopathy) [17]. It has been suggested, however, that within these fields, some of the work may lack the robustness to be truly generalisable to all clinical situations or, indeed, to be as accurate as medical professionals [18].

Within dentistry, radiographs are combined with a full clinical examination and special tests to aid in assessment, diagnosis, and treatment planning. The type of radiograph taken depends upon the disease or pathology being investigated or the procedure being undertaken, and may include bitewing, periapical, or orthopantomography [19]. Common justifications for taking dental radiographs include the diagnosis of dental caries staging and grading of periodontal disease [20], detection of apical pathology [21], or in the assessment of peri-implant health. A dentist will then report on these images. Still, it has been shown that in detecting both dental caries and periodontal bone loss, inter-ratee and intra-rater agreement is poor [22,23], lending this analysis to CNN assistance.

Image-based diagnostics is not the only area in which CNNs are currently used within medicine and dentistry. AI’s ability to assess large volumes of data for regression purposes lends itself to data analysis within larger data fields where traditional methods struggle. In medicine, success has been achieved using patient health records and metabolite data to predict Alzheimer’s disease, depression, sepsis, and dementia [11]. As the data processing technology has improved, specific to medicine and dentistry, a term introduced by the American Medical Association is that of ‘augmented intelligence’. This describes a conceptualisation of AI in healthcare, highlighting its assistive role to medical professionals.

Whilst the benefits of utilising AI within healthcare can be clear to see, for example reducing human error, assisting in diagnosis, and streamlining data analysis and task performance, which may ultimately lead to more efficient and cost-effective services, its adoption is not without issue. Specifically, these can include a lack of data curation, hardware, code sharing, and readability [24], as well as the inherent issues of introducing any new technology within a service such as reluctance to change, embedding new technology within current infrastructures, and ongoing cost maintenance. More recent criticism addressed the presence of implicit biases in the training datasets and the direct consequences in AI performance [25].

Within periodontology and implantology, AI is still in its relative infancy and has not yet been used to its full potential. With the advantages of diagnostic assistance, data analysis, and detailed regression, it would appear that much could be gained through applying this tool. Given the relative paucity of literature in the subject area, this scoping review aimed to assess the current evidence on the use of artificial intelligence within the field of periodontics and implant dentistry. This would both describe current practice and guide further research in this field.

2. Materials and Methods

This prospective scoping review was conducted by considering the original guidance of Arksey et al. [26] and more recent guidance from Munn et al. [27].

2.1. Focused Question and Study Eligibility

The focused question used for the current literature search was “What are the current clinical applications of machine learning and/or artificial intelligence in the field of periodontology and implantology?”

The secondary questions were as follows:

  1. Which methods were used in these studies to establish datasets; develop, train, and test the model; and report on its performance?

  2. In cases where these models were tested against human performance:
    • a.
      What metrics were used to compare performance?
    • b.
      What were the outcomes?

The inclusion criteria for the studies:

  1. Original articles published in English.

  2. Implant- and periodontal-based literature using ML or AI models for diagnostic purposes, detection of abnormalities/pathologies, patient group analysis, or planning of surgical procedures.

  3. Study designs whereby the use of ML or AI was used as the independent variable.

The exclusion criteria for the studies:

  1. Studies not in English.

  2. Studies using classic software rather than CNN derivative protocols for machine-based learning.

  3. Studies using AI for purposes other than periodontology and peri-implant health.

2.2. Study Search Strategy and Process

An electronic search was performed via the following databases:

  • Medline—the most widely used medical database for publishing journal articles. The search strategy for this is outlined in Figure 1.

  • Scopus—the largest database of scientific journals.

  • CINAHL—an index that focuses on allied health literature.

  • IEEE Xplore—a digital library that includes journal articles, technical standards, conference proceedings, and related materials on computer science.

  • arXiv—arXiv is an open-access repository of electronic preprints and postprints approved for posting after moderation but not peer review.

  • Google Scholar—Google Scholar is a freely accessible web search engine that indexes the full text or metadata of scholarly literature across an array of publishing formats.

Figure 1.

Figure 1

Medline via OVID search strategy.

This electronic search was supplemented with hand searches of included texts references lists. The search strategy was compiled in collaboration with the librarian at the University of Sheffield Medical Library. Keywords were a combination of Medical Subject Headings (MeSH) terms and frank descriptors, which were employed to reflect the intricacies of each database. The publication period was set to 20 years and was restricted to literature in English. All articles were included for initial review in line with the scoping review methodology. Records were collated in reference manager software (EndnoteTM; Version: 20, Clarivate Analytics, New York, NY, USA) [28], and the titles were screened for duplicates.

2.3. Study Selection

A single reviewer screened titles and abstracts. For records appearing to meet inclusion criteria, or where there was uncertainty, full texts were reviewed to determine their eligibility. This was completed twice by the reviewer at a 2-month interval to assess intra-rater agreement. Additional manual hand searching was performed of included full-text articles with reference lists from these studies included. These selected full texts were similarly read, and suitability for inclusion was as per the original criteria.

2.4. Data Extraction and Outcome of Interest

Data was extracted from the studies and recorded in a tabulated form. The standardised data collation sheet included the author title, year of publication, data format, application of ML/AI technique, the workflow of the ML/AI model, the subsequent training/testing datasets, the validation technique, the form of comparison used, and then some description of the performance of the AI model. The primary outcome of interest was the scope of current clinical applications of ML/AI in the field of periodontology and peri-implant health and the performance of these AI models in clinician or patient assistance. As this was a scoping review, all texts meeting eligibility criteria were subject to qualitative review.

3. Results

3.1. Study Selection and Data Compilation

The study selection process is outlined in Figure 2. A total of 401 articles were identified for screening after the removal of duplicates. In total, 293 texts were excluded after screening for factors not meeting inclusion criteria. Full-text review and hand searching identified 50 studies for inclusion in the qualitative analysis. Most of these excluded articles tested software rather than an AI architecture to assess the input data.

Figure 2.

Figure 2

Study selection flowchart.

The included articles were appraised with key information compiled into a single data table (Table 1) for display in text format.

3.2. Location of Research

The first authors were from a wide variety of locations, illustrated in Figure 3. However, over half were from institutions in the USA (n = 12), China (n = 11), or South Korea (n = 7). The remainder were from Europe (n = 6), the Middle East (n = 4), the Indo-Pacific (n = 4), Brazil (n = 3), Turkey (n = 2), and Canada (n = 1). There is also strong evidence of international collaboration, with the first and last authors’ locations being geographically disparate in a number of cases (n = 8).

Figure 3.

Figure 3

Location of research.

3.3. Year of Publication

Figure 4 depicts the year that studies were published. The first was in 2014, with a steady increase in the frequency of publication over the next decade. This number appears to be stabilising at 14–15 per year post-2020, in line with six found in 2022 prior to April when the searches were conducted.

Figure 4.

Figure 4

Publication of research.

3.4. Input Data

All included studies had a periodontal focus; however, the input data varied significantly (Figure 5). The majority (68%) of studies focused on imaging data, using either photographs (n = 12), radiographs (n = 20), or ultrasound images (n = 2). Patient data (Electronic Health Record) were used in to attempt to predict periodontal or dental outcomes (n = 7). Metabolites and saliva markers were used to classify, diagnose, and predict periodontal and dental outcomes (n = 7).

Figure 5.

Figure 5

Input data.

3.5. Datasets

The dataset size had a large span due to the variability of inputs and outcomes of the study types. Datasets for image processing studies ranged from 30 to 12,179, with a mean of 1774 for solely panoramic radiographs, 1064 for solely periapical radiographs, and 1431 for photographs. Due to the novel nature of ultrasound imaging, the two included studies contained less data (n = 35 and 627). Patient datasets varied between 216 and 41,543 with no meaningful descriptive statistics due to outcome and methodological heterogeneity. For all imaging studies, a mean of 1666 images were used.

However, it is worth noting that the number of images included in studies with relatively heterogenous image types (i.e., plain film radiography) was not distributed in a Gaussian or normally distributed curve, with the median being significantly different to the mean. This was demonstrated with median (585 images) divergent from the mean (1940 images). Figure 6 shows the relative frequency peaks for visual analysis.

Figure 6.

Figure 6

Numbers of images used in radiographic training datasets.

3.6. ML Architectures

A wide variety of convolutional architectures were used in this literature body (n = 67). The most common architecture described was a U-Net (n = 9). Several other architectures were used, including ResNet (n = 6), GoogLeNet Inception (n = 4), R-CNN or Faster R-CNN’s (n = 4), and AlexNet (n = 2).

For image recognition, there appears to have been a shift to the use of U-Net over other architectures with all nine studies utilising or comparing this platform published in 2021 or 2022 [29,30,31,32,33,34,35,36,37]. When compared against other architectures, Dense U-Net out performed a standard U-Net [35], or U-Net was found to be optimised with a ResNet Encoder [36].

A number of studies involved patient data (e.g., metabolites or Electronic Health Record (EHR) data used support vector machine algorithms (SVM) [38,39,40,41,42,43]. The difference in data studies showed significant heterogeneity in methods, observations, and outcomes, and as such, no relevant statistical outcomes can be formed. However, descriptively, SVM showed either comparative outcomes [39] or reduced predictive capabilities [43] when compared to other ML formats such as ANN multilayer perceptron (MLP), random forest (RF), or naïve Bayes (NB).

3.7. Training and Annotation

The majority of patient data research in this literature body focused on using CNNs to assist with regression analyses. In these cases, training is not required as the models are sourcing regression analyses rather than replicating human activity.

In the case of image data processing, CNNs are mimicking humans and, as such, require training. General dentists performed training annotation in 47% of papers (n = 16), some form of specialist dentist or trainee specialist dentist performed training annotation in 16% of papers (n = 6), and radiologists in 8% (n = 3) of studies. Dental hygienists were also used (n = 1), as well as mixtures of clinicians (n = 6).

The methods of labelling were, in the majority, manual annotation by drawing or labelling the external pixels of required features (n = 25). However, in a more recent paper [44], a process of ‘dye staining’ was used, whereby annotators merely highlighted areas of interest with a CNN used to ascertain characteristics around these single or multiple-point annotations had occurred.

3.8. Outcome Metrics and Comparative Texts

As is the nature of a scoping review there is vast heterogeneity in data forms, methodologies which result in very different outcome metrics [26,27]. These included an array of best fit measurements including F1 and F2 scores, precision, and accuracy, alongside sensitivity and specificity. Area under the curve analyses and ICC between test sets and representative expert labels were also frequently quoted. More specific imaging outcomes were also used with Jaccard’s Index, Pixel Accuracy, and Hausdorff Difference utilised. This made a meaningful statistical comparison of outcomes difficult due to the vast number of analyses presented.

Descriptively, as one would expect, when ML was asked to produce nominal outcomes, accuracy increased. In cases where outcomes were dichotomous, 90–98% accuracy was reported [45]. Within research that is more image focused, this is reflected when the task is more simple, such as in the work of Kong et al., with gross recognition of periodontal bone loss reporting an accuracy of 98% [46].

However, as the task asked of the ML increases in complexity, the accuracy was shown to drop. This is possibly best illustrated by the eloquent research of Lee et al. [36]. This study assessed several parameters to provide best outcomes for the radiographic staging of periodontitis, showing a U-Net with ResNet Encoder-50 for the majority of its image analysis. A power calculation was performed for gross bone loss. Here, Lee et al. describes greatest accuracy (0.98) where there was no bone loss, with reduction in accuracy (0.89) for fine increments such as for minimal bone loss (stage 1).

Table 1.

Description of included studies.

Study Country Year Data Type Subject Total ML
Architecture
Annotators Performance Comparison CNN Performance Comment Brief Description
1 Papantonopoulos [45] Greece 2014 Patient data 29 MLP ANN n.a. Not comparative ANN’s gave 90–98% accuracy in classifying patients into AgP or CP. ANNs used to classify periodontitis by immune response profile to aggressive periodontitis (AgP) or chronic periodontitis (CP) class.
2 Bezruk [47] Ukraine 2017 Saliva 141 CNN
(no description)
n.a. Not comparative Precision of CNN 0.8 in predicting gingivitis based upon crevicular fluid markers. CNN was used for the learning task to build an information model of salivary lipid peroxidation and periodontal status and to evaluate the correlation between antioxidant levels in unstimulated saliva and inflammation in periodontal tissues.
3 Rana [48] USA 2017 Photographs 405 CNN Autoencoder Dentist Not comparative AU ROC curve of 0.746 for classifier to distinguish between inflamed and healthy gingiva. Machine learning classifier used to provide pixel-wise inflammation segmentations from photographs of colour-augmented intraoral images.
4 Feres [41] Brazil 2018 Plaque 435 SVM n.a. Not comparative AUC > 0.95 for SVM to distinguish between disease and health. AUC for ability to distinguish between CP and AgP was 0.83. SVM was used to assess whether 40 bacterial species could be used to classify patients into CP, AgP, or periodontal health.
5 Lee [49] South Korea 2018 Periapical 1740 CNN encoder
+ 3 dense layers
Periodontist Periodontist CNN showed AU ROC curve of 73.4–82.6 (95% CI 60.9–91.1) in predicting hopeless teeth. The accuracy of predicting extraction was evaluated and compared between the CNN and blinded board-certified periodontists using 64 premolars and 64 molars diagnosed as severe n the test dataset. For premolars, the deep CNN had an accuracy of 82.8%
6 Yoon [50] USA 2018 Patient data 4623 Deep neural network-BigML n.a. Not comparative DNN used as multi-regressional tool found correlation between ageing and mobility. 78 variables assessed by DNN were used to find a correlation that can predict tooth mobility.
7 Aberin [51] Philippines 2019 Plaque 1000 AlexNet Pathologists Not comparative Accuracy in predicting health or periodontitis from plaque slides reported at 75%. CNN was used to classify which microscopic dental plaque images were associated with gingival health.
8 Askarian [52] USA 2019 Photographs 30 SVM n.a. Not comparative 94.3% accuracy of SVM in detection of periodontal infection. Smartphone-based standardised photograph detection using CNN to classify gingival disease presence.
9 Duong [31] Canada 2019 Ultrasound 35 U-Net n.a. Orthodontist CNN yielded 75% average dicemetric for ultrasound segmentation. The proposed method was evaluated over 15 ultrasound images of teeth acquired from porcine specimens.
10 Hegde [39] USA 2019 Patient data 41,543 SVM n.a. Not comparative Comparison of ML vs. MLP vs. RF vs. SVM for data analysis. Similar accuracy found between all methods. The objective was to develop a predictive model using medical-dental data from an integrated electronic health record (iEHR) to identify individuals with undiagnosed diabetes mellitus (DM) in dental settings.
11 Joo [53] South Korea 2019 Photographs 451 CNN encoder
+ 1 dense layer
n.a. Not comparative Reported CNN accuracy of 70–81% for validation data. Descriptive analysis of preliminary data for concepts of imaging analysis.
12 Kim [54] South Korea 2019 Panoramic 12,179 DeNTNet Hygienists Hygienists Superior F1 score (0.75 vs. 0.69), PPV (0.73 vs. 0.62), and AUC (0.95 vs. 0.85) for balanced setting DeNTNet vs. clinicians for assessing periodontal bone loss. CNN used to develop an automated diagnostic support system assessing periodontal bone loss in panoramic dental radiographs.
13 Krois [55] Germany 2019 Panoramic 85 CNN encoder
+ 3 dense layers
Dentist Dentists CNN performed less accurately than the original examiner segmentation and independent dentists’ observers. CNNs used to detect periodontal bone loss (PBL) on panoramic dental radiographs.
14 Moriyama [56] Japan 2019 Photographs 820 AlexNet Dentist Not comparative Changes in ROC curves can have a significant effect on outcomes—looking at predicted pocket depth photographs and distorting images to improve accuracy. CNN was used to establish if there is a correlation between pocket depth probing and images of the diseased area.
15 Yauney [57] USA 2019 Patient data 1215 EED-net
(custom net)
Dentist Not comparative AUC of 0.677 for prediction of periodontal disease based on intraoral fluorescent porphyrin biomarker imaging. CNN was used to establish a link between intraoral fluorescent porphyrin biomarker imaging, clinical examinations, and systemic health conditions with periodontal disease.
16 Alalharith [58] Saudi Arabia 2020 Photographs 134 Faster R-CNN Dentist Previously published outcomes Faster R-CNN had tooth detection accuracy of 100% to determine region of interest and 77.12% accuracy to detect inflammation. An evaluation of the effectiveness of deep learning based CNNs for the pre-emptive detection and diagnosis of periodontal disease and gingivitis by using intraoral images.
17 Bayrakdar [59] Turkey 2020 Panoramic 2276 GoogLeNet Inception v3 Radiologist and periodontist Radiologist and periodontist CNN showed 0.9 accuracy to detect alveolar bone loss. CNN used to detect alveolar bone loss from dental panoramic radiographic images.
18 Chang [60] South Korea 2020 Panoramic 340 ResNet Radiologist Radiologists 0.8–0.9 agreement between radiologists and CNN performance. Automatic method for staging periodontitis on dental panoramic radiographs using the deep learning hybrid method.
19 Chen [61] China (and the UK) 2020 Photographs 180 ANN
(no description)
n.a. Not comparative ANN accuracy of 71–75.44% for presence of gingivitis from photographs. Visual recognition of gingivitis testing a novel ANN for binary classification exercise—gingivitis or healthy.
20 Farhadian [38] Iran 2020 Patient data 320 SVM n.a. Not comparative The SVM model gave an 88.4% accuracy to diagnose periodontal disease. The study aimed to design a support vector machine (SVM)-based decision-making support system to diagnose various periodontal diseases.
21 Huang [40] China 2020 Gingival crevicular fluid 25 SVM n.a. n.a. Classification models achieved greater than or equal to 91% in classifying SP patients, with LDA being the highest at 97.5% accuracy. This study highlights the potential of antibody arrays to diagnose severe periodontal disease by testing five models (SVM, RF, kNN, LDA, CART).
22 Kim [42] South Korea 2020 Saliva 692 SVM n.a. n.a. Accuracy ranged from 0.78 to 0.93 comparing neural network, random forest, and support vector machines with linear kernel, and regularised logistic regression in the R caret package. CNN was used to assess whether biomarkers can differentiate between healthy controls and those with differing severities of periodontitis.
23 Kong [46] China 2020 Panoramic 2602 EED-net
(custom net)
Expert? Expert? The custom CNN performed better than U-Net or FCN-8, all with accuracies above 98% for anatomical segmentation. CNN was used to complete maxillofacial segmentation of images, including periodontal bone loss recognition.
24 Lee [62] South Korea 2020 Periapicals and panoramic 10,770 GoogLeNet Inception v3 Periodontist Periodontist The CNN (0.95) performed better than human (0.90) for OPGs, but the same for PAs (0.97). CNN used for identification of implants systems and their associated health.
25 Li [63] Saudi Arabia 2020 Panoramic 302 R-CNN Dentist Other CNNs and dentist Proposed architecture gave accuracies of 93% for detecting no periodontitis, 89% for mild, 95% for moderate, and 99% for severe. This study compared different CNN models for bone loss recognition.
26 Moran [64] Brazil 2020 Periapicals 467 ResNet, Inception Radiologist and dentist Compares two CNN approaches for accuracy AUC ROC curve for ResNet and Inception was 0.86 for identification of regions of periodontal bone destruction. Assessment of whether a CNN can recognise of periodontal bone loss improve post-image enhancement?
27 Romm [65] USA 2020 Metabolites N/A CNN
(No description),
PCA
n.a. n.a. Oral cancer identified rather than a periodontal disease with 81.28% accuracy. CNN to analyse metabolite sets for different oral diseases to distinguish between different forms of oral disease.
28 Shimpi [43] USA 2020 Patient data N/A SVM, ANN n.a. n.a. ANN presented more reliable outcomes than NB, LR, and SVM. The study reviewed classic and CNN regression to assess accuracy in prediction for periodontal risk assessment based on EHIR.
29 Thanathornwong [66] Thailand 2020 Panoramic 100 Faster R-CNN Periodontist Periodontist 0.8 precision for identifying periodontally compromised teeth using radiographs. CNN used to assess periodontally compromised teeth on OPG.
30 You [67] China 2020 Photographs 886 DeepLabv3+ Orthodontist Orthodontist No statistically significant difference in the ability to discern plaque on photographs compared to clinician. CNN used to assess plaque presence in paediatric teeth.
31 Cetiner [68] Turkey 2021 Patient data 216 MLP ANN n.a. n.a. The DT was most accurate, with accuracy of 0.871 compared to LR (0.832) and LP (0.852). Assessment of three models of data mining to provide a predictive decision model for peri-implant health.
32 Chen [69] China 2021 Periapicals 2900 r-CNN Dentist n.a. CNN used to locate periodontitis, caries, and PA pathology on PAs.
33 Danks [70] UK 2021 Periapicals 340 ResNet Dentist Dentist Predicting periodontitis stage accuracy of 68.3%. CNN used to find bone loss landmarks using different tools to provide staging of disease.
34 Kabir [30] USA 2021 Periapicals 700 Custom CNN combining Res-Net and U-Net Periodontitis Periodontitis Agreement between professors and HYNETS of 0.69. CNN calibrated with bone loss on PAs applied to OPGs for staging and grading of whole-mouth periodontal status.
35 Khaleel [71] Iraq 2021 Photographs 120 BAT algorithm, PCA, SOM Dentist n.a. BAT method provided 95% accuracy against ground truth Assessment of different algorithms’ efficacy in recognising gingival disease.
36 Kouznetsova [72] USA 2021 Salivary metabolites N/A DNN n.a. n.a. Model performance assessment only of different CNNs. CNN predicts which molecules should be assessed for metabolic diagnosis of periodontitis or oral cancers.
37 Lee [35] South Korea 2021 Panoramic 530 U-Net, Dense U-Net, ResNet, SegNet Radiologists Radiologists The accuracy of the resulting model was 79.54%. Assessment of a variety of CNN architectures for detecting and quantifying the missing teeth, bone loss, and staging on panoramic radiographs.
38 Li [73] China 2021 Photographs 3932 Fnet, Lnet, cnet Dentist Dentists Low agreement between three dentists and CNN in heatmap analysis. CNN used for gingivitis detection photographs.
39 Li [29] China 2021 Photographs 110 DeepLabv3+ Dentist n.a. MobileNetV2 performed in a similar manner to Xception65; however, Mob, was 20× quicker. Different CNNs trialled for RGB assessment of gingival tissues to assess inflamed gum detection on photographs.
40 Ma [74] Taiwan 2021 Panoramic 432 ConvNet, U-Net Unknown n.a. ConvNet analysis post U-Net segmentation—the model showed moderate levels of agreement (F2 score of between 0.523 and 0.903) and the ability to predict periodontitis and ASCVD. CNNs used to assess for atherosclerotic cardiovascular disease and periodontitis on OPGs.
41 Moran [75] Brazil 2021 Periapicals 5 Inception and for super-resolution SRCNN Dentist n.a. Minimal enhancement of CNN performance was noted from super resolution, which may introduce additional artefacts. The study compared the effects of super resolution methods on the ability of CNNs to perform segmentation and bone loss identification.
42 Ning [76] Germany 2021 Saliva N/A DisGeNet, HisgAtlas n.a. n.a. DL-based model able to predict immunosuppression genes in periodontitis with an accuracy of 92.78%. CNN to identify immune subtypes of periodontitis and pivotal immunosuppression genes that discriminated periodontitis from the healthy.
43 Shang [32] China 2021 Photographs 7220 U-Net Dentist Dentist U-Net to have a 10% increased recognition of calculus, wear facets, gingivitis, and decay Comparison of U-Net vs. comparison between U-Net and DeepLabV3/PSPNet architecture for image recognition on oral pictures for wear, decay, calculus, and gingivitis.
44 Wang [77] USA 2021 Metabolites N/A FARDEEP n.a. n.a. ML successfully used in logistic regression of plaque samples. CNN is used as a processing tool for clinical, immune, and microbial profiling of peri-implantitis patients against health.
45 Jiang [37] China 2022 Panoramic 640 U-Net, YOLO-v4 Periodontist Periodontist Compared to the ground truth, accuracy of 0.77 was achieved by the proposed architecture. CNN used to provide % bone loss and resorption/furcation lesion and staging of periodontal disease from OPGs.
46 Lee [36] USA 2022 Periapicals 693 U-Net, ResNet Dentist Dentist The accuracy of the diagnosis based upon staging and grading was 0.85 Full mouth PA films were used to review bone loss—staging and grading were then performed.
47 Li [73] China 2022 Photographs 2884 OCNet, Anet Dentist Dentist CNN provided AUC prediction of 87.11% for gingivitis and 80.11% for calculus. Research trialling different methods of segmentation to assess plaque on photographs of tooth surfaces (inc ‘dye labelling’).
48 Liu [78] China 2022 Periapicals 1670 Faster R-CNN Dentist Dentist The results confirm the advantage of utilising multiple CNN architectures for joint optimisation to increase UTC ROC boosts of up to 8%. CNN used to assess implant marginal bone loss with dichotomous outcomes.
49 Pan [33] USA 2022 Ultrasound 627 U-Net Dentist Dentist Showed a significant difference between CNN outcome and dental experts’ labelling. CNN was used to provide an estimation of gingival height in porcine models.
50 Zadrozny [34] Poland 2022 Panoramic 30 U-Net Radiologists Dentists Tested CNN showed unacceptable reliability for assessment of caries (ICC = 0.681) and periapical lesions (ICC = 0.619), but acceptable for fillings (ICC = 0.920), endodontically treated teeth (ICC = 0.948), and periodontal bone loss (ICC = 0.764). Testing of commercially available product Diagnocat in the evaluation of panoramic radiographs.

4. Discussion

CNNs are becoming increasingly clinically relevant in their ability to assess imaging data and can be an excellent utility for analysing large clinically relevant datasets. In the present review, we systematically compiled the application of CNNs in the field of periodontology, evaluating the application and outcomes of these studies. The majority of these studies were completed in the USA and China, which is line with the majority of DL papers in the medical spheres [79].

The use of CNNs in periodontal and dental research has continuously grown over the last decade. Since the first publications using CNNs in the early 2010s, there has been an exponential increase in the number of publications using this tool. In 2021 alone, there were 2911 registered studies on PubMed with CNN in the title, up from 42 in 2010. This, of course, makes empirical sense. The power, utility and applicability of this tool are endless and are improving as the architectures evolve, providing both generic and task-focused utility.

Previous systematic and scoping reviews in dentistry have highlighted the underuse of this tool and a lag for dental research in this area [80,81]. However, with the total number of periodontal imaging papers alone now equalling the number of dental imaging papers in 2018, this lag is likely to have been overcome. This is unequivocally to the betterment of dental patient care when considering the benefits patients have enjoyed through similar endeavours in medicine [82].

The advantage of a broad search is the volume of literature that is assessed. However, it must be noted that a significant portion of the literature was derived from technical standards, conference proceedings, and related materials on computer-science-oriented repositories rather than journal articles. This is advantageous for the authors because the time to publication can be significantly reduced by removing the requirement for peer or public review. This may suit the rapidly evolving world of computer science, where breakthroughs can occur at breakneck speed, but it is unclear if the intrinsic validity of these publications is reduced due to a lack of public/peer scrutiny. This literature is published by technical scientists and therefore reported differently to how clinicians might expect it.

The majority of studies focused on the processing of images and the recognition of structures. Radiography accounted for two-thirds of these data (Figure 5), reasonably evenly split between periapical and panoramic radiographs. With both imaging modalities indicated in the assessment of patients with periodontitis, virtual assistance in diagnosis will be relevant to the clinician. In these studies, the majority of studies focused on image segmentation rather than pathology detection. We can only assume that this was due to the relative complexity of a detection tool compared to segmentation. However, moving forwards, relative detection from consecutive radiographs or more pathology identification would be of use as an assistive tool.

Database collation is an issue for all data science. This is of significant difficulty when compiling data from medical records. When considering the field of machine learning, suitable numbers of records are required to train and refine an AI tool. Supervised training is recognised as a central tenant to improve a CNN’s performance. This process requires that the CNN is shown labelled images to define structures for the CNN to segment. It is, however, possible to over-train CNNs, resulting in errors due to over-recognition.

With homogenous data, it would not be unreasonable to believe that the distribution of training input utilised would be Gaussian or normally distributed. However, we find that the mean number of plain film radiographs utilised for training was multinomial in distribution. This is best illustrated by the divergent mean (1940) and median (585). Figure 6 shows a histogram of the data, where peaks can be seen between 100 and 1000 images and over 4000 images used. The authors query whether this was related to convenient sampling implicit with smaller numbers of data or whether this is reflective of the inherent variabilities of the requirements of the differing architectures of ML used. A power calculation was performed in a single paper, but only for a single outcome, and as such may not have offered the researcher team an accurate assessment of data volume required [37]. Further to this, there was no notable descriptive correlation between outcome-reported accuracy and the number of images used, with larger studies reporting a variety of accuracies or other outcomes [30,54,59,62,78]. This may be solely reflective of the noted variety of reporting outcomes and methodologies rather than due to a lack of correlation.

Labelling these images for training and subsequent reference tests is also of paramount importance. Gold standards were applied in several studies whereby more than one clinician or a specialist radiologist was used to perform this manual task to ensure a consensus approach to best fit was taken. However, in many studies, this was performed by a single evaluator, reducing both the external validity of the results due to single operator bias and the internal validity, introducing potential systematic error. Whilst some efforts have been made to standardise methodologies [83,84], these are still yet to be adopted or referenced in wider practice. The majority of studies used pixel-by-pixel annotation tools, with single study moving to ‘Dye Staining or ‘Grab Cut’ methodology [73]. This practice changes the digital annotation sequencing, essentially highlighting areas of interest to the CNN rather than circumscribing the anatomical feature of relevance. This is already mainstream in several other fields [85], possibly highlighting the digital technical lag present in dental research and the opportunities available in this field.

The performance of CNNs was reported in very heterogeneous manners, almost all of which come with drawbacks. Area under the curve (AUC) analyses are important, but only partially informative when it comes to outcomes. Lying above the curve is just a minimum requirement and, typically, only very elevated values are representative of applications suitable to handle real-world data; thus, sensitivity and specificity need to be reported as well to indicate performance. This was the case in some later papers, but those additional values were missing from most papers where AUC was a reported outcome. Accuracy was the most commonly reported outcome known to distort results when class imbalances are in place [86]. This is due to the class distribution being unknown in the training data, meaning that there is an assumption as to which population is more present (i.e., bone loss or no bone loss). This assumption skews data, making the reported ≈70–90% accuracy less meaningful due to inherent guesswork in ascertaining the class.

It has been suggested that the gold standard employed in these papers should be for models to be tested against independent expert assessors on truly unseen data, or indeed for the models to be used in a clinical trial [83]. However, no studies included compare the outcomes from a truly independent examiner team on unseen data against the proposed AI/ML outcomes. This leads to a ‘fuzzy gold standard’, whereby the AI/ML outcome is being marked against the clinician examiners that were used to train the tool. Until literature providing baseline information on the efficacy of examiners is universally accepted or studies showing performance against truly independent dentist performance in a clinical environment is shown, the assistive nature of these tools in a ‘real world setting’ will remain unproven.

With the marked reporting heterogeneity and uncertainty of gold standard testing, it is not surprising that gauging meaningful comparison and pooling data for meta-analysis has not been done. This may be indicated in a further systematic review in the future as more standardisation of this form of research occurs. However, descriptively, when considering outcomes, this body of research could show little improvement of accuracies over the last nine years, with accuracies of 90–98% reported in 2014 [45] and accuracies of 0.85 described in 2022. However, we feel that whilst the reported figures have remained similar, the question has markedly changed over this time. Ever increasingly complex problems are being asked of ML. Whilst historical papers asked more nominal questions, more recent literature such as Jiang et al. [37] looked to radiographically stage periodontitis from OPG radiographs. This represents a continuous problem that is grouped into ordinal datasets. Here, the accuracy established was similar to the periodontists who had completed the original manual labelling in the region of 0.85, depending upon tooth position and severity of bone loss. This shows a significant improvement in the quality of the outcome and indicates that as the power of ML increases, the assistive nature of these tools may become more powerful.

This scoping review comes with its inherent limitations. The chosen question is as broad; as is inherent within the purpose of a scoping review [27], reflected by a broad search strategy. Whilst including articles from resources such as ArXiv offers the readers a larger pool of references, it should be noted that these articles are not peer-reviewed and therefore may lack some of the methodological rigour of published literature. In addition to this, the broad inclusion criterion has led to the authors somewhat controversially opting to include papers using a broad variety of machine learning and artificial intelligence modalities employed to analyse a broad range of data types. The resulting heterogeneity of the literature reduced the opportunity for meaningful outcome data comparison. However, the authors would add that they agree with the findings of referenced homogenisation efforts to reduce the variability in results expressed in this field.

Methodologically, the main frailty revolves around searches and synthesis being performed by a single reviewer, with cursory checks by second and third authors. Whilst the single reviewer completed the synthesis twice with an extended time interval between reviews, this still forms an area of inherent bias. However, this was necessary with the practicality of this study.

5. Conclusions

Overall, this review gives insight into the application of machine learning in the field of Periodontology. Given artificial intelligence’s relative infancy in healthcare, it is not surprising that significant heterogeneity was found in the methodology and reporting outcomes. All efforts should be made to bring further research in line with increasingly recognised gold standard for research and reporting. International agreement on a gold standard against which to measure these tools would also significantly assist readers in assessing the utility of this modality of tool. As such, at this juncture, no accurate conclusions can be drawn as to the efficacy and usefulness of this tool in the field of periodontology.

Acknowledgments

Thanks are given to the research support services at The University of Sheffield Library for their assistance in completing searches.

Author Contributions

Conceptualisation, J.S.; methodology, J.S., A.M.B., D.A.; validation, J.S., A.M.B., D.A. formal analysis J.S.; data curation, J.S.; writing—original draft preparation, J.S.; writing—review and editing, J.S., A.M.B., O.J., D.A. All authors have read and agreed to the published version of the manuscript.

Institutional Review Board Statement

Ethical review and approval were waived for this study due to the retrospective analytical nature of this study.

Informed Consent Statement

Not applicable.

Data Availability Statement

Further data required may be requested from the corresponding author.

Conflicts of Interest

The authors declare no conflict of interest.

Funding Statement

This research received no external funding.

Footnotes

Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

References

  • 1.Anderson P.C. The Post-Physician Era: Medicine in the 21st Century. JAMA. 1977;237:2336. doi: 10.1001/jama.1977.03270480076033. [DOI] [Google Scholar]
  • 2.Park W.J., Park J.-B. History and application of artificial neural networks in dentistry. Eur. J. Dent. 2018;12:594–601. doi: 10.4103/ejd.ejd_325_18. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3.Tonetti M.S., Greenwell H., Kornman K.S. Staging and grading of periodontitis: Framework and proposal of a new classification and case definition. J. Clin. Periodontol. 2018;45((Suppl. S20)):S149–S161. doi: 10.1111/jcpe.12945. [DOI] [PubMed] [Google Scholar]
  • 4.Papapanou P.N., Susin C. Periodontitis epidemiology: Is periodontitis under-recognized, over-diagnosed, or both? Periodontology 2000. 2017;75:45–51. doi: 10.1111/prd.12200. [DOI] [PubMed] [Google Scholar]
  • 5.Pirih F.Q., Monajemzadeh S., Singh N., Sinacola R.S., Shin J.M., Chen T., Fenno J.C., Kamarajan P., Rickard A.H., Travan S., et al. Association between metabolic syndrome and periodontitis: The role of lipids, inflammatory cytokines, altered host response, and the microbiome. Periodontology 2000. 2021;87:50–75. doi: 10.1111/prd.12379. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6.Casanova L., Hughes F., Preshaw P.M. Diabetes and periodontal disease: A two-way relationship. Br. Dent. J. 2014;217:433–437. doi: 10.1038/sj.bdj.2014.907. [DOI] [PubMed] [Google Scholar]
  • 7.Kebschull M., Demmer R.T., Papapanou P.N. “Gum bug, leave my heart alone!”--epidemiologic and mechanistic evidence linking periodontal infections and atherosclerosis. J. Dent. Res. 2010;89:879–902. doi: 10.1177/0022034510375281. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.Leroy R., Eaton K.A., Savage A. Methodological issues in epidemiological studies of periodontitis--how can it be improved? BMC Oral. Health. 2010;10:8. doi: 10.1186/1472-6831-10-8. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9.Chapple I.L.C., Wilson N.H.F. Manifesto for a paradigm shift: Periodontal health for a better life. Br. Dent. J. 2014;216:159–162. doi: 10.1038/sj.bdj.2014.97. [DOI] [PubMed] [Google Scholar]
  • 10.Moutinho R.P., Coelho L., Silva A., Pereira J., Pinto M., Baptista I. Validation of a dental image-analyzer tool to measure the radiographic defect angle of the intrabony defect in periodontitis patients. J. Periodontal Res. 2012;47:695–700. doi: 10.1111/j.1600-0765.2012.01483.x. [DOI] [PubMed] [Google Scholar]
  • 11.Topol E.J. High-performance medicine: The convergence of human and artificial intelligence. Nat. Med. 2019;25:44–56. doi: 10.1038/s41591-018-0300-7. [DOI] [PubMed] [Google Scholar]
  • 12.Goodfellow I., Bengio Y., Courville A. Deep Learning. MIT Press; Cambridge, MA, USA: 2016. [Google Scholar]
  • 13.LeCun Y., Boser B., Denker J., Henderson D., Howard R., Hubbard W., Jackel L. Handwritten digit recognition with a back-propagation network. Adv. Neural Inf. Process. Systems. 1989;2:303–318. [Google Scholar]
  • 14.Du G., Cao X., Liang J., Chen X., Zhan Y. Medical Image Segmentation based on U-Net: A Review. J. Imaging Sci. Technol. 2020;64:20508-1. doi: 10.2352/j.imagingsci.technol.2020.64.2.020508. [DOI] [Google Scholar]
  • 15.Astley J.R., Wild J.M., Tahir B.A. Deep learning in structural and functional lung image analysis. Br. J. Radiol. 2022;95:20201107. doi: 10.1259/bjr.20201107. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16.Haenssle H.A., Fink C., Schneiderbauer R., Toberer F., Buhl T., Blum A., Kalloo A., Hassen A.B.H., Thomas L., Enk A., et al. Man against machine: Diagnostic performance of a deep learning convolutional neural network for dermoscopic melanoma recognition in comparison to 58 dermatologists. Ann. Oncol. Off. J. Eur. Soc. Med. Oncol. 2018;29:1836–1842. doi: 10.1093/annonc/mdy166. [DOI] [PubMed] [Google Scholar]
  • 17.Schmidt-Erfurth U., Sadeghipour A., Gerendas B.S., Waldstein S.M., Bogunović H. Artificial intelligence in retina. Prog. Retin. Eye Res. 2018;67:1–29. doi: 10.1016/j.preteyeres.2018.07.004. [DOI] [PubMed] [Google Scholar]
  • 18.Marcus G. Deep learning: A critical appraisal. arXiv. 2018180100631 [Google Scholar]
  • 19.Pendlebury M., Horner K., Eaton K. Selection Criteria for Dental Radiography. Faculty of General Dental Practitioners; London, UK: Royal College of Surgeons; London, UK: 2004. [Google Scholar]
  • 20.Dietrich T., Ower P., Tank M., West N.X., Walter C., Needleman I., Hughes F.J., Wadia R., Milward M.R., Hodge P.J., et al. Periodontal diagnosis in the context of the 2017 classification system of periodontal diseases and conditions–implementation in clinical practice. Br. Dent. J. 2019;226:16–22. doi: 10.1038/sj.bdj.2019.3. [DOI] [PubMed] [Google Scholar]
  • 21.Endodontology ESo Quality guidelines for endodontic treatment: Consensus report of the European Society of Endodontology. Int. Endod. J. 2006;39:921–930. doi: 10.1111/j.1365-2591.2006.01180.x. [DOI] [PubMed] [Google Scholar]
  • 22.Langlais R.P., Skoczylas L.J., Prihoda T.J., Langland O.E., Schiff T. Interpretation of bitewing radiographs: Application of the kappa statistic to determine rater agreements. Oral Surg. Oral Med. Oral Pathol. 1987;64:751–756. doi: 10.1016/0030-4220(87)90181-2. [DOI] [PubMed] [Google Scholar]
  • 23.Gröndahl K., Sundén S., Gröndahl H.-G. Inter- and intraobserver variability in radiographic bone level assessment at Brånemark fixtures. Clin. Oral Implant. Res. 1998;9:243–250. doi: 10.1034/j.1600-0501.1998.090405.x. [DOI] [PubMed] [Google Scholar]
  • 24.Schwendicke F., Samek W., Krois J. Artificial Intelligence in Dentistry: Chances and Challenges. J. Dent. Res. 2020;99:769–774. doi: 10.1177/0022034520915714. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25.Inker L.A., Eneanya N.D., Coresh J., Tighiouart H., Wang D., Sang Y., Crews D.C., Doria A., Estrella M.M., Froissart M., et al. New Creatinine- and Cystatin C–Based Equations to Estimate GFR without Race. N. Engl. J. Med. 2021;385:1737–1749. doi: 10.1056/NEJMoa2102953. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 26.Arksey H., O’Malley L. Scoping studies: Towards a methodological framework. Int. J. Soc. Res. Methodol. 2005;8:19–32. doi: 10.1080/1364557032000119616. [DOI] [Google Scholar]
  • 27.Munn Z., Peters M.D.J., Stern C., Tufanaru C., McArthur A., Aromataris E. Systematic Review or Scoping Review? Guidance for Authors When Choosing between a Systematic or Scoping Review Approach. BMC Med. Res. Methodol. 2018;18:143. doi: 10.1186/s12874-018-0611-x. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 28.Hupe M. EndNote X9. Journal of Electronic Resources in Medical Libraries. 2019;16:117–119. doi: 10.1080/15424065.2019.1691963. [DOI] [Google Scholar]
  • 29.Li G.-H., Hsung T.-C., Ling W.-K., Lam W.Y.-H., Pelekos G., McGrath C. Automatic Site-Specific Multiple Level Gum Disease Detection Based on Deep Neural Network. 15th ISMICT. 2021;15:201–205. doi: 10.1109/ismict51748.2021.9434936. [DOI] [Google Scholar]
  • 30.Kabir T., Lee C.-T., Nelson J., Sheng S., Meng H.-W., Chen L., Walji M.F., Jiang X., Shams S. An End-to-end Entangled Segmentation and Classification Convolutional Neural Network for Periodontitis Stage Grading from Periapical Radiographic Images. arXiv. 20212109.13120 [Google Scholar]
  • 31.Duong D.Q., Nguyen K.-C.T., Kaipatur N.R., Lou E.H.M., Noga M., Major P.W., Punithakumar K., Le L.H. Fully automated segmentation of alveolar bone using deep convolutional neural networks from intraoral ultrasound images; Proceedings of the Annual International Conference of the IEEE Engineering in Medicine and Biology Society IEEE Engineering in Medicine and Biology Society Annual International Conference; Berlin, Germany. 23–27 July 2019; pp. 6632–6635. [DOI] [PubMed] [Google Scholar]
  • 32.Shang W., Li Z., Li Y., editors. Identification of common oral disease lesions based on U-Net; Proceedings of the 2021 IEEE 3rd International Conference on Frontiers Technology of Information and Computer (ICFTIC); Greenville, SC, USA. 12–14 November 2021. [Google Scholar]
  • 33.Pan Y.-C., Chan H.-L., Kong X., Hadjiiski L.M., Kripfgans O.D. Multi-class deep learning segmentation and automated measurements in periodontal sonograms of a porcine model. Dentomaxillofacial Radiol. 2022;51:214–218. doi: 10.1259/dmfr.20210363. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 34.Zadrożny L., Regulski P., Brus-Sawczuk K., Czajkowska M., Parkanyi L., Ganz S., Mijiritsky E. Artificial Intelligence Application in Assessment of Panoramic Radiographs. Diagnostics. 2022;12:224. doi: 10.3390/diagnostics12010224. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 35.Lee S. A deep learning-based computer-aided diagnosis method for radiographic bone loss and periodontitis stage: A multi-device study; Proceedings of the 2021 52nd Korean Electric Society Summer Conference; July 2021. [Google Scholar]
  • 36.Lee C., Kabir T., Nelson J., Sheng S., Meng H., Van Dyke T.E., Walji M.F., Jiang X., Shams S. Use of the deep learning approach to measure alveolar bone level. J. Clin. Periodontol. 2021;49:260–269. doi: 10.1111/jcpe.13574. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 37.Jiang L., Chen D., Cao Z., Wu F., Zhu H., Zhu F. A two-stage deep learning architecture for radiographic staging of periodontal bone loss. BMC Oral Heal. 2022;22:1–9. doi: 10.1186/s12903-022-02119-z. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 38.Farhadian M., Shokouhi P., Torkzaban P. A decision support system based on support vector machine for diagnosis of periodontal disease. BMC Res. Notes. 2020;13:1–6. doi: 10.1186/s13104-020-05180-5. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 39.Hegde H., Shimpi N., Panny A., Glurich I., Christie P., Acharya A. Development of non-invasive diabetes risk prediction models as decision support tools designed for application in the dental clinical environment. Inform. Med. Unlocked. 2019;17:100254. doi: 10.1016/j.imu.2019.100254. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 40.Huang W., Wu J., Mao Y., Zhu S., Huang G.F., Petritis B., Huang R. Developing a periodontal disease antibody array for the prediction of severe periodontal disease using machine learning classifiers. J. Periodontol. 2019;91:232–243. doi: 10.1002/jper.19-0173. [DOI] [PubMed] [Google Scholar]
  • 41.Feres M., Louzoun Y., Haber S., Faveri M., Figueiredo L.C., Levin L. Support vector machine-based differentiation between aggressive and chronic periodontitis using microbial profiles. Int. Dent. J. 2018;68:39–46. doi: 10.1111/idj.12326. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 42.Kim E.-H., Kim S., Kim H.-J., Jeong H.-O., Lee J., Jang J., Joo J.-Y., Shin Y., Kang J., Park A.K., et al. Prediction of Chronic Periodontitis Severity Using Machine Learning Models Based On Salivary Bacterial Copy Number. Front. Cell. Infect. Microbiol. 2020;10:571515. doi: 10.3389/fcimb.2020.571515. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 43.Shimpi N., McRoy S., Zhao H., Wu M., Acharya A. Development of a periodontitis risk assessment model for primary care providers in an interdisciplinary setting. Technol. Heal. Care. 2020;28:143–154. doi: 10.3233/THC-191642. [DOI] [PubMed] [Google Scholar]
  • 44.Li S., Liu J., Zhou Z., Zhou Z., Wu X., Li Y., Wang S., Liao W., Ying S., Zhao Z. Artificial intelligence for caries and periapical periodontitis detection. J. Dent. 2022;122:104107. doi: 10.1016/j.jdent.2022.104107. [DOI] [PubMed] [Google Scholar]
  • 45.Papantonopoulos G., Takahashi K., Bountis T., Loos B.G. Artificial Neural Networks for the Diagnosis of Aggressive Periodontitis Trained by Immunologic Parameters. PLoS ONE. 2014;9:e89757. doi: 10.1371/journal.pone.0089757. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 46.Kong Z., Xiong F., Zhang C., Fu Z., Zhang M., Weng J., Fan M. Automated Maxillofacial Segmentation in Panoramic Dental X-Ray Images Using an Efficient Encoder-Decoder Network. IEEE Access. 2020;8:207822–207833. doi: 10.1109/ACCESS.2020.3037677. [DOI] [Google Scholar]
  • 47.Bezruk V., Krivenko S., Kryvenko L., editors. Salivary lipid peroxidation and periodontal status detection in ukrainian atopic children with convolutional neural networks; Proceedings of the 2017 4th International Scientific-Practical Conference Problems of Infocommunications Science and Technology (PIC S&T); Kharkov, Ukraine. 10–13 October 2017. [Google Scholar]
  • 48.Rana A., Yauney G., Wong L.C., Gupta O., Muftu A., Shah P., editors. Automated segmentation of gingival diseases from oral images; Proceedings of the 2017 IEEE Healthcare Innovations and Point of Care Technologies, HI-POCT 2017; Bethesda, MD, USA. 6–8 November 2017. [Google Scholar]
  • 49.Lee J.-H., Kim D.-H., Jeong S.-N., Choi S.-H. Diagnosis and prediction of periodontally compromised teeth using a deep learning-based convolutional neural network algorithm. J. Periodontal Implant. Sci. 2018;48:114–123. doi: 10.5051/jpis.2018.48.2.114. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 50.Yoon S., Odlum M., Lee Y., Choi T., Kronish I.M., Davidson K.W., Finkelstein J. Applying Deep Learning to Understand Predictors of Tooth Mobility Among Urban Latinos. Stud. Health Technol. Inform. 2018;251:241–244. [PMC free article] [PubMed] [Google Scholar]
  • 51.Aberin S.T.A., De Goma J.C. Detecting periodontal disease using convolutional neural networks; Proceedings of the 2018 IEEE 10th International Conference on Humanoid, Nanotechnology, Information Technology, Communication and Control, Environment and Management (HNICEM); Baguio City, Philippines. 29 November–2 December 2018; pp. 1–6. [DOI] [Google Scholar]
  • 52.Askarian B., Tabei F., Tipton G.A., Chong J.W., editors. Smartphone-based method for detecting periodontal disease; Proceedings of the 2019 IEEE Healthcare Innovations and Point of Care Technologies, HI-POCT 2019; Bethesda, Maryland. 20–22 November 2019; [(accessed on 25 November 2022)]. Available online: https://ieeexplore.ieee.org/document/8962638. [Google Scholar]
  • 53.Joo J., Jeong S., Jin H., Lee U., Yoon J.Y., Kim S.C., editors. Periodontal disease detection using convolutional neural networks; Proceedings of the 2019 International Conference on Artificial Intelligence in Information and Communication (ICAIIC); Okinawa, Japan. 11–13 February 2019. [Google Scholar]
  • 54.Kim J., Lee H.-S., Song I.-S., Jung K.-H. DeNTNet: Deep Neural Transfer Network for the detection of periodontal bone loss using panoramic dental radiographs. Sci. Rep. 2019;9:17615. doi: 10.1038/s41598-019-53758-2. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 55.Krois J., Ekert T., Meinhold L., Golla T., Kharbot B., Wittemeier A., Dörfer C., Schwendicke F. Deep Learning for the Radiographic Detection of Periodontal Bone Loss. Sci. Rep. 2019;9:8495. doi: 10.1038/s41598-019-44839-3. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 56.Moriyama Y., Lee C., Date S., Kashiwagi Y., Narukawa Y., Nozaki K., Murakami S. Evaluation of dental image augmentation for the severity assessment of periodontal disease; Proceedings of the 6th Annual Conference on Computational Science and Computational Intelligence, CSCI 2019; Las Vegas, NV, USA. 5–7 December 2019. [Google Scholar]
  • 57.Yauney G., Rana A., Wong L.C., Javia P., Muftu A., Shah P. Automated process incorporating machine learning segmentation and correlation of oral diseases with systemic health; Proceedings of the Annual International Conference of the IEEE Engineering in Medicine and Biology Society IEEE Engineering in Medicine and Biology Society Annual International Conference; Berlin, Germany. 23–27 July 2019; pp. 3387–3393. [DOI] [PubMed] [Google Scholar]
  • 58.Alalharith D.M., Alharthi H.M., Alghamdi W.M., Alsenbel Y.M., Aslam N., Khan I.U., Shahin S.Y., Dianišková S., Alhareky M.S., Barouch K.K. A Deep Learning-Based Approach for the Detection of Early Signs of Gingivitis in Orthodontic Patients Using Faster Region-Based Convolutional Neural Networks. Int. J. Environ. Res. Public Heal. 2020;17:8447. doi: 10.3390/ijerph17228447. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 59.Bayrakdar S.K., Ҫelik Ö., Bayrakdar I.S., Orhan K., Bilgir E., Odabaş A., Aslan A.F. Success Of Artificial Intelligence System In Determining Alveolar Bone Loss From Dental Panoramic Radiography Images. Cumhur. Dent. J. 2020;23:318–324. [Google Scholar]
  • 60.Chang H.-J., Lee S.-J., Yong T.-H., Shin N.-Y., Jang B.-G., Kim J.-E., Huh K.-H., Lee S.-S., Heo M.-S., Choi S.-C., et al. Deep Learning Hybrid Method to Automatically Diagnose Periodontal Bone Loss and Stage Periodontitis. Sci. Rep. 2020;10:7531. doi: 10.1038/s41598-020-64509-z. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 61.Chen Y., Chen X. Medical Imaging and Computer-Aided Diagnosis: Proceeding of 2020 International Conference on Medical Imaging and Computer-Aided Diagnosis (MICAD 2020) Vol. 633. Springer; Singapore: 2020. Gingivitis identification via GLCM and artificial neural network; pp. 95–106. [DOI] [Google Scholar]
  • 62.Lee J.-H.D., Jeong S.-N. Efficacy of deep convolutional neural network algorithm for the identification and classification of dental implant systems, using panoramic and periapical radiographs. Medicine. 2020;99:e20787. doi: 10.1097/md.0000000000020787. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 63.Li H., Zhou J., Zhou Y., Chen J., Gao F., Xu Y., Gao X. Automatic and interpretable model for periodontitis diagnosis in panoramic radiographs; Proceedings of the Medical Image Computing and Computer Assisted Intervention—MICCAI 2020: 23rd International Conference; Lima, Peru. 4–8 October 2020; [(accessed on 25 November 2022)]. pp. 454–463. Available online: https://dl.acm.org/doi/abs/10.1007/978-3-030-59713-9_44. [Google Scholar]
  • 64.Moran M.B.H., Faria M., Giraldi G., Bastos L., Inacio B.D.S., Conci A. On using convolutional neural networks to classify periodontal bone destruction in periapical radiographs; Proceedings of the 2020 IEEE International Conference on Bioinformatics and Biomedicine (BIBM); Seoul, Republic of Korea. 16–19 December 2020; pp. 2036–2039. [DOI] [Google Scholar]
  • 65.Romm E., Li J., Kouznetsova V.L., Tsigelny I.F. Machine Learning Strategies to Distinguish Oral Cancer from Periodontitis Using Salivary Metabolites. Adv. Intell. Syst. Comput. 2020;1252:511–526. doi: 10.1007/978-3-030-55190-2_38. [DOI] [Google Scholar]
  • 66.Thanathornwong B., Suebnukarn S. Automatic detection of periodontal compromised teeth in digital panoramic radiographs using faster regional convolutional neural networks. Imaging Sci. Dent. 2020;50:169–174. doi: 10.5624/isd.2020.50.2.169. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 67.You W., Hao A., Li S., Wang Y., Xia B. Deep learning-based dental plaque detection on primary teeth: A comparison with clinical assessments. BMC Oral Heal. 2020;20:1–7. doi: 10.1186/s12903-020-01114-6. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 68.Cetiner D., Isler S., Bakirarar B., Uraz A. Identification of a Predictive Decision Model Using Different Data Mining Algorithms for Diagnosing Peri-implant Health and Disease: A Cross-Sectional Study. Int. J. Oral Maxillofac. Implant. 2021;36:952–965. doi: 10.11607/jomi.8965. [DOI] [PubMed] [Google Scholar]
  • 69.Chen H., Li H., Zhao Y., Zhao J., Wang Y. Dental disease detection on periapical radiographs based on deep convolutional neural networks. Int. J. Comput. Assist. Radiol. Surg. 2021;16:649–661. doi: 10.1007/s11548-021-02319-y. [DOI] [PubMed] [Google Scholar]
  • 70.Danks R.P., Bano S., Orishko A., Tan H.J., Sancho F.M., D’Aiuto F., Stoyanov D. Automating Periodontal bone loss measurement via dental landmark localisation. Int. J. Comput. Assist. Radiol. Surg. 2021;16:1189–1199. doi: 10.1007/s11548-021-02431-z. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 71.Khaleel B.I., Aziz M.S. Using Artificial Intelligence Methods For Diagnosis Of Gingivitis Diseases. J. Phys. Conf. Ser. 2021;1897:012027. doi: 10.1088/1742-6596/1897/1/012027. [DOI] [Google Scholar]
  • 72.Kouznetsova V.L., Li J., Romm E., Tsigelny I.F. Finding distinctions between oral cancer and periodontitis using saliva metabolites and machine learning. Oral Dis. 2020;27:484–493. doi: 10.1111/odi.13591. [DOI] [PubMed] [Google Scholar]
  • 73.Li W., Liang Y., Zhang X., Liu C., He L., Miao L., Sun W. A deep learning approach to automatic gingivitis screening based on classification and localization in RGB photos. Sci. Rep. 2021;11:1–8. doi: 10.1038/s41598-021-96091-3. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 74.Ma K.S.-K., Liou Y.-J., Huang P.-H., Lin P.-S., Chen Y.-W., Chang R.-F. Identifying medically-compromised patients with periodontitis-associated cardiovascular diseases using convolutional neural network-facilitated multilabel classification of panoramic radiographs; Proceedings of the 2021 International Conference on Applied Artificial Intelligence (ICAPAI); Halden, Norway. 19–21 May 2021. [Google Scholar]
  • 75.Moran M., Faria M., Giraldi G., Bastos L., Conci A. Do Radiographic Assessments of Periodontal Bone Loss Improve with Deep Learning Methods for Enhanced Image Resolution? Sensors. 2021;21:2013. doi: 10.3390/s21062013. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 76.Ning W., Acharya A., Sun Z., Ogbuehi A.C., Li C., Hua S., Ou Q., Zeng M., Liu X., Deng Y., et al. Deep Learning Reveals Key Immunosuppression Genes and Distinct Immunotypes in Periodontitis. Front. Genet. 2021;12 doi: 10.3389/fgene.2021.648329. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 77.Wang C.-W., Hao Y., Di Gianfilippo R., Sugai J., Li J., Gong W., Kornman K.S., Wang H.-L., Kamada N., Xie Y., et al. Machine learning-assisted immune profiling stratifies peri-implantitis patients with unique microbial colonization and clinical outcomes. Theranostics. 2021;11:6703–6716. doi: 10.7150/thno.57775. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 78.Liu M., Wang S., Chen H., Liu Y. A pilot study of a deep learning approach to detect marginal bone loss around implants. BMC Oral Heal. 2022;22:1–8. doi: 10.1186/s12903-021-02035-8. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 79.Celi L.A., Cellini J., Charpignon M.-L., Dee E.C., Dernoncourt F., Eber R., Mitchell W.G., Moukheiber L., Schirmer J., Situ J., et al. Sources of bias in artificial intelligence that perpetuate healthcare disparities—A global review. PLOS Digit. Heal. 2022;1:e0000022. doi: 10.1371/journal.pdig.0000022. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 80.Schwendicke F., Golla T., Dreher M., Krois J. Convolutional neural networks for dental image diagnostics: A scoping review. J. Dent. 2019;91:103226. doi: 10.1016/j.jdent.2019.103226. [DOI] [PubMed] [Google Scholar]
  • 81.Hung K., Montalvao C., Tanaka R., Kawai T., Bornstein M.M. The use and performance of artificial intelligence applications in dental and maxillofacial radiology: A systematic review. Dentomaxillofacial Radiol. 2020;49:20190107. doi: 10.1259/dmfr.20190107. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 82.Rajpurkar P., Chen E., Banerjee O., Topol E.J. AI in health and medicine. Nat. Med. 2022;28:31–38. doi: 10.1038/s41591-021-01614-0. [DOI] [PubMed] [Google Scholar]
  • 83.Schwendicke F., Singh T., Lee J.-H., Gaudin R., Chaurasia A., Wiegand T., Uribe S., Krois J. Artificial intelligence in dental research: Checklist for authors, reviewers, readers. J. Dent. 2021;107:103610. doi: 10.1016/j.jdent.2021.103610. [DOI] [Google Scholar]
  • 84.Mongan J., Moy L., Kahn C.E., Jr. Checklist for artificial intelligence in medical imaging (CLAIM): A guide for authors and reviewers. Radiol. Artif. Intell. 2020;2 doi: 10.1148/ryai.2020200029. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 85.Rother C., Kolmogorov V., Blake A. “GrabCut” interactive foreground extraction using iterated graph cuts. ACM Trans. Graph. (TOG) 2004;23:309–314. [Google Scholar]
  • 86.Kuhn M., Johnson K. Applied Predictive Modelling. Springer; Berlin/Heidelberg, Germany: 2013. [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Data Availability Statement

Further data required may be requested from the corresponding author.


Articles from Dentistry Journal are provided here courtesy of Multidisciplinary Digital Publishing Institute (MDPI)

RESOURCES