Skip to main content
PLOS One logoLink to PLOS One
. 2025 Aug 13;20(8):e0329986. doi: 10.1371/journal.pone.0329986

Examining the diagnostic accuracy of artificial intelligence for detecting dental caries across a range of imaging modalities: An umbrella review with meta-analysis

Sarah Arzani 1,*, Ali Karimi 2, Pedram Iranmanesh 3, Maryam Yazdi 1,4,5, Mohammad A Sabeti 6, Mohammad Hossein Nekoofar 7, Jafar Kolahi 8,*, Heejung Bang 9, Paul MH Dummer 10
Editor: Giang Truong Vu11
PMCID: PMC12349118  PMID: 40802701

Abstract

The objective of this systematic review was to systematically collect and analyze multiple published systematic reviews to address the following research question “Are artificial intelligence (AI) algorithms effective for the detection of dental caries?”. A systematic search of five electronic databases, including the Cochrane Library, Embase, PubMed, Scopus, and Web of Science, was conducted until October 15, 2024, with a language restriction to English. All fourteen systematic reviews which assessed the performance of AI algorithms for the detection of dental caries were included. From 137 primary original research studies within the systematic reviews, only 20 reported the data necessary for inclusion in the meta-analysis. Pooled sensitivity was 0.85 (95% Confidence Interval (CI): 0.83 to 0.93), specificity was 0.90 (95% CI: 0.85 to 0.95), and log diagnostic odds ratio was 4.37 (95% CI: 3.16 to 6.27). Area under the summary ROC curve was 0.86. Positive post-test probability was 79% and negative post-test probability was 6%. In conclusion, this meta-analysis has revealed that caries diagnosis using AI is accurate and its use in clinical practice is justified. Future studies should focus on specific subpopulations, depth of caries, and real-world performance validation to further improve the accuracy of AI in caries diagnosis.

Introduction

On a global level, it is estimated that dental caries affects the permanent dentition of approximately 2.3 billion adults and the primary dentition of approximately 530 million children [1]. The annual treatment costs of dental caries for individuals aged 12–65 years worldwide is estimated at US$357 billion (331 billion), or 4.9% of global healthcare expenditure [2].

Early and precise detection of dental caries can lead to effective prevention and treatment with less invasive methods, potentially resulting in improved outcomes and reduced healthcare costs [3]. The conventional strategy for diagnosis of carious lesions is visual examination, supplemented with intraoral radiographs, preferably bitewings [4]. Obviously, the level of clinical expertise of dentists may affect the reliability and accuracy of visual examination methods. The meta-analysis of Walsh T, et al., revealed a pooled sensitivity of visual caries diagnosis of 0.83 and specificity of 0.81 [5]. The meta-analysis of Schwendicke F, et al., assessed diagnostic accuracy of intraoral radiography, i.e., bitewing or periapical radiographs. They reported the pooled sensitivity for radiographic detection of any type of occlusal carious lesion in clinical studies was 0.35, and pooled specificity was 0.78. For radiographic detection of any type of proximal caries in clinical studies, the pooled sensitivity was 0.24, and pooled specificity was 0.97 [6].

Development of reliable, automated, user-friendly, and low-cost tools for diagnosis of dental caries can play an important role in the management of the disease, and improve oral healthcare access and quality globally. Artificial intelligence (AI) algorithms, particularly convolutional neural networks (CNN), are revolutionizing dental care [7]. AI algorithms can detect dental caries from various imaging modalities, such as intraoral photographic images, periapical radiographs, bitewing radiographs, panoramic radiographs, cone beam computed tomography (CBCT) and near-infrared-light transillumination. As an example, the “Videa Dental Assist” is an AI-based caries detection system approved by the U.S. Food and Drug Administration (FDA) that can analyze bitewing, periapical, and panoramic radiographs acquired from patients aged 3 years or older [8].

AI algorithms can automate the diagnostic process, reducing reliance on human expertise, and be used on smart phone apps [9] or cloud platforms, enhancing access to dental care particularly in underserved regions. Despite the promising results, challenges remain in diagnostic accuracy and generalizability of AI platforms across diverse populations and imaging modalities.

Several systematic reviews have attempted to answer concerns that surround the accuracy of AI algorithms as a diagnostic tool for caries detection. The systematic reviews rarely undertook meta-analyses and pooled sensitivity and specificity values were therefore unavailable for evidence-based clinical decision making. The reviews looked at a range of factors when using AI for detection of dental caries, which has led to an overlap of the primary studies in several reviews. Therefore, the aim of the present study is to systematically collect and assess multiple published systematic reviews to answer the question “Are AI algorithms effective for the detection of dental caries?” by including a meta-analysis and reporting pooled sensitivity, specificity and diagnostic odds ratio.

Materials and methods

Protocol registration

The protocol of the review was registered in PROSPERO (#CRD42024568618). The Preferred Reporting Items for Overviews of Reviews (PRIOR), the Preferred Reporting Items for Systematic reviews and Meta-Analyses of Diagnostic Test Accuracy Studies (PRISMA-DTA) and PRISMA-AI criteria were followed to provide accurate and transparent reporting of the review methodology and results [1012].

Eligibility criteria

The selection of studies on the accuracy of diagnostics are based on the PIRD criteria, which include the population, index test, reference test, and diagnosis of interest [13]. The utilization of the PIRD format is suggested as a framework for defining the inclusion criteria used in systematic reviews focusing on the accuracy of diagnostic tests [13]. The PIRD elements of the research were defined as: Population: Systematic reviews and meta-analyses evaluating the diagnostic accuracy of AI algorithms for dental caries detection, Index Test: Application of AI algorithms in dental caries detection, Reference Test: Visual-tactile or clinical based assessment of dental caries, Diagnosis of Interest: The diagnostic accuracy measures of the AI model for dental caries including sensitivity, specificity, log diagnostic odds ratio, and area under receiver operating characteristic (ROC) curve (AUC).

All systematic reviews involving human subjects and relevant dental images, reporting performance metrics or results of AI algorithms in dental caries detection were included. Research that covered non-AI methods for detecting dental caries were excluded. Additionally, guidelines, comments, editorials, duplicate publications, studies that were not systematic reviews, and abstracts without full-text availability were also excluded from the analysis.

Search strategy

Five electronic databases, including Cochrane Library, Embase, PubMed, Scopus, and Web of Science, limited to the English language, were searched systematically until October 15, 2024. To capture grey literature, WorldCat and the first 100 hits of Google Scholar were also explored. The search strategy comprised terms presented in Table 1.

Table 1. Search strategy and number of records retrieved from each database on Oct 15, 2024.

Database Search line Number of retrieval records
PubMed (“Artificial Intelligence”[MeSH Terms] OR “Support Vector Machine”[MeSH Terms] OR “Deep Learning”[MeSH Terms] OR “neural networks, computer”[MeSH Terms] OR (“Machine Learning”[MeSH Terms] OR “Unsupervised Machine Learning”[MeSH Terms] OR “Supervised Machine Learning”[MeSH Terms]) OR “diagnosis, computer assisted”[MeSH Terms] OR “Electronic Data Processing”[MeSH Terms] OR “Convolutional Neural Network”[Text Word] OR “CNN”[Text Word]) AND (“Dental Caries”[MeSH Terms] OR “Root Caries”[MeSH Terms]) 253
EmBase (‘artificial intelligence’/exp OR ‘support vector machine’/exp OR ‘deep learning’/exp OR ‘artificial neural network’/exp OR ‘ann (artificial neural network)’ OR ‘ann analysis’ OR ‘ann approach’ OR ‘ann method’ OR ‘ann methodology’ OR ‘ann methods’ OR ‘ann model’ OR ‘ann modeling’ OR ‘ann modelling’ OR ‘ann models’ OR ‘ann output’ OR ‘ann technique’ OR ‘ann techniques’ OR ‘ann training’ OR ‘anns (artificial neural networks)’ OR ‘algorithmic neural network’ OR ‘artificial nn’ OR ‘artificial nns’ OR ‘artificial neural network’ OR ‘artificial neural networks’ OR ‘computational neural network’ OR ‘computer neural network’ OR ‘computer neural networks’ OR ‘computerized neural network’ OR ‘connectionist model’ OR ‘connectionist network’ OR ‘connectionist neural network’ OR ‘connectionist system’ OR ‘mathematical neural network’ OR ‘neural network (artificial)’ OR ‘neural network (computer)’ OR ‘neural network algorithm’ OR ‘neural network model’ OR ‘neural networks (computer)’ OR ‘neural networks, computer’ OR ‘machine learning’/exp OR ‘convolutional neural network’/exp OR ‘cnn (convolutional neural network)’ OR ‘cnns (convolutional neural networks)’ OR ‘convnet’ OR ‘convoluted neural network’ OR ‘convolution neural network’ OR ‘convolutional anns’ OR ‘convolutional nn’ OR ‘convolutional artificial neural network’ OR ‘convolutional deep neural network’ OR ‘convolutional neural network’ OR ‘convolutionary neural network’ OR ‘deep convolutional neural network’) AND (‘dental caries’/exp OR ‘dental caries’ OR ‘white spot lesion’/exp OR ‘white spot lesion’) 352
Scopus #1 TITLE-ABS-KEY (“Artificial intelligence” OR ai OR “support vector machine” OR “Deep learning” OR “Machine based algorithm*” OR “Neural network*” OR nn OR ann OR cnn OR “Machine learning”)
#2 TITLE-ABS-KEY (dental AND (caries OR cavit* OR decay* OR “White Spot*” OR lesion*))
#3 TITLE-ABS-KEY (caries AND (detection OR diagnosis))
#4 TITLE-ABS-KEY (t**th AND (lesion* OR caries))
#5 TITLE-ABS-KEY (carious AND (lesion* OR dentin*))
#6 TITLE-ABS-KEY (bitewing OR periapical OR pa OR panoramic OR opg OR “Panoramic Radiograph*” OR orthopantomography OR “Cone Beam Computed Tomography” OR cbct OR “Cone Beam CT” OR “Dental CT” OR “Intraoral Photographic Image*” OR “Photographic Image*” OR “Near Infrared Light Transillumination” OR nilt OR “Dental PRE/2 Tomography”)
#7: #2 OR #3 OR #4 OR #5
#8: #1 AND #6 AND #7
426
Web of Science #1 TS=(“Artificial intelligence” OR AI OR “Support Vector Machine” OR “Deep learning” OR “Machine based algorithm*” OR “Neural network*” OR NN OR ANN OR CNN OR “Machine learning”)
#2 TS=(Dental AND (Caries OR Cavit* OR Decay* OR “White Spot*” OR Lesion*))
#3 TS=(Caries AND (Detection OR Diagnosis))
#4 TS=(T**th AND (Lesion* OR Caries))
#5 TS=(Carious AND (Lesion* OR Dentin*))
#6 TS=(bitewing OR periapical OR pa OR panoramic OR opg OR “Panoramic Radiograph*” OR orthopantomography OR “Cone Beam Computed Tomography” OR cbct OR “Cone Beam CT” OR “Dental CT” OR “Intraoral Photographic Image*” OR “Photographic Image*” OR “Near Infrared Light Transillumination” OR nilt OR “Dental PRE/2 Tomography”)
#7: #2 OR #3 OR #4 OR #5
#8: #1 AND #6 AND #7
277
Cochrane Reviews (“artificial intelligence”) AND (“dental caries”)” (Word variations have been searched) 1
WorldCat ti:(“Artificial intelligence” OR “support vector machine” OR “Deep learning” OR “Neural network*” OR “Machine learning”) AND (“Dental Caries” OR “Dental Cavit*” OR “Dental Decay*” OR “Dental White Spot*” OR “Carious Lesion*” OR “Caries Detection”) 119
Google Scholar (“Artificial intelligence” OR “support vector machine” OR “Deep learning” OR “Neural network*” OR “Machine learning”) AND (“Dental Caries” OR “Dental Cavit*” OR “Dental Decay*” OR “Dental White Spot*” OR “Carious Lesion*” OR “Caries Detection”) limited to review articles First 100 hits

Data extraction and synthesis

Duplicate records were removed using Mendeley Reference Manager and two independent reviewers (SA and JK) performed the preliminary screening of titles and abstracts in accordance with the eligibility criteria. Full-text relevant records were retrieved and screened. During the screening stage, text mining was also conducted using the SWIFT-Review software (Sciome LLC, NC, USA). This program automatically groups abstracts related to comparable subjects using machine learning methods [14]. We employed AI algorithms to search, categorize, and prioritize large number of primary studies during the screening stage using SWIFT-Review software [15]. However, the final decisions on inclusion were based on human judgment. Discrepancies among the independent reviewers were resolved using the Delphi methodology during each phase [16].

The characteristics of included systematic reviews was extracted by two independent authors (SA and AK) using a standard Joanna Briggs Institute extraction form [17], which included the following details:

  • Study characteristics: the first author’s name, year of publication, number of databases searched, number and types of included studies, date of search, quality assessment tool, and results of the meta-analysis.

  • AI techniques employed: artificial neural network, machine learning algorithms, deep learning method, or convolutional neural network.

  • The types of caries: proximal, root, occlusal, or other parts of tooth.

  • Dental imaging modalities used: bitewing, periapical, panoramic radiographs, cone beam computed tomography, and intraoral photographic images.

  • Measures to be pooled: sensitivity, specificity, log diagnostic odds ratio, area under curve.

The retrieved data were analyzed and qualitatively summarized to assess the diagnostic accuracy of artificial intelligence for detecting dental caries across a range of imaging modalities. Overlapping studies within the included reviews were handled by creation of a citation matrix and calculation of corrected covered area (CCA) using the ccaR package of R software (R Foundation for Statistical Computing, Vienna, Austria) [18].

Meta-analysis

Original studies provided data on True Positive (TP), False Positive (FP), False Negative (FN), and True Negative (TN) counts included in the meta-analysis. We also used a backward calculation of TP, TN, FP, and FN using: 1) Prevalence (TP + FN/ Total sample size), sensitivity, specificity, and total sample size, 2) Prevalence, positive likelihood ratio, negative likelihood ratio, and total sample size, or 3) Sensitivity, specificity, total number of positive and total number of negative test results [19].

A random effect model was used to combine sensitivity, specificity, log diagnostic odds ratio, AUC estimates of diagnostic tests among the original studies included in the systematic reviews. The Wald test was used to calculate the confidence intervals (CIs). The Galbraith plot was employed to assess heterogeneity and detect potential outliers. To quantify the potential effect of these outliers on the estimation of the pooled variables, leave-one-out meta-analyses were carried out. The regression-based Egger test and nonparametric rank correlation Begg test were employed for assessment of small-study effects. Nonparametric trim-and-fill analysis was employed to evaluate the number of studies potentially missing from the meta-analysis. To assess post-test probabilities, the Fagan nomogram was used [20]. The meta-analysis and visualizations were carried out using Stata 18 (StataCorp, College Station, TX, USA) and the mada, MetaDTA, and nsROC packages of R software. We employed an online diagnostic test calculator hosted by the University of Illinois at Chicago to draw Fagan nomogram. Interaction between imaging method and AI algorithm for sensitivity, specificity, and log diagnostic odds ratio assessed by random forests model (a machine learning algorithm) using a metaforest package.

Quality and risk of bias assessment

The quality and risk of bias of the included systematic reviews were assessed by two independent researchers (SA and MY) using the AMSTAR 2 (A MeaSurement Tool to Assess systematic Reviews) tool [21]. The risk of bias in the original studies included in the systematic reviews eligible for meta-analysis was evaluated using the QUADAS-2 tool [22]. Disagreements regarding the quality evaluation were resolved by a Delphi technique [16].

Results

Characteristics of the included reviews

The initial search yielded a total of 1120 studies (Table 1). Following the removal of duplicates and the title-abstract screening process, 25 records met the criteria for full-text review. Ultimately, after excluding 11 records (Table 2), the remaining 14 records which met the eligible criteria were included (Fig 1). Twelve of the included records were systematic reviews and two were systematic reviews with meta-analyses (Table 3). The reviews covered a range of publication years, with 2 conducted prior to 2022, 6 conducted in 2022, 1 conducted in 2023, and 5 conducted in 2024. The original studies included in the reviews spanned from 1984 to 2023, providing a broad range of evidence.

Table 2. Excluded systematic reviews and the reason for exclusion.

Study Year of publication Title Reason for exclusion
Chifor et al. [23] 2022 Automated diagnosis using artificial intelligence a step forward for preventive dentistry: A systematic review. Conditions not concerning dental caries diagnosis
Hegde et al. [24] 2022 Deep learning algorithms show some potential as an adjunctive tool in caries diagnosis. Editorial
Alqutaibi et al. [25] 2023 Artificial intelligence (AI) as an aid in restorative dentistry is promising, but still a work in progress. Editorial
Fatima et al. [26] 2023 A systematic review on artificial intelligence applications in restorative dentistry. Duplicate publication of Revilla-León M, et al. [27]
Futyma-Gabka et al. [28] 2021 The use of artificial intelligence in radiological diagnosis and detection of dental caries: a systematic review Critically low quality (based on AMSTAR 2)
Musri et al. [29] 2021 Deep learning convolutional neural network algorithms for the early detection and
diagnosis of dental caries on periapical radiographs: A systematic review
Critically low quality (based on AMSTAR 2)
Singh et al. [30] 2022 Progress in deep learning-based dental and maxillofacial image analysis: A systematic review Critically low quality (based on AMSTAR 2)
Forouzeshfar et al. [31] 2023 Dental caries diagnosis using neural networks and deep learning: a systematic review Critically low quality (based on AMSTAR 2)
Bhat et al. [32] 2023 A comprehensive survey of deep learning algorithms and applications in dental radiograph analysis Critically low quality (based on AMSTAR 2)
Al-Namankany et al. [33] 2023 Influence of artificial intelligence-driven diagnostic tools on treatment decision making in early childhood caries: A systematic review of accuracy and clinical outcomes Conditions not concerning dental caries diagnosis
Alam et al. [34] 2024 Applications of artificial intelligence in the utilisation of imaging modalities in dentistry: A systematic review and meta-analysis of in-vitro studies Conditions not concerning dental caries diagnosis
Pecorari et al. [35] 2024 The use of artificial intelligence in the diagnosis of carious lesions: Systematic review and meta-analysis Not peer-reviewed and a preprint article

Fig 1. Systematic review flowchart according to PRISMA 2020.

Fig 1

Table 3. Data extracted from the included systematic reviews.

Study Aim No. of databases, included studies Types of included studies Date range of included studies Quality assessment tool Meta-analysis AI techniques The types of caries Dental imaging modalities used Findings
Talpur S et al. [36], 2022 Diagnosis of DC with ML 4, 12 Clinical studies and conference papers 2008−22 Tool developed by the authors No DL with algorithms CNN and ANN, Hidden layer propagation, SVM, Random Forest, Logistic Regression, Feedforward propagation and Feedback propagation CNN- and ANN based research Proximal, root, and occlusal Different types of imaging modalities Acc: 69.6–99.0%
Reyes LT et al. [37], 2022 Diagnosis and prognostic prediction of DC with ML 5, 15 (10 articles related to the aim of present review) Cross-sectional and longitudinal studies 2008−20 QUADAS-2 No ANNs, J48, Random Tree, Random Forest, SVM, and Naïve Bayes Proximal, occlusal, and smooth free surface Conventional or digital radiographic, oral photographs, and NILT AUC: 74.5–98.7%
Moharrami M et al. [38], 2023 Diagnosis of DC with AI 3, 19 Clinical studies and conference papers 2019−22 Modified version of QUADAS-2 No CNNs, Mask R-CNN, SVM, Decision Tree, SSD-MobileNetV2, Faster R-CNN, Customized CNNs, Custom UCDA, YOLO-based CNN, ResNet architectures, Shufflenet V2, DenseNet, ResNext, RetinaNet, Squeeze Net, RDFNet, VGG NET-16, VGG NET-19, InceptionV3, SSD-ConvNet, DentalNet Occlusal and smooth free surface Oral photographs F1-scores: classification task: 68.3–94.3%, and detection task: 42.8–95.4%
Prados-Privado M et al. [39], 2020 Diagnosis of DC with DL 3,13 In vitro and clinical studies and conference papers 2008−20 Cochrane RoB No ResNet architectures, ResNext50, ANN, CNN, DNN Mask R-CNN, Feedforward NN, RT and DCT, FCNN, MLP, Back propagation NN Proximal, occlusal, pre-cavitated, and initial caries PA, NILT, bitewing, panoramic, radiovisiography, oral photographs Acc: 68.6–99.0%, Precision: 61.5–98.7%, AUC: 74.0–97.1%
Mohammad-Rahimi H et al. [40], 2022 Diagnosis of DC with DL 4, 42 Clinical and in vitro studies 2015−21 Modified version of QUADAS-2 No Customized CNN structures, Transfer learning models, MLP, Auto-encoders (e.g., U-Net), One-stage object detectors (e.g., YoLo) or Two-stage object detectors (e.g., Faster R-CNN) NA Oral photographs, PA, bitewing, NILT, OCT, panoramic, and CBCT Acc: 68.0–99.0%
Revilla-León M et al. [27], 2022 Diagnostic performance of AI in restorative dentistry 5, 34 (29 articles related to the aim of present review) Clinical and in vitro studies 1984-2020 JBI critical appraisal checklist No Expert systems, Regression analysis, Fuzzy logic learning, Perceptron NN, MLP, Back propagation NN, CNNs, ANNs, Decision tree learning NA PA, bitewing, oral photographs, NILT, and fiber optic displacement sensor Acc: 76.0–88.3%, Se: 73.0–90.0%, Sp: 61.5–93.0%
Khanagar SB et al. [41], 2022 Diagnosis and prediction of DC with AI 7, 34 (24 articles related to the aim of present review) No limitation for the type of study 2018−22 QUADAS-2 No DCNNs, CNNs, ANNs NA PA, NILT, bitewing, oral photographs, panoramic, OCT, and micro-CT images Acc: 73.3–98.8%, Se: 59.0–98.9%, Sp: 68.9–98.2%, AUC: 74.0–98.7%
Khanagar SB et al. [42], 2022 Application of AI in pediatric dentistry 7, 20 (1 articles related to the aim of present review) No limitation for the type of study 2021 QUADAS-2 No CNNs, ANNs NA Oral photographs Se: 88.1–96.0%,
Sp: 97.0–99.0%
Khanagar SB et al. [43], 2021 Application of AI in dentistry 7, 43 (4 articles related to the aim of present review) No limitation for the type of study 2008−20 QUADAS-2 No DCNNs, CNNs, ANNs NA PA, bitewing, and NILT AUC: 74.0–89.0%
Zanini LGK et al. [44], 2024 Application of AI in caries detection, classification, and segmentation in X-ray images 4, 42 No limitation for the type of study 2016−23 Tool developed by the authors No IP, CSL, Inception V3, Hybrid graph-cut segmentation, CNN-SVM, AlexNet, VGG19, VGG16, DenseNet-121, U-Net, ResNet, GoogLeNet, MobileNet V2, Hybrid NN, XNet, SegNet, 3-layer CNN, Network MI-DCNNE, Mask R-CNN, Faster R-CNN, YOLOv3, CariesNet, ResNeXt, Deep gradient–based LeNet, EfficientNet-B0 NA PA, bitewing, x-ray, panoramic, and CBCT NA
Ndiaye AD et al. [45], 2024 Application of AI in caries detection, classification, and segmentation from radiographic databases 4, 35 Radiographic human studies 2018−22 QUADAS-AI No CNNs, SVM, Random Forest, Naive Bayes, K-nearest neighbors, auto-encoder, semi-supervised learning, support tensor machine, network-in-network, Transfer learning NA PA, bitewing, CBCT, and panoramic Acc: 70.8–99.1%, Se: 41.6–99.4%, Sp: 66.3–98.9%
Ammar N et al. [46], 2024 Diagnosis of DC with AI on bitewing radiographs 5, 14 (5 articles included in meta-analysis) No limitation for the type of study 2020−23 QUADAS-2 Yes AlexNet, ResNet, Inception, U-Net, Faster R-CNN, YOLOv3 Proximal Bitewing Odds Ratio: 55.8 (95% CI = 28.8–108.3), Se: 76.0–94.0%, Sp: 75.0–96.0%
Rokhshad R et al. [47], 2024 Application of AI in pediatric dentistry 8, 33 (2 articles related to the aim of present review) Clinical studies 2008−23 QUADAS-2 Yes ANNs NA Oral photographs Acc: 78.0%, Se: 88.1–96.0%,
Sp: 97.0–99.0%
Albano D et al. [48], 2024 Detection of DC with AI 5, 20 (19 articles related to the aim of present review) No limitation for the type of study 2008−22 QUADAS-2 No DCNNs, CNNs, ANNs NA PA, bitewing, orthopantomography Acc: 73.0–98.0%, Se: 44.0–86.0%, Sp: 85.0–98.0%, precision: 50.0–94.0%, AUC: 84.0–98.0%, F1-score: 64.0–92.0%

Abbreviations: DC, Dental Caries; ML, Machine Learning; DL, Deep Learning; ECC, Early Childhood Caries; RoB, Risk of Bias; CNN/ConvNet, Convolutional Neural Network; ANN, Artificial Neural Network; SVM, Support Vector Machine; FCNN, Fully Convolutional Neural Network; FCM, Fuzzy C-Means algorithm; RT, Radon transformation; DCT, Discrete Cosine Transformation; ResNet, Residual Network; NN, Neural Network; R-CNN, Region-based Convolutional Neural Network; SSD, Single Shot Detector; UCDA, Unified Caries Detection and Assessment; DCNN, Deep Convolutional Neural Network; RDFNet, RGB-D Fusion Network; VGG, Visual Geometry Group; LSTM, Long Short-Term Memory; MLP, Multi-Layer Perceptron; MPCA, Modified Principal Component Analysis; IP, Image Processing; CSL, Classical Supervised Learning; MI-DCNNE, Multi-Input Deep Convolutional Neural Network Ensemble; NA, Not Available; PA, Peri-Apical; CBCT, Cone-beam computed tomography; CT, Computed Tomography scans; OCT, Optical Coherence Tomography; NILT, Near-Infrared-Light Transillumination; Acc, Accuracy; Se, Sensitivity; Sp, Specificity; AUC, Area Under Curve.

The included reviews encompassed a range of AI techniques, prominently CNN, employed for dental caries detection. These reviews assessed the performance of AI algorithms in detecting dental caries using diverse dental imaging modalities, including intraoral photographic images, periapical radiographs, bitewing radiographs, CBCT images, near-infrared-light transillumination, panoramic radiographs, and others. Quality assessment among the included systematic reviews exhibited predominantly moderate quality (Table 4).

Table 4. Quality assessment of each included systematic review according to AMSTAR 2.

Systematic Review 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 Overall Quality
Reyes LT et al. [37] Yes Yes Yes Yes Yes Yes Yes Yes Yes No No MA No MA Yes Yes No MA Yes High
Mohammad-Rahimi H et al. [40] Yes Yes Yes Yes Yes Yes Yes Yes Yes No No MA No MA Yes Yes No MA Yes High
Khanagar SB et al. [41] Yes PY Yes Yes Yes Yes Yes Yes Yes No No MA No MA Yes Yes No MA Yes High
Ndiaye AD et al. [45] Yes Yes Yes Yes Yes Yes Yes Yes Yes No No MA No MA Yes Yes No MA Yes High
Moharrami M et al. [38] Yes Yes Yes Yes Yes No Yes Yes Yes No No MA No MA Yes Yes No MA Yes Moderate
Prados-Privado M et al. [39] Yes PY Yes Yes Yes Yes Yes Yes Yes No No MA No MA Yes Yes No MA Yes Moderate
Revilla-León M et al. [27] Yes PY Yes Yes Yes Yes Yes Yes Yes No No MA No MA Yes No No MA Yes Moderate
Khanagar SB et al. [42] Yes No Yes PY Yes Yes Yes Yes Yes No No MA No MA Yes Yes No MA Yes Moderate
Khanagar SB et al. [43] Yes PY Yes Yes Yes Yes Yes Yes Yes No No MA No MA Yes Yes No MA Yes Moderate
Zanini LGK et al. [44] Yes Yes No PA Yes Yes Yes Yes Yes No No MA No MA Yes Yes No MA Yes Moderate
Ammar N et al. [46] Yes PY Yes Yes Yes Yes Yes Yes PY No Yes No Yes Yes Yes Yes Moderate
Rokhshad R et al. [47] Yes PY Yes Yes Yes Yes Yes Yes Yes No Yes No Yes Yes Yes Yes Moderate
Talpur S et al. [36] Yes PY Yes PY No No Yes Yes PY No No MA No MA Yes Yes No MA Yes Low
Albano D et al. [48] Yes No Yes Yes Yes Yes Yes Yes Yes No No MA No MA Yes No No MA Yes Low

Abbreviations: PY, Partial Yes; MA, Meta-Analysis; Low, Low quality. Numbers at first row showed the items in AMSTAR 2. [21]

The citation matrix of primary studies included in the systematic reviews is presented in S1 Table. The CCA_Proportion was 0.07 and CCA_Percentage was 6.90, showing moderate overlap. The pairwise CCA presented in S1 Fig shows which combinations of paired reviews had the highest overlap.

Meta-analysis of eligible original studies

Among the fourteen systematic reviews included, 137 relevant original studies were identified (S1 Table). Only 20 original articles reported numbers of TP, TN, FP, and FN and were included in the meta-analysis (Table 5). Quality assessment of diagnostic accuracy studies among the included original studies revealed the lowest risk regarding index test applicability concerns and the highest risk regarding index test risk of bias (Fig 2).

Table 5. Data extracted from the included original studies for meta-analysis.

Study Dental imaging modality Caries detection Reference standard AI algorithms (in detail)
Liu L [49] 2015 Intraoral photographic images Overall Dental experts CNNs (Mask R-CNN)
Kühnisch J [50] 2022 Intraoral photographic images Overall A dental expert CNNs (MobileNetV2)
Detection of cavitations CNNs (MobileNetV2)
Park EY [51] 2022 Intraoral photographic images Overall A dental expert CNNs (ResNet-18)
CNNs (ResNet-18 with U-net)
CNNs (Faster R-CNN)
CNNs (Faster R-CNN with U-net)
Vinayahalingam S[52] 2021 Panoramic radiographs Overall Dental experts CNNs (MobileNet V2)
Ezhov M [53] 2021 CBCT N/A Dental and radiology experts CNNs (U-net)
Devlin H [54] 2021 Bitewing radiographs Enamel Dental experts CNNs (AssistDent AI software)
Zheng L [55] 2021 Periapical radiographs Dentin Dental experts CNNs (VGG19)
CNNs (Inception V3)
CNNs (ResNet-18)
De Araujo Faria V[56], 2021 Panoramic radiographs Overall Dental experts ANNs
Oztekin F [57] 2023 Panoramic radiographs Overall A dental expert CNNs (EfficientNet-B0)
CNNs (DenseNet-121)
CNNs (ResNet-50)
Imak A [58] 2022 Periapical radiographs Overall A dental expert CNNs (MI-DCNNE)
Chen X [59], 2022 Bitewing radiographs Overall 2 endodontic experts and 1 radiologist CNNs (Faster R-CNN)
Li S [60] 2022 Periapical radiographs Overall Dental experts CNNs (VGG16)
CNNs (modified ResNet-18 backbone)
Fariza A [61] 2022 Panoramic radiographs Overall Dental experts CNNs (ResNet-18)
CNNs (ResNeXt50 32 × 4d)
Jayasinghe H [62] 2022 Periapical radiographs Overall Dental experts CNNs (ResNet-101)
Ari T [63] 2022 Periapical radiographs Overall An oral and maxillofacial radiologist CNNs (U-net)
Zadro˙zny L [64] 2022 Panoramic radiographs Overall Dental experts CNNs (Diagnocat)
Suttapak W [65] 2022 Bitewing radiographs Overall An oral and maxillofacial radiologist CNNs (ResNet-101)
Liu F [66] 2022 Periapical radiographs Overall Dental experts CNNs (VGG16)
CNNs (Inception V3)
CNNs (ResNet-50)
CNNs (DenseNet-121)
Estai M [67] 2022 Bitewing radiographs Overall Dental experts CNNs (Inception-ResNet-V2)
Bayraktar Y [68] 2022 Bitewing radiographs Overall Dental experts CNNs (YOLOv3)

Fig 2. Quality assessment of diagnostic accuracy studies among the included original studies for meta-analysis according to QUADAS-2.

Fig 2

Of the 29423 diagnostic tests analyzed from the results retrieved from the 20 articles (Fig 3), the pooled sensitivity was 0.85 (95% CI: 0.83 to 0.93), specificity was 0.90 (95% CI: 0.85 to 0.95), and log diagnostic odds ratio was 4.37 (95% CI: 3.16 to 6.27) (diagnostic odd ratio: 70.9 (95% CI: 44.9 to 111.9)) as shown in Fig 4. Results of heterogeneity assessments related to sensitivity, specificity, and log diagnostic odds ratio among the included studies are provided in Fig 5. The studies of Zadrozny L, 2022 [64], Park EY, 2022 (Faster R-CNN) [51], and De Araujo Faria V, 2021 [56], were outliers for sensitivity, specificity, and log diagnostic odds ratio, respectively. Results of the leave-one-out meta-analysis are shown in Fig 6. When omitting each study, results changed minimally, e.g., 0.01 in sensitivity and specificity and the first decimal place in log diagnostic odds ratio. The results of the regression-based Egger test and nonparametric rank correlation Begg test for assessment of small-study effects are provided in Table 6. The nonparametric trim-and-fill analysis of publication bias are in Table 6 and Fig 7. Nonparametric trim-and-fill analysis estimated 6 unpublished studies estimated for sensitivity and log diagnostic odds ratio. The meta-analysis for non-parametric ROC curves are presented in Fig 8. The area under the pooled ROC curve is 0.867. Positive and negative likelihood ratios were 10.443 (95% CI: 7.505 to 14.531) and 0.168 (95% CI: 0.138 to 0.205), respectively. Total number of tests were 29424, TP were 6836, and FN were 1209. Hence the prevalence of dental caries was 27.3%. The positive post-test probability was 79% and negative post-test probability was 6% (Fig 9). The interaction between imaging method and AI algorithm for sensitivity, specificity, and log diagnostic odds ratio showed in Fig 10. Finally, Fig 11 provides a summary of meta-analysis performance, along with visual evaluations of threshold effect, hierarchical summary ROC curve and heterogeneity of data.

Fig 3. Summary of the 29423 diagnostic tests included in the meta-analysis.

Fig 3

Fig 4. Forest plots showing pooled sensitivity, specificity, and log diagnostic odds ratio among the included studies.

Fig 4

Test of sub-group differences for sensitivity, specificity, and log diagnostic odds ratio were significant (P < 0.001).

Fig 5. Galbraith plots showing the results of heterogeneity assessments related to sensitivity, specificity, and log diagnostic odds ratio among the included studies.

Fig 5

Fig 6. Forest plots showing the results of the leave-one-out meta-analysis for sensitivity, specificity, and log diagnostic odds ratio among the included studies.

Fig 6

Table 6. Results of small-study effects related to publication bias and results of nonparametric trim-and-fill analysis, showing the impact of publication bias on the pooled estimates.

Egger’s Regression Egger’s test Kendalls Tau Begg’s test Number of imputed studies Observed pooled effect size Observed + Imputed pooled effect size
Sensitivity −3/201 P = 0.00137 −0/338 P = 0.0058 6 0.850 (95% CI: 0.811 to 0.889) 0.818 (95% CI: 0.778 to 0.858)
Specificity −3/062 P = 0.0022 −0/485 P < 0.001 0 0.898 (95% CI: 0.867 to 0.929) 0.898 (95% CI: 0.867 to 0.929)
Log diagnostic odds ratio 6/408 P < 0.001 0/481 P < 0.001 6 4.368 (95% CI: 3.798 to 4.938) 3.913 (95% CI: 3.219 to 4.606)

Fig 7. Funnel plot showing the results of the nonparametric trim-and-fill analysis of publication bias related to sensitivity, specificity, and log diagnostic odds ratio among the included studies.

Fig 7

Fig 8. Meta-analysis for non-parametric receiver operating characteristic (ROC) curve.

Fig 8

Studies numbered according to the order seen in forest plots in Fig 6.

Fig 9. Fagan nomogram to determine the post-test probabilities of presence of dental caries based on the likelihood ratios.

Fig 9

The post-test probability of a patient having dental caries was 79% with the positive test result and post-test probability of a patient having dental caries was 6% with the negative test result.

Fig 10. Pinned scatter plot of the bivariate partial prediction of interaction between imaging method and AI algorithm for sensitivity, specificity, and log diagnostic odds ratio (DOR).

Fig 10

This plot was created by means of random forests model (a machine learning algorithm) (Number of trees in forest: 500, Minimum terminal node size: 5).

Fig 11. Summary receiver operating characteristic plot showing meta-analysis performance (HSROC: Hierarchical summary receiver operating characteristic).

Fig 11

Discussion

The core question in this umbrella review and meta-analysis was “Are AI algorithms effective for the detection of dental caries?” Given the global burden of caries and the exponential growth of AI in diagnostics, the study is timely and necessary in order to inform clinicians, scientists and other stakeholders on the effectiveness of this emerging technology.

In this umbrella review, 14 systematic reviews were included (Table 3) and 12 systematic reviews excluded (Table 2). Among the included systematic reviews only 2 conducted a meta-analysis (Table 4, item 11), most likely due to the failure of the original studies they included to report adequate details. We analyzed 29423 diagnostic tests, which resulted in a pooled sensitivity of 0.85 (95% CI: 0.83 to 0.93) and specificity of 0.90 (95% CI: 0.85 to 0.95) (Fig 4). Readers must note, the type of dental imaging strongly influences diagnostic performance. AI evaluating panoramic radiographs (with tooth overlap) operates under very different conditions than AI analyzing bitewing images (with clear approximal visibility). Pooling such results may reduce the clinical interpretability of the findings. To address this issue we conducted a subgroup meta-analysis for different imaging modalities with the results presented in Fig 4.

It is well-known that, the most reliable conclusions would come from reviews comparing studies conducted under the same diagnostic protocols and caries definitions. As showed in Table 5, caries detection method, reference standard and AI algorithm were the same among the studies included in the meta-analysis.

Nevertheless, Ammar N et al., [46] (2024) in a recent systematic review and meta-analysis assessed diagnostic performance of AI algorithms for caries detection on bitewing radiographs. Among 5 included studies, the pooled sensitivity and specificity were 0.87 (95% CI: 0.76 to 0.94) and 0.89 (95% CI: 0.75 to 0.96). Macey R, et al., (2021) [69] in a recent systematic review and meta-analysis, which included 67 studies reporting a total of 19590 tooth sites/surfaces, assessed the diagnostic accuracy of several visual classification systems for the detection and diagnosis of non‐cavitated coronal dental caries. For all visual classification systems, the pooled sensitivity and specificity were 0.86 (95% CI 0.80 to 0.90) and 0.77 (95% CI 0.72 to 0.82) respectively. In another systematic review and meta-analysis, Iranzo-Cortés JE, et al., (2019) [70]examined the accuracy of caries diagnostic tools based on laser fluorescence in pre-cavitated carious lesions. For 655 nm light wavelength lasers, which included 25 studies, the pooled sensitivity and specificity were 0.77 (95% CI: 0.70 to 0.83) and 0.75 (95% CI: 0.69 to 0.80), respectively. For 405 nm light wavelength lasers, which included 13 studies, the pooled sensitivity and specificity were 0.81 (95% CI: 0.68 to 0.89) and 0.75 (95% CI: 0.62 to 0.85), respectively. The meta-analysis of Walsh T, et al., (2022) [5] which included 64 studies, reported pooled sensitivity and specificity for fluorescence-based devices 0.76 and 0.83, for analog and digital radiographs 0.50 and 0.89, for electrical conductance or impedance 0.83 and 0.72, and for transillumination and optical coherence tomography 0.76 and 0.82, respectively. Macey R, et al. (2020) [71] in a systematic review and meta-analysis of diagnostic test accuracy of fluorescence-based devices for diagnosis of enamel caries reported estimated sensitivity of 0.70 and specificity of 0.78 among 79 included studies.

Among 29423 diagnostic tests analyzed in the present study, the prevalence estimate of dental caries was 27.3% (Fig 3). Results of the meta-analysis revealed a positive likelihood ratio of 10.443 (95% CI: 7.505 to 14.531), meaning that a positive test result is 8.47 time more likely to occur in someone who has caries, compared to someone who does not have caries, and a negative likelihood ratio of 0.168 (95% CI: 0.138 to 0.205), meaning that a negative test result is 0.17 time more likely to occur in someone who has caries compared to someone who does not have caries. The positive post-test probability was 79%, meaning that if a patient test is positive, there is a 79% chance they actually have caries, and negative post-test probability of 6%, meaning that if a patient test is negative, there is a 6% chance they actually have caries (Fig 9).

Sensitivity and specificity are a well-known pair of indicators for assessment of diagnostic test accuracy. There have been many efforts to combine the results of a diagnostic study into one single measure, for example the diagnostic odd ratio [72]. The results of a diagnostic odd ratio assessment ranges from 0 to infinity (with 1 as null value), where higher values indicate better diagnostic test performance [73]. In the present study the pooled log diagnostic odd ratio was 4.37 (diagnostic odd ratio: 70.9). This value provides a measure of how much likely a positive test result occurs in a person with dental caries compared to person without dental caries. Ammar N et al., (2024) [46] reported a pooled diagnostic odds ratio of 55.8 for AI-based caries detection on bitewing radiographs. In the study of Macey R, et al., (2021) [69] that included all visual caries classification systems, the pooled diagnostic odds ratio was 20.38. In another meta-analysis, Macey R, et al., (2020) [71] reported that the pooled diagnostic odds ratio for fluorescence-based enamel caries diagnostic devices was 14.1.

As a general rule, the AUC serves as an overall summary of diagnostic test performance. In the present study, the area under the pooled ROC curve for AI algorithms for the detection of dental caries was 0.86. This means that there is high probability that a randomly chosen individual with caries will have higher test score than a randomly chosen individual without disease. Iranzo-Cortés JE, et al., (2019) [70] reported that the area under the pooled ROC curve was 0.81 for 655 nm light wavelength lasers and was 0.80 for 405 nm light wavelength lasers.

The Egger’s and Bag’s test results were significant (p < 0.05) for sensitivity, specificity, and log diagnostic odds ratio confirming that publication bias existed within the 20 original studies included in the meta-analysis. The nonparametric trim-and-fill analysis estimated 6 unpublished studies and presence of publication bias for sensitivity, and log diagnostic odds ratio. This publication bias may be related to difficulties related to the publication of innovative interdisciplinary high-tech research outcomes.

High levels of heterogeneity (I2 > 95) were found regarding sensitivity, specificity, and log diagnostic odds ratio among the included studies. These levels of heterogeneity can be explained by diversity in imaging modalities (e.g., bitewing vs. panoramic radiographs), differences in AI algorithm architectures, and training dataset characteristics.

The main limitation of this umbrella review and meta-analysis include the fact that only 20 original studies were included in the meta-analysis quantitative data synthesis. The number of primary original research studies was 137, with only 20 (14.5%) original research articles reporting the necessary details of AI–based caries diagnostic test results including numbers TP, TN, FP, and FN involved in the meta-analysis. To facilitate future meta-analysis, authors are encouraged to report details of AI–based caries diagnostic tests including numbers of TP, TN, FP, and FN, all of which are essential for a meta-analysis.

The prevalence of dental caries among the 20 original research articles included in the meta-analysis was 27.3%, which is noticeably lower than the prevalence of dental caries in the real community suggesting a degree of selection bias among images datasets used for training and validation of AI algorithms. Future studies must involve the training and validation of AI algorithms using image datasets from a broader range demographic locations and healthcare settings to better reflect the real-world prevalence of dental caries.

We tried to conduct sub-group analysis according to the depth of the carious lesions and their location, for instance, enamel caries, dentine caries or root caries. Yet, among the majority of the original studies included, the AI algorithms were trained to detect dental caries and could not distinguish different depths and locations. Future research should focus on training AI algorithms to diagnose caries at different sites and at different depth of progression.

However, in future analyses, the most valuable evidence will likely come from studies that simulate real-world diagnostic conditions, including time pressure, clinician–patient interaction, and the integration of AI as part of the clinical workflow rather than in isolated image review [74].

In summary, this meta-analysis supports the use of AI in clinical practice for the detection of dental caries. However, AI algorithms are being developed and implemented rapidly and their accuracy for detection of dental caries is likely to increase in future. The majority of the included studies in the meta-analysis used CNN for image processing and detection of dental caries. This type of deep learning algorithm is a powerful type of feed-forward artificial neural network [75] that learns features automatically via filter optimization. The accuracy of the CNN model for detection of dental caries can be improved by dealing with large datasets and high-resolution inputs from diverse demographics (tooth type, age, sex, and ethnicity) and imaging equipment. Dental professionals could participate in public volunteer computing programs and potentially upload their radiographic images and donate their unused CPU and GPU cycles to develop powerful CNN models for detection of dental caries with high levels of accuracy [76].

Finally, a recent survey published in Nature involving 1,600 researchers from around the world concluded that scientists are worried and expressed fears over the lack of transparency and the ‘Black Box Effect’ among AI systems, over training data including biased information, AI spreading misinformation, and AI-generated deepfakes [77]. An editorial in the journal Science stated that “Excitement about AI has been tempered by concerns about potential downsides” [78]. The dental research community, journal editors, clinicians, and policymakers should be aware and be attentive of the significant and emerging concerns regarding AI safety [79] and ethical worries regarding the use of AI systems in biomedical research and clinical practice [80,81]. Although AI can assist in dental caries diagnosis, it should not be a substitute for human judgment and dental practitioners must take responsibility for the use of AI in caries diagnosis.

Conclusions

In this umbrella meta-analysis, the analysis of 29423 diagnostic tests resulted in a pooled sensitivity of 0.85, specificity of 0.90, log diagnostic odd ratio of 4.37, AUC of 0.86, positive post-test probability of 79%, and negative post-test probability of 6%, which support the implementation of this diagnostic tool in clinical practice. Future studies should focus on specific subpopulations, depth of caries, and real-world performance validation. Although AI can assist in dental caries diagnosis, it should not be a substitute for human judgment.

Supporting information

S1 Table. The citation matrix of primary studies included in the systematic reviews for the use of AI in the detection of dental caries.

The “1” implies a checkmark, that is the study is included “0” implies that the study is not included in the review in question.

(DOCX)

pone.0329986.s001.docx (562.6KB, docx)
S1 Fig. Visualization of the pairwise CCA (%) with a heatmap.

(TIF)

pone.0329986.s002.tif (112.3KB, tif)
S1 File. PRISMA+DTA+Checklist.

(DOCX)

pone.0329986.s003.docx (30.1KB, docx)

Acknowledgments

The authors would like to extend their gratitude to Prof. Roya Kelishadi for her invaluable guidance and support in the preparation of this manuscript.

Data Availability

All relevant data are within the manuscript and its Supporting Information files.

Funding Statement

The author(s) received no specific funding for this work.

References

  • 1.Wen PYF, Chen MX, Zhong YJ, Dong QQ, Wong HM. Global Burden and Inequality of Dental Caries, 1990 to 2019. J Dent Res. 2022;101(4):392–9. doi: 10.1177/00220345211056247 [DOI] [PubMed] [Google Scholar]
  • 2.EFP reveals stunning global cost of gingivitis, caries, tooth loss. https://www.dental-tribune.com/news/efp-reveals-stunning-global-cost-of-gingivitis-caries-tooth-loss/#:~:text=Using%20new%20modelling%20to%20estimate,4.9%25%20of%20global%20healthcare%20expenditure
  • 3.Neuhaus KW, Ellwood R, Lussi A, Pitts NB. Traditional lesion detection aids. Monogr Oral Sci. 2009;21:42–51. doi: 10.1159/000224211 [DOI] [PubMed] [Google Scholar]
  • 4.Kühnisch J, Aps JK, Splieth C, Lussi A, Jablonski-Momeni A, Mendes FM, et al. ORCA-EFCD consensus report on clinical recommendation for caries diagnosis. Paper I: caries lesion detection and depth assessment. Clin Oral Investig. 2024;28(4):227. doi: 10.1007/s00784-024-05597-3 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5.Walsh T, Macey R, Ricketts D, Carrasco Labra A, Worthington H, Sutton AJ, et al. Enamel Caries Detection and Diagnosis: An Analysis of Systematic Reviews. J Dent Res. 2022;101(3):261–9. doi: 10.1177/00220345211042795 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6.Schwendicke F, Tzschoppe M, Paris S. Radiographic caries detection: A systematic review and meta-analysis. J Dent. 2015;43(8):924–33. doi: 10.1016/j.jdent.2015.02.009 [DOI] [PubMed] [Google Scholar]
  • 7.Alzaid N, Ghulam O, Albani M, Alharbi R, Othman M, Taher H, et al. Revolutionizing Dental Care: A Comprehensive Review of Artificial Intelligence Applications Among Various Dental Specialties. Cureus. 2023;15(10):e47033. doi: 10.7759/cureus.47033 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.Foresman A d a m. U.S. Food & Drug Administration. https://www.accessdata.fda.gov/cdrh_docs/pdf23/K232384.pdf 2023. Accessed 2025 June 27.
  • 9.Dhanak N, Chougule VT, Nalluri K, Kakkad A, Dhimole A, Parihar AS. Artificial intelligence enabled smart phone app for real-time caries detection on bitewing radiographs. Bioinformation. 2024;20(3):243–7. doi: 10.6026/973206300200243 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10.Pollock M, Fernandes RM, Pieper D, Tricco AC, Gates M, Gates A, et al. Preferred Reporting Items for Overviews of Reviews (PRIOR): a protocol for development of a reporting guideline for overviews of reviews of healthcare interventions. Syst Rev. 2019;8(1):335. doi: 10.1186/s13643-019-1252-9 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11.McInnes MDF, Moher D, Thombs BD, McGrath TA, Bossuyt PM, and the PRISMA-DTA Group, et al. Preferred Reporting Items for a Systematic Review and Meta-analysis of Diagnostic Test Accuracy Studies: The PRISMA-DTA Statement. JAMA. 2018;319(4):388–96. doi: 10.1001/jama.2017.19163 [DOI] [PubMed] [Google Scholar]
  • 12.Cacciamani GE, Chu TN, Sanford DI, Abreu A, Duddalwar V, Oberai A, et al. PRISMA AI reporting guidelines for systematic reviews and meta-analyses on AI in healthcare. Nat Med. 2023;29(1):14–5. doi: 10.1038/s41591-022-02139-w [DOI] [PubMed] [Google Scholar]
  • 13.Campbell JM, Klugar M, Ding S, Carmody DP, Hakonsen SJ, Jadotte YT. Diagnostic test accuracy: methods for systematic review and meta-analysis. JBI Evidence Implementation. 2015;13. [DOI] [PubMed] [Google Scholar]
  • 14.Marshall IJ, Wallace BC. Toward systematic review automation: a practical guide to using machine learning tools in research synthesis. Syst Rev. 2019;8(1):163. doi: 10.1186/s13643-019-1074-9 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15.Fatalla AA, Arzani S, Veseli E, Khademi A, Khandan A, Fahmy MD, et al. Revolutionizing Systematic Reviews and Meta-analyses: The Role of Artificial Intelligence in Evidence Synthesis. Dent Hypotheses. 2023;14. Available: https://journals.lww.com/dhyp/fulltext/2023/14040/revolutionizing_systematic_reviews_and.1.aspx
  • 16.Thangaratinam S, Redman CW. The Delphi technique. The Obstetrician & Gynaecologist. 2005;7(2):120–5. doi: 10.1576/toag.7.2.120.27071 [DOI] [Google Scholar]
  • 17.Armataris E, Redman CWE. JBI Manual for Evidence Synthesis. JBI. 2020. doi: 10.46658/jbimes-20-01 [DOI] [Google Scholar]
  • 18.Lunny C, Pieper D, Thabet P, Kanji S. Managing overlap of primary study results across systematic reviews: practical considerations for authors of overviews of reviews. BMC Med Res Methodol. 2021;21(1):140. doi: 10.1186/s12874-021-01269-y [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19.Kolahi J. Backward calculation of TP, TN, FP, and FN. figshare. 10.6084/m9.figshare.27908817.v1.2024. Accessed 2025 June 27. [DOI]
  • 20.Fagan TJ. Letter: Nomogram for Bayes’s theorem. N Engl J Med. 1975;293(5):257. doi: 10.1056/NEJM197507312930513 [DOI] [PubMed] [Google Scholar]
  • 21.Shea BJ, Reeves BC, Wells G, Thuku M, Hamel C, Moran J, et al. AMSTAR 2: a critical appraisal tool for systematic reviews that include randomised or non-randomised studies of healthcare interventions, or both. BMJ. 2017;358:j4008. doi: 10.1136/bmj.j4008 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22.Whiting PF, Rutjes AWS, Westwood ME, Mallett S, Deeks JJ, Reitsma JB, et al. QUADAS-2: a revised tool for the quality assessment of diagnostic accuracy studies. Ann Intern Med. 2011;155(8):529–36. doi: 10.7326/0003-4819-155-8-201110180-00009 [DOI] [PubMed] [Google Scholar]
  • 23.Chifor R, Arsenescu T, Dascalu (Rusu) LM, Badea AF. Automated diagnosis using artificial intelligence a step forward for preventive dentistry: A systematic review. Ro J Stomatol. 2022;68(3):106–15. doi: 10.37897/rjs.2022.3.7 [DOI] [Google Scholar]
  • 24.Hegde S, Gao J. Deep learning algorithms show some potential as an adjunctive tool in caries diagnosis. J Evid Based Dent Pract. 2022;22(4):101772. doi: 10.1016/j.jebdp.2022.101772 [DOI] [PubMed] [Google Scholar]
  • 25.Alqutaibi AY, Aboalrejal AN. Artificial intelligence (ai) as an aid in restorative dentistry is promising, but still a work in progress. J Evid Based Dent Pract. 2023;23(1):101837. doi: 10.1016/j.jebdp.2023.101837 [DOI] [PubMed] [Google Scholar]
  • 26.Fatima ST, Ali Akbar SM, Ahmed A, Asghar SK, Hussain M, Iqbal SS. A Systematic Review on Artificial Intelligence Applications in Restorative Dentistry. PJMHS. 2023;17(2):783–6. doi: 10.53350/pjmhs2023172783 [DOI] [Google Scholar]
  • 27.Revilla-León M, Gómez-Polo M, Vyas S, Barmak AB, Özcan M, Att W, et al. Artificial intelligence applications in restorative dentistry: A systematic review. J Prosthet Dent. 2022;128(5):867–75. doi: 10.1016/j.prosdent.2021.02.010 [DOI] [PubMed] [Google Scholar]
  • 28.Futyma-Gabka K, Rózylo-Kalinowska I. The use of artificial intelligence in radiological diagnosis and detection of dental caries: A systematic review. Journal of Stomatology. 2021;74. 10.5114/jos.2021.111664 [DOI]
  • 29.Musri N, Christie B, Ichwan SJA, Cahyanto A. Deep learning convolutional neural network algorithms for the early detection and diagnosis of dental caries on periapical radiographs: A systematic review. Imaging Sci Dent. 2021;51(3):237–42. doi: 10.5624/isd.20210074 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 30.Singh NK, Raza K. Progress in deep learning-based dental and maxillofacial image analysis: A systematic review. Expert Systems with Applications. 2022;199:116968. doi: 10.1016/j.eswa.2022.116968 [DOI] [Google Scholar]
  • 31.Forouzeshfar P, Safaei AA, Ghaderi F, Hashemi Kamangar S, Kaviani H, Haghi S. Dental caries diagnosis using neural networks and deep learning: a systematic review. Multimed Tools Appl. 2023;83(10):30423–66. doi: 10.1007/s11042-023-16599-w [DOI] [Google Scholar]
  • 32.Bhat S, Birajdar GK, Patil MD. A comprehensive survey of deep learning algorithms and applications in dental radiograph analysis. Healthcare Analytics. 2023;4:100282. doi: 10.1016/j.health.2023.100282 [DOI] [Google Scholar]
  • 33.Al-Namankany A. Influence of artificial intelligence-driven diagnostic tools on treatment decision-making in early childhood caries: a systematic review of accuracy and clinical outcomes. Dent J (Basel). 2023;11(9):214. doi: 10.3390/dj11090214 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 34.Alam MK, Alftaikhah SAA, Issrani R, Ronsivalle V, Lo Giudice A, Cicciù M, et al. Applications of artificial intelligence in the utilisation of imaging modalities in dentistry: A systematic review and meta-analysis of in-vitro studies. Heliyon. 2024;10(3):e24221. doi: 10.1016/j.heliyon.2024.e24221 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 35.Arias Pecorari VG, Almeida Cezário LR, de Barros Arato CV, de Lima Costa T, Cortellazzi KL, Pecorari RF, et al. The use of artificial intelligence in the diagnosis of carious lesions: Systematic review and meta-analysis. Cold Spring Harbor Laboratory. 2024. doi: 10.1101/2024.05.03.24306821 [DOI] [Google Scholar]
  • 36.Talpur S, Azim F, Rashid M, Syed SA, Talpur BA, Khan SJ. Uses of Different Machine Learning Algorithms for Diagnosis of Dental Caries. J Healthc Eng. 2022;2022:5032435. doi: 10.1155/2022/5032435 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 37.Reyes LT, Knorst JK, Ortiz FR, Ardenghi TM. Machine Learning in the Diagnosis and Prognostic Prediction of Dental Caries: A Systematic Review. Caries Res. 2022;56(3):161–70. doi: 10.1159/000524167 [DOI] [PubMed] [Google Scholar]
  • 38.Moharrami M, Farmer J, Singhal S, Watson E, Glogauer M, Johnson AEW, et al. Detecting dental caries on oral photographs using artificial intelligence: A systematic review. Oral Dis. 2024;30(4):1765–83. doi: 10.1111/odi.14659 [DOI] [PubMed] [Google Scholar]
  • 39.Prados-Privado M, García Villalón J, Martínez-Martínez CH, Ivorra C, Prados-Frutos JC. Dental Caries Diagnosis and Detection Using Neural Networks: A Systematic Review. J Clin Med. 2020;9(11):3579. doi: 10.3390/jcm9113579 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 40.Mohammad-Rahimi H, Motamedian SR, Rohban MH, Krois J, Uribe SE, Mahmoudinia E, et al. Deep learning for caries detection: A systematic review. J Dent. 2022;122:104115. doi: 10.1016/j.jdent.2022.104115 [DOI] [PubMed] [Google Scholar]
  • 41.Khanagar SB, Alfouzan K, Awawdeh M, Alkadi L, Albalawi F, Alfadley A. Application and performance of artificial intelligence technology in detection, diagnosis and prediction of dental caries (DC)-a systematic review. Diagnostics (Basel). 2022;12(5):1083. doi: 10.3390/diagnostics12051083 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 42.Khanagar SB, Alfouzan K, Alkadi L, Albalawi F, Iyer K, Awawdeh M. Performance of Artificial Intelligence (AI) Models Designed for Application in Pediatric Dentistry—A Systematic Review. Applied Sciences. 2022;12(19):9819. doi: 10.3390/app12199819 [DOI] [Google Scholar]
  • 43.Khanagar SB, Al-Ehaideb A, Maganur PC, Vishwanathaiah S, Patil S, Baeshen HA, et al. Developments, application, and performance of artificial intelligence in dentistry - A systematic review. J Dent Sci. 2021;16(1):508–22. doi: 10.1016/j.jds.2020.06.019 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 44.Zanini LGK, Rubira-Bullen IRF, Nunes F de LDS. A systematic review on caries detection, classification, and segmentation from x-ray images: methods, datasets, evaluation, and open opportunities. J Imaging Inform Med. 2024;37(4):1824–45. doi: 10.1007/s10278-024-01054-5 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 45.Ndiaye AD, Gasqui MA, Millioz F, Perard M, Leye Benoist F, Grosgogeat B. Exploring the Methodological Approaches of Studies on Radiographic Databases Used in Cariology to Feed Artificial Intelligence: A Systematic Review. Caries Res. 2024;58(3):117–40. doi: 10.1159/000536277 [DOI] [PubMed] [Google Scholar]
  • 46.Ammar N, Kühnisch J. Diagnostic performance of artificial intelligence-aided caries detection on bitewing radiographs: a systematic review and meta-analysis. Jpn Dent Sci Rev. 2024;60:128–36. doi: 10.1016/j.jdsr.2024.02.001 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 47.Rokhshad R, Zhang P, Mohammad-Rahimi H, Shobeiri P, Schwendicke F. Current applications of artificial intelligence for pediatric dentistry: a systematic review and meta-analysis. Pediatr Dent. 2024;46(1):27–35. [PubMed] [Google Scholar]
  • 48.Albano D, Galiano V, Basile M, Di Luca F, Gitto S, Messina C, et al. Artificial intelligence for radiographic imaging detection of caries lesions: a systematic review. BMC Oral Health. 2024;24(1):274. doi: 10.1186/s12903-024-04046-7 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 49.Liu L, Xu J, Huan Y, Zou Z, Yeh S-C, Zheng L-R. A Smart Dental Health-IoT Platform Based on Intelligent Hardware, Deep Learning, and Mobile Terminal. IEEE J Biomed Health Inform. 2020;24(3):898–906. doi: 10.1109/JBHI.2019.2919916 [DOI] [PubMed] [Google Scholar]
  • 50.Kühnisch J, Meyer O, Hesenius M, Hickel R, Gruhn V. Caries Detection on Intraoral Images Using Artificial Intelligence. J Dent Res. 2022;101(2):158–65. doi: 10.1177/00220345211032524 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 51.Park EY, Cho H, Kang S, Jeong S, Kim E-K. Caries detection with tooth surface segmentation on intraoral photographic images using deep learning. BMC Oral Health. 2022;22(1):573. doi: 10.1186/s12903-022-02589-1 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 52.Vinayahalingam S, Kempers S, Limon L, Deibel D, Maal T, Bergé S, et al. The Automatic Detection of Caries in Third Molars on Panoramic Radiographs Using Deep Learning: A Pilot Study. Research Square Platform LLC. 2021. doi: 10.21203/rs.3.rs-379636/v1 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 53.Ezhov M, Gusarev M, Golitsyna M, Yates J, Kushnerev E, Tamimi D, et al. Development and Validation of a Cbct-Based Artificial Intelligence System for Accurate Diagnoses of Dental Diseases. Research Square Platform LLC. 2021. doi: 10.21203/rs.3.rs-303329/v1 [DOI] [Google Scholar]
  • 54.Devlin H, Williams T, Graham J, Ashley M. The ADEPT study: a comparative study of dentists’ ability to detect enamel-only proximal caries in bitewing radiographs with and without the use of AssistDent artificial intelligence software. Br Dent J. 2021;231(8):481–5. doi: 10.1038/s41415-021-3526-6 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 55.Zheng L, Wang H, Mei L, Chen Q, Zhang Y, Zhang H. Artificial intelligence in digital cariology: a new tool for the diagnosis of deep caries and pulpitis using convolutional neural networks. Ann Transl Med. 2021;9(9):763. doi: 10.21037/atm-21-119 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 56.De Araujo Faria V, Azimbagirad M, Viani Arruda G, Fernandes Pavoni J, Cezar Felipe J, Dos Santos EMCMF, et al. Prediction of radiation-related dental caries through pyradiomics features and artificial neural network on panoramic radiography. J Digit Imaging. 2021;34(5):1237–48. doi: 10.1007/s10278-021-00487-6 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 57.Oztekin F, Katar O, Sadak F, Yildirim M, Cakar H, Aydogan M, et al. An explainable deep learning model to prediction dental caries using panoramic radiograph images. Diagnostics (Basel). 2023;13(2):226. doi: 10.3390/diagnostics13020226 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 58.Imak A, Celebi A, Siddique K, Turkoglu M, Sengur A, Salam I. Dental caries detection using score-based multi-input deep convolutional neural network. IEEE Access. 2022;10:18320–9. doi: 10.1109/access.2022.3150358 [DOI] [Google Scholar]
  • 59.Chen X, Guo J, Ye J, Zhang M, Liang Y. Detection of Proximal Caries Lesions on Bitewing Radiographs Using Deep Learning Method. Caries Res. 2022;56(5–6):455–63. doi: 10.1159/000527418 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 60.Li S, Liu J, Zhou Z, Zhou Z, Wu X, Li Y, et al. Artificial intelligence for caries and periapical periodontitis detection. J Dent. 2022;122:104107. doi: 10.1016/j.jdent.2022.104107 [DOI] [PubMed] [Google Scholar]
  • 61.Fariza A, Asmara R, Rojaby MOF, Astuti ER, Putra RH. Evaluation of Convolutional Neural Network for Automatic Caries Detection in Digital Radiograph Panoramic on Small Dataset. 2022 International Conference on Data and Software Engineering (ICoDSE). 2022. 65–70. 10.1109/icodse56892.2022.9972183 [DOI]
  • 62.Jayasinghe H, Pallepitiya N, Chandrasiri A, Heenkenda C, Vidhanaarachchi S, Kugathasan A, et al. Effectiveness of Using Radiology Images and Mask R-CNN for Stomatology. In: 2022 4th International Conference on Advancements in Computing (ICAC), 2022. 60–5. doi: 10.1109/icac57685.2022.10025034 [DOI] [Google Scholar]
  • 63.Ari T, Sağlam H, Öksüzoğlu H, Kazan O, Bayrakdar İŞ, Duman SB, et al. Automatic Feature Segmentation in Dental Periapical Radiographs. Diagnostics (Basel). 2022;12(12):3081. doi: 10.3390/diagnostics12123081 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 64.Zadrożny Ł, Regulski P, Brus-Sawczuk K, Czajkowska M, Parkanyi L, Ganz S, et al. Artificial Intelligence Application in Assessment of Panoramic Radiographs. Diagnostics (Basel). 2022;12(1):224. doi: 10.3390/diagnostics12010224 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 65.Wantanajittikul K, Panyarak W, Jira-apiwattana D, Wantanajittikul K. A unified convolution neural network for dental caries classification. ECTI-CIT Transactions. 2022;16(2):186–95. doi: 10.37936/ecticit.2022162.245901 [DOI] [Google Scholar]
  • 66.Liu F, Gao L, Wan J, Lyu Z-L, Huang Y-Y, Liu C, et al. Recognition of digital dental X-ray images using a convolutional neural network. J Digit Imaging. 2023;36(1):73–9. doi: 10.1007/s10278-022-00694-9 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 67.Estai M, Tennant M, Gebauer D, Brostek A, Vignarajan J, Mehdizadeh M, et al. Evaluation of a deep learning system for automatic detection of proximal surface dental caries on bitewing radiographs. Oral Surg Oral Med Oral Pathol Oral Radiol. 2022;134(2):262–70. doi: 10.1016/j.oooo.2022.03.008 [DOI] [PubMed] [Google Scholar]
  • 68.Bayraktar Y, Ayan E. Diagnosis of interproximal caries lesions with deep convolutional neural network in digital bitewing radiographs. Clin Oral Investig. 2022;26(1):623–32. doi: 10.1007/s00784-021-04040-1 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 69.Macey R, Walsh T, Riley P, Glenny A-M, Worthington HV, O’Malley L, et al. Visual or visual-tactile examination to detect and inform the diagnosis of enamel caries. Cochrane Database Syst Rev. 2021;6(6):CD014546. doi: 10.1002/14651858.CD014546 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 70.Iranzo-Cortés JE, Montiel-Company JM, Almerich-Torres T, Bellot-Arcís C, Almerich-Silla JM. Use of DIAGNOdent and VistaProof in diagnostic of Pre-Cavitated Caries Lesions-A Systematic Review and Meta-Analysis. J Clin Med. 2019;9(1):20. doi: 10.3390/jcm9010020 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 71.Macey R, Walsh T, Riley P, Glenny A-M, Worthington HV, Fee PA, et al. Fluorescence devices for the detection of dental caries. Cochrane Database Syst Rev. 2020;12(12):CD013811. doi: 10.1002/14651858.CD013811 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 72.Glas AS, Lijmer JG, Prins MH, Bonsel GJ, Bossuyt PMM. The diagnostic odds ratio: a single indicator of test performance. J Clin Epidemiol. 2003;56(11):1129–35. doi: 10.1016/s0895-4356(03)00177-x [DOI] [PubMed] [Google Scholar]
  • 73.Schlattmann P. Tutorial: statistical methods for the meta-analysis of diagnostic test accuracy studies. Clin Chem Lab Med. 2023;61(5):777–94. doi: 10.1515/cclm-2022-1256 [DOI] [PubMed] [Google Scholar]
  • 74.Kwiatek J, Leśna M, Piskórz W, Kaczewiak J. Comparison of the Diagnostic Accuracy of an AI-Based System for Dental Caries Detection and Clinical Evaluation Conducted by Dentists. J Clin Med. 2025;14(5):1566. doi: 10.3390/jcm14051566 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 75.Mirmohammadi H, Kolahi J, Khademi A. The transformative power of artificial neural networks in scientific statistical analysis. Dent Hypotheses. 2024;15. [Google Scholar]
  • 76.Kolahi J, Dunning D, Piskun A, Khazaei S, Iranmanesh P, Soltani P. Public volunteer computing in dental science: A lesson learned from the search for extraterrestrial intelligence (SETI@Home) to battle against COVID-19 (Folding@home). 2021. 10.6084/m9.figshare.17124320.v2 [DOI]
  • 77.AI will transform science - now researchers must tame it. Nature. 2023;621(7980):658. doi: 10.1038/d41586-023-02988-6 [DOI] [PubMed] [Google Scholar]
  • 78.Horvitz E. AI, people, and society. Science. 2017;357(6346):7. doi: 10.1126/science.aao2466 [DOI] [PubMed] [Google Scholar]
  • 79.Lazar S, Nelson A. AI safety on whose terms?. Science. 2023;381(6654):138. doi: 10.1126/science.adi8982 [DOI] [PubMed] [Google Scholar]
  • 80.Franco R, Taghizadeh M, Iranmanesh P, Mirmohammadi H, Hasselgren G, Bang H, et al. Whether Enough Attention is Being Paid to the Ethical Concerns Regarding the Use of Artificial Intelligence in Dentistry?. Dental Hypotheses. 2023;14(3):69–70. doi: 10.4103/denthyp.denthyp_96_23 [DOI] [Google Scholar]
  • 81.AI diagnostics need attention. Nature. 2018;555(7696):285. doi: 10.1038/d41586-018-03067-x [DOI] [PubMed] [Google Scholar]

Decision Letter 0

Giang Vu

11 Jun 2025

Dear Dr. Arzani,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

Please submit your revised manuscript by Jul 26 2025 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org . When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols . Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols .

We look forward to receiving your revised manuscript.

Kind regards,

Giang Truong Vu, D.D.S., M.S., Ph.D.

Academic Editor

PLOS ONE

Journal Requirements:

When submitting your revision, we need you to address these additional requirements.

1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at

https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and

https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf

2. Please remove all personal information, ensure that the data shared are in accordance with participant consent, and re-upload a fully anonymized data set.

Note: spreadsheet columns with personal information must be removed and not hidden as all hidden columns will appear in the published file.

Additional guidance on preparing raw data for publication can be found in our Data Policy (https://journals.plos.org/plosone/s/data-availability#loc-human-research-participant-data-and-other-sensitive-data) and in the following article: http://www.bmj.com/content/340/bmj.c181.long.

3. Please include captions for your Supporting Information files at the end of your manuscript, and update any in-text citations to match accordingly. Please see our Supporting Information guidelines for more information: http://journals.plos.org/plosone/s/supporting-information

4. Please review your reference list to ensure that it is complete and correct. If you have cited papers that have been retracted, please include the rationale for doing so in the manuscript text, or remove these references and replace them with relevant current references. Any changes to the reference list should be mentioned in the rebuttal letter that accompanies your revised manuscript. If you need to cite a retracted article, indicate the article’s retracted status in the References list and also include a citation and full reference for the retraction notice.

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Is the manuscript technically sound, and do the data support the conclusions?

Reviewer #1: Yes

Reviewer #2: Yes

**********

2. Has the statistical analysis been performed appropriately and rigorously? -->?>

Reviewer #1: Yes

Reviewer #2: Yes

**********

3. Have the authors made all data underlying the findings in their manuscript fully available??>

The PLOS Data policy

Reviewer #1: Yes

Reviewer #2: Yes

**********

4. Is the manuscript presented in an intelligible fashion and written in standard English??>

Reviewer #1: Yes

Reviewer #2: Yes

**********

Reviewer #1: The manuscript is technically sound and well written. However, some minor corrections would improve the quality.

Heterogeneity has been discussed, however the implications of high variability across studies (e.g., imaging modality, AI model type) could be discussed further. The impact of publication bias on the pooled estimates should be elaborated.

The references have to be formatted according to journal guidelines.

Reviewer #2: Review Comments to the Author:

The topic of this manuscript is timely and relevant. The authors present a broad summary of existing evidence on the use of artificial intelligence (AI) for detecting dental caries, which seems important and useful for the field.

Strengths:

The methodology is clearly described, and the tools used for analysis appear to be appropriate and well chosen for the study’s objective.

The conclusions are well linked to the presented data, and the results are carefully analyzed.

It is also appreciated that the underlying data are fully available.

Suggestions:

It might be helpful to discuss more clearly the limitations related to the fact that only a small number of original studies were included in the meta-analysis.

A short explanation on how overlapping studies across the included reviews were handled could also be useful.

The manuscript would benefit from more critical reflection on how the presence or absence of caries was determined in the original studies. There is little detail on the reference standard, and it remains unclear whether definitions of caries were consistent, validated, or clinically reliable.

It would strengthen the discussion to acknowledge that the most reliable conclusions would come from reviews comparing studies conducted under the same diagnostic protocols and caries definitions.

The type of dental imaging strongly influences diagnostic performance. AI evaluating panoramic radiographs (with tooth overlap) operates under very different conditions than AI analyzing bitewing images (with clear interproximal visibility). Pooling such results may reduce the clinical interpretability of the findings and should be acknowledged as a limitation.

In future analyses, the most valuable evidence will likely come from studies that simulate real-world diagnostic conditions, including time pressure, clinician–patient interaction, and the integration of AI as part of the clinical workflow rather than in isolated image review.

A recently published study demonstrates the importance of incorporating such clinical constraints in AI evaluation and could be considered for citation:

PMID: 40095536

Conclusion:

Overall, this seems to be a well-prepared and solid piece of work. I recommend it for publication after minor clarifications.

**********

what does this mean? ). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy

Reviewer #1: No

Reviewer #2: No

**********

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/ . PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org

PLoS One. 2025 Aug 13;20(8):e0329986. doi: 10.1371/journal.pone.0329986.r003

Author response to Decision Letter 1


4 Jul 2025

The manuscript was revised based on the journal requirements and guidlines.

Comments to the Author

Reviewer #1: The manuscript is technically sound and well written. However, some minor corrections would improve the quality.

Heterogeneity has been discussed, however the implications of high variability across studies (e.g., imaging modality, AI model type) could be discussed further.

Discussion section (page 22):

-High levels of heterogeneity (I2>95) were found regarding sensitivity, specificity, and log diagnostic odds ratio among the included studies. These levels of heterogeneity can be explained by diversity in imaging modalities (e.g., bitewing vs. panoramic radiographs), differences in AI algorithm architectures, and training dataset characteristics.

The impact of publication bias on the pooled estimates should be elaborated.

Results section (page 17):

-The nonparametric trim-and-fill analysis was employed to assess the impact of publication bias on the pooled estimates and results were presented at Table 6

The references have to be formatted according to journal guidelines.

-Done

Reviewer #2: Review Comments to the Author:

The topic of this manuscript is timely and relevant. The authors present a broad summary of existing evidence on the use of artificial intelligence (AI) for detecting dental caries, which seems important and useful for the field.

Strengths:

The methodology is clearly described, and the tools used for analysis appear to be appropriate and well chosen for the study’s objective.

The conclusions are well linked to the presented data, and the results are carefully analyzed.

It is also appreciated that the underlying data are fully available.

Suggestions:

It might be helpful to discuss more clearly the limitations related to the fact that only a small number of original studies were included in the meta-analysis.

Discussion section (page 22):

-The main limitation of this umbrella review and meta-analysis include the fact that only 20 original studies were included in meta-analysis quantitative data synthesis. The number of primary original research studies was 137, with only 20 (14.5 %) original research articles reporting the necessary details of AI–based caries diagnostic test results including numbers TP, TN, FP, and FN involved in the meta-analysis. To facilitate future meta-analysis, authors are encouraged to report details of AI–based caries diagnostic tests including numbers of TP, TN, FP, and FN, all of which are essential for a meta-analysis.

A short explanation on how overlapping studies across the included reviews were handled could also be useful.

Methods section (page 7):

-Overlapping studies across the included reviews were handled by creation of citation matrix and calculation of corrected covered area (CCA) using ccaR package of R software (R Foundation for Statistical Computing, Vienna, Austria).

Results section (page 15):

-The citation matrix of primary studies included in the systematic reviews was presented at Supplementary Table 1. The CCA_Proportion was 0.07 and CCA_Percentage was 6.90, showing moderate overlap. The pairwise CCA presented at Supplementary Figure 1, showed which combinations of paired reviews had the highest overlap.

The manuscript would benefit from more critical reflection on how the presence or absence of caries was determined in the original studies. There is little detail on the reference standard, and it remains unclear whether definitions of caries were consistent, validated, or clinically reliable.

Results section (page 15):

-Table 5 was added to manuscript. At this table the caries detection method, reference standard and AI algorithm were presented.

It would strengthen the discussion to acknowledge that the most reliable conclusions would come from reviews comparing studies conducted under the same diagnostic protocols and caries definitions.

Discussion section (page 20):

-It is well-known that, the most reliable conclusions would come from reviews comparing studies conducted under the same diagnostic protocols and caries definitions. As showed at Table 5, caries detection method, reference standard and AI algorithm were same among included studies in meta-analysis.

The type of dental imaging strongly influences diagnostic performance. AI evaluating panoramic radiographs (with tooth overlap) operates under very different conditions than AI analyzing bitewing images (with clear interproximal visibility). Pooling such results may reduce the clinical interpretability of the findings and should be acknowledged as a limitation.

Discussion section (page 19):

-Readers must note, the type of dental imaging strongly influences diagnostic performance. AI evaluating panoramic radiographs (with tooth overlap) operates under very different conditions than AI analyzing bitewing images (with clear interproximal visibility). Pooling such results may reduce the clinical interpretability of the findings. To address this issue we conduct subgroup meta-analysis for different imaging modalities and results presented at Figure 4.

In future analyses, the most valuable evidence will likely come from studies that simulate real-world diagnostic conditions, including time pressure, clinician–patient interaction, and the integration of AI as part of the clinical workflow rather than in isolated image review.

A recently published study demonstrates the importance of incorporating such clinical constraints in AI evaluation and could be considered for citation:

-Done

Attachment

Submitted filename: Reviewers comments.docx

pone.0329986.s005.docx (18.8KB, docx)

Decision Letter 1

Giang Vu

25 Jul 2025

Examining the diagnostic accuracy of artificial intelligence for detecting dental caries across a range of imaging modalities: An umbrella review with meta-analysis

PONE-D-25-05467R1

Dear Dr. Arzani,

We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements.

Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication.

An invoice will be generated when your article is formally accepted. Please note, if your institution has a publishing partnership with PLOS and your article meets the relevant criteria, all or part of your publication costs will be covered. Please make sure your user information is up-to-date by logging into Editorial Manager at Editorial Manager®  and clicking the ‘Update My Information' link at the top of the page. For questions related to billing, please contact billing support .

If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

Kind regards,

Giang Truong Vu, D.D.S., M.S., Ph.D.

Academic Editor

PLOS ONE

Acceptance letter

Giang Vu

PONE-D-25-05467R1

PLOS ONE

Dear Dr. Arzani,

I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now being handed over to our production team.

At this stage, our production department will prepare your paper for publication. This includes ensuring the following:

* All references, tables, and figures are properly cited

* All relevant supporting information is included in the manuscript submission,

* There are no issues that prevent the paper from being properly typeset

You will receive further instructions from the production team, including instructions on how to review your proof when it is ready. Please keep in mind that we are working through a large volume of accepted articles, so please give us a few days to review your paper and let you know the next and final steps.

Lastly, if your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

You will receive an invoice from PLOS for your publication fee after your manuscript has reached the completed accept phase. If you receive an email requesting payment before acceptance or for any other service, this may be a phishing scheme. Learn how to identify phishing emails and protect your accounts at https://explore.plos.org/phishing.

If we can help with anything else, please email us at customercare@plos.org.

Thank you for submitting your work to PLOS ONE and supporting open access.

Kind regards,

PLOS ONE Editorial Office Staff

on behalf of

Dr. Giang Truong Vu

Academic Editor

PLOS ONE

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    S1 Table. The citation matrix of primary studies included in the systematic reviews for the use of AI in the detection of dental caries.

    The “1” implies a checkmark, that is the study is included “0” implies that the study is not included in the review in question.

    (DOCX)

    pone.0329986.s001.docx (562.6KB, docx)
    S1 Fig. Visualization of the pairwise CCA (%) with a heatmap.

    (TIF)

    pone.0329986.s002.tif (112.3KB, tif)
    S1 File. PRISMA+DTA+Checklist.

    (DOCX)

    pone.0329986.s003.docx (30.1KB, docx)
    Attachment

    Submitted filename: Rebuttal letter.docx

    pone.0329986.s004.docx (12.8KB, docx)
    Attachment

    Submitted filename: Reviewers comments.docx

    pone.0329986.s005.docx (18.8KB, docx)

    Data Availability Statement

    All relevant data are within the manuscript and its Supporting Information files.


    Articles from PLOS One are provided here courtesy of PLOS

    RESOURCES