Abstract
Purpose
To compare the trueness of artificial intelligence (AI)‐based, manual, and global segmentation protocols by superimposing the resulting segmented 3D models onto reference gold standard surface scan models.
Materials and Methods
Twelve dry human mandibles were used. A cone beam computed tomography (CBCT) scanner was used to scan the mandibles, and the acquired digital imaging and communications in medicine (DICOM) files were segmented using three protocols: global thresholding, manual, and AI‐based segmentation (Diagnocat; Diagnocat, San Francisco, CA). The segmented files were exported as study 3D models. A structured light surface scanner (GoSCAN Spark; Creaform 3D, Levis, Canada) was used to scan all mandibles, and the resulting reference 3D models were exported. The study 3D models were compared with the respective reference 3D models by using a mesh comparison software (Geomagic Design X; 3D Systems Inc, Rock Hill, SC). Root mean square (RMS) error values were recorded to measure the magnitude of deviation (trueness), and color maps were obtained to visualize the differences. Comparisons of the trueness of three segmentation methods for differences in RMS were made using repeated measures analysis of variance (ANOVA). A two‐sided 5% significance level was used for all tests in the software program.
Results
AI‐based segmentations had significantly higher RMS values than manual segmentations for the entire mandible (p < 0.001), alveolar process (p < 0.001), and body of the mandible (p < 0.001). AI‐based segmentations had significantly lower RMS values than manual segmentations for the condyles (p = 0.018) and ramus (p = 0.013). No significant differences were found between the AI‐based and manual segmentations for the coronoid process (p = 0.275), symphysis (p = 0.346), and angle of the mandible (p = 0.344). Global thresholding had significantly higher RMS values than manual segmentations for the alveolus (p < 0.001), angle of the mandible (p < 0.001), body of the mandible (p < 0.001), condyles (p < 0.001), coronoid (p = 0.002), entire mandible (p < 0.001), ramus (p < 0.001), and symphysis (p < 0.001). Global thresholding had significantly higher RMS values than AI‐based segmentation for the alveolar process (p = 0.002), angle of the mandible (p < 0.001), body of the mandible (p < 0.001), condyles (p < 0.001), coronoid (p = 0.017), mandible (p < 0.001), ramus (p < 0.001), and symphysis (p < 0.001).
Conclusion
AI‐based segmentations produced lower RMS values, indicating truer 3D models, compared to global thresholding, and showed no significant differences in some areas compared to manual segmentation. Thus, AI‐based segmentation offers a level of segmentation trueness acceptable for use as an alternative to manual or global thresholding segmentation protocols.
Keywords: AI, artificial intelligence, CBCT, cone beam computed tomography, 3D model, machine learning
Segmentation involves partitioning scanned images to isolate and define boundaries within regions of interest. 1 , 2 Accurate segmentation results in 3D models crucial for diagnosis and treatment planning. 3 , 4 , 5 Success in segmentation and 3D modeling heavily relies on cone beam computed tomography (CBCT) imaging and data acquisition, with factors such as the type of scanner, imaging technique, angle, and resolution significantly influencing algorithm success. 6 , 7 CBCT artifacts, often seen in the lingual and posterior borders of the mandible, condyles, and alveolar crest, can introduce streak and ring halation defects, distorting the digital dataset. 8 , 9 Additionally, image slice spacing and surface geometry changes can impact the accuracy of the 3D model. 1 Loss of image detail may also occur during post‐processing due to manual editing. 3 , 10 Finally, user skill plays a crucial role, as segmentations by commercial software companies have been shown to yield more accurate 3D models than those performed by clinicians. 9 , 11
Global thresholding uses a threshold value “t” to include voxel gray values greater or equal to “t” in the segmented image. 5 The method's simplicity makes it widely available but has drawbacks, leading to the least accurate 3D models. Challenges in delineating borders arise when different tissue voxels are at the borders, and the method does not account for CBCT scan errors like metal artifacts or noise. 5 , 10 , 12 Global thresholding requires extensive post‐processing and expertise, increasing the cost and time for accurate 3D models. 5 , 12 In manual segmentation, each part of a CBCT image is outlined slice‐by‐slice and then compiled into 2D or 3D images. It can address issues such as missing slices, recognized as ground truth in medical image analysis due to expert input. 13 This method produces highly accurate segmentations with volumes matching the original region of interest. 14 However, due to its precision and the expertise required, manual segmentation is costly, time‐intensive, and subject to variability and subjective bias. 13 , 14
To address the limitations of manual segmentation, semi‐automatic and automatic segmentation methods, including artificial intelligence (AI)‐based auto‐segmentation, have been developed. These use computational tools to delineate regions of interest rapidly with little or no manual input. 15 , 16 , 17 , 18 For example, one AI technology (Diagnocat; Diagnocat, San Francisco, CA) uses a deep multilayer convolutional neural network (CNN) that makes use of a U‐net‐like architecture, which can simulate human brain architecture and processes data through a network of so‐called interconnected neurons. 19 AI‐based segmented 3D models are instrumental in dental and maxillofacial radiology, assisting in diagnosing cystic lesions, treatment planning, and creating surgical implant guides. 20 However, despite their potential, many advanced segmentation software programs are not yet commercially available. 5
Limited research has compared AI‐based segmentation to global thresholding, the most straightforward and prevalent technique, and the highly accurate manual segmentation. Moreover, the accuracy of AI‐based segmentation has been underexplored despite its potential to outperform human and other methods and become a market‐ready software. This study aimed to evaluate the trueness of AI‐based segmentation by comparing it with global thresholding and manual segmentation, using superimposition and comparison against a reference gold standard surface scan model. The null hypothesis was that the segmentation protocol has no effect on the trueness of the resulting 3D models.
MATERIALS AND METHODS
This study was reviewed by the Human Research Protection Program (HRPP) before inception, and it was determined that further review was not required (#19615). The inclusion criteria specified that each mandible must contain at least one tooth, as the AI‐based segmentation software (Diagnocat; Diagnocat) included in this study requires the presence of one remaining tooth. The exclusion criteria ruled out mandibles with fractures or defects that could compromise anonymity or interfere with accurately comparing anatomical landmarks. Twelve anonymized, dry, and macerated mandibles were obtained from the Department of Anatomy, Cell Biology, and Physiology and included in this study (Figure 1a). With a sample size of 12 mandibles, the study had 80% power to detect an effect size of 1.0 between methods, based on a paired t‐test calculation at a two‐sided 5% significance level.
FIGURE 1.

(a) Representative image of a dried mandible included in the study; (b) All mandibles were scanned using a structured light 3D surface scanner. The resulting 3D surface models were exported as reference standard tessellation language (STL) files.
All procedures were standardized among researchers before data experimentation. An experienced radiologist defined each segmentation process and mandibular landmarks on the reference 3D models. Examiners underwent training, and calibration was assessed by repeating intra‐examiner readings at a one‐week interval until intraclass correlation coefficients (ICC) exceeded 0.80. After calibration, the first examiner randomized and numbered the segmented digital imaging and communications in medicine (DICOM) files, maintained a master sheet for data matching, and sent blinded files to the second examiner, who completed segmentation, alignment, measurement, and comparisons.
All mandibles were scanned using a structured light surface scanner (GoSCAN Spark; Creaform 3D, Levis, Canada) (Figure 1b). The resulting 3D surface models were exported as standard tessellation language (STL) files. These models were then imported into computer‐aided design and computer‐aided manufacturing (CAD‐CAM) software (Autodesk Meshmixer; Autodesk Inc, San Francisco, CA) to select and delineate the mandible's anatomical regions for comparative analysis. Seven anatomical landmarks critical for implant‐related planning in the mandible were identified on the 3D surface models. These included the right and left condyles, coronoid processes, rami, angles, body, remaining teeth and alveolar process, and symphysis. Each region and the entire mandible were saved as a separate STL file to serve as a reference 3D model for subsequent comparisons in the study (Figure 2).
FIGURE 2.

Anatomical regions, including the right and left condyles, coronoid processes, rami, angles, the body of the mandible, remaining teeth, alveolar process and remaining teeth, symphysis, and the entire mandible (regions a to g combined), were saved as individual standard tessellation language (STL) files. These files served as references for subsequent comparative analyses in the study. (a) Condyles; (b) Coronoid processes; (c) Rami; (d) Angles of the mandible; (e) Body of the mandible; (f) Symphysis; (g) Remaining teeth and alveolar process.
All mandibles were scanned using a CBCT scanner (Planmeca Viso G7; Planmeca USA Inc, Hoffman Estates, IL) with a field of view (FOV) measuring 100 × 140 mm, voxel size of 0.15 mm, anode voltage of 100 kV, and anode current of 63 mAs per second. The acquired DICOM files were exported and stored. Three different segmentation protocols were employed to create the 3D models: global thresholding segmentation, manual segmentation, and AI‐based segmentation. Initially, the DICOM file was uploaded to the segmentation and implant planning software (BlueSkyPlan; BlueSky Bio, Liberty Drive, IL, USA), where the FOV and panoramic curve were identified. To obtain the global thresholding segmentation 3D model, a single threshold value was selected for inclusion in the segmented model, effectively removing all scatter, noise, and artifacts from the imported CBCT scans. The CBCT scan was then manually segmented slice‐by‐slice using the same software. The segmentation process began in the axial view, using the interactive lasso tool to select the entire slice, followed by the manual brush tool, if necessary, to define the whole outline of the slice. Subsequently, the “fill hole” option was used to segment the remaining areas within the borders. After completing the segmentation of all slices from the axial view, the process was repeated in the cross‐sectional view until all slices were segmented. For the AI‐based segmentations, the CBCT scans were imported into an AI‐based segmentation software (Diagnocat; Diagnocat), where the software fully automated the segmentations. The resulting study STL files of 3D models from all segmentation protocols were saved and exported for comparative analysis.
The STL files generated by the three segmentation protocols were compared with the respective reference files from the surface scanner using point‐based, global, and fine alignment methods. All STL files were imported into CAD software (Geomagic Design X; 3D Systems Inc, Rock Hill, SC). An initial alignment was executed using four consistent anatomical reference points: the right and left mental foramina and the right and left antegonial notches to superimpose the study STL files onto the reference files. A best‐fit algorithm incorporating global and fine registration was applied to automatically align all study and reference files. A 3D compare tool was employed to evaluate mesh deviations (Figure 3). Root mean square (RMS) values were recorded to measure the magnitude of deviation (trueness) for each segmented file compared to the reference surface scans. Color maps were created to visualize the differences, with shades from yellow to red indicating positive deviations, signifying an overestimation of the segmented files relative to the references. Conversely, blue indicated negative deviations, signifying an underestimation of the segmented files compared to the references.
FIGURE 3.

(a) Standard tessellation language (STL) file from the surface scan used as a reference; (b) STL file of a study 3D model from segmentation; (c) A best‐fit algorithm incorporating global and fine registration was applied to align all study and reference files automatically.
Comparisons among the global thresholding, manual segmentation, and AI‐based segmentation methods for differences in RMS were made using repeated measures analysis of variance (ANOVA), which allowed correlations among the measurements from the three methods on the same mandible and different variances for each method. A two‐sided 5% significance level was used for all tests in a software program (SAS version 9.4; SAS Institute Inc, Cary, NC).
RESULTS
The RMS values for global thresholding, manual, and AI‐based segmentation are summarized in Table 1. A boxplot was used to depict the graphic summary of RMS values from each subgroup (Figure 4). Comparisons among the global thresholding, manual segmentation, and AI‐based segmentation methods for differences in RMS were made using repeated measures ANOVA and pair‐wise comparisons (Table 2). Based on pair‐wise comparison data, AI‐based segmentations had significantly higher RMS values than manual segmentations for the entire mandible (p < 0.001), alveolar process and remaining teeth (p < 0.001), and body of the mandible (p < 0.001). AI‐based segmentations had significantly lower RMS values than manual segmentations for the condyles (p = 0.018) and rami (p = 0.013). No significant differences were found between pair‐wise comparisons between RMS values for AI‐based and manual segmentations for the coronoid process (p = 0.275), symphysis (p = 0.346), and angles of the mandible (p = 0.344).
TABLE 1.
The descriptive statistics of each group's root mean square (RMS, mm) and mean (standard deviation).
| Location | Segmentation methods | Mean (standard deviation) |
|---|---|---|
| Alveolar process and remaining teeth | AI‐based | 1.304 (0.093) |
| Global thresholding | 1.529 (0.202) | |
| Manual | 0.943 (0.168) | |
| Angles of the mandible | AI‐based | 1.089 (0.238) |
| Global thresholding | 1.927 (0.179) | |
| Manual | 0.986 (0.226) | |
| Body of the mandible | AI‐based | 1.432 (0.209) |
| Global thresholding | 2.100 (0.114) | |
| Manual | 0.659 (0.154) | |
| Condyles | AI‐based | 0.742 (0.443) |
| Global thresholding | 1.951 (0.342) | |
| Manual | 1.134 (0.331) | |
| Coronoid processes | AI‐based | 0.748 (0.246) |
| Global thresholding | 1.026 (0.345) | |
| Manual | 0.709 (0.217) | |
| Rami | AI‐based | 0.684 (0.117) |
| Global thresholding | 1.839 (0.278) | |
| Manual | 0.947 (0.257) | |
| Symphysis | AI‐based | 0.516 (0.570) |
| Global thresholding | 2.257 (0.243) | |
| Manual | 0.668 (0.316) | |
| Mandible (entire mandible) | AI‐based | 1.148 (0.112) |
| Global thresholding | 1.871 (0.150) | |
| Manual | 0.794 (0.169) |
Abbreviation: AI, artificial intelligence; RMS, root mean square.
FIGURE 4.

Boxplot showing the graphic summary of descriptive statistics on root mean square (RMS) values.
TABLE 2.
Pair‐wise comparisons were made to compare RMS values from all subgroups.
| Locations | Comparison made | Difference | SE | p‐value | |
|---|---|---|---|---|---|
| Alveolar process and remaining teeth | AI‐based | Global thresholding | −0.2247 | 0.0536 | 0.002 * |
| AI‐based | Manual | 0.3607 | 0.0385 | <0.001 * | |
| Global thresholding | Manual | 0.5854 | 0.0510 | <0.001 * | |
| Angles of the mandible | AI‐based | Global thresholding | −0.8385 | 0.0952 | <0.001 * |
| AI‐based | Manual | 0.1028 | 0.1039 | 0.344 | |
| Global thresholding | Manual | 0.9413 | 0.0563 | <0.001 * | |
| Body of the mandible | AI‐based | Global thresholding | −0.6679 | 0.0685 | <0.001 * |
| AI‐based | Manual | 0.7728 | 0.0626 | <0.001 * | |
| Global thresholding | Manual | 1.4407 | 0.0411 | <0.001 * | |
| Condyles | AI‐based | Global thresholding | −1.2087 | 0.1860 | <0.001 * |
| AI‐based | Manual | −0.3920 | 0.1417 | 0.018 * | |
| Global thresholding | Manual | 0.8167 | 0.1406 | <0.001 * | |
| Coronoid processes | AI‐based | Global thresholding | −0.2781 | 0.0990 | 0.017 * |
| AI‐based | Manual | 0.0390 | 0.0340 | 0.275 | |
| Global thresholding | Manual | 0.3171 | 0.0806 | 0.002 * | |
| Rami | AI‐based | Global thresholding | −1.1552 | 0.0952 | <0.001 * |
| AI‐based | Manual | −0.2632 | 0.0888 | 0.013 * | |
| Global thresholding | Manual | 0.8921 | 0.0520 | <0.001 * | |
| Symphysis | AI‐based | Global thresholding | −1.7406 | 0.1762 | <0.001 * |
| AI‐based | Manual | −0.1522 | 0.1547 | 0.346 | |
| Global thresholding | Manual | 1.5884 | 0.1124 | <0.001 * | |
| Mandible (entire mandible) | AI‐based | Global thresholding | −0.7234 | 0.0467 | <0.001 * |
| AI‐based | Manual | 0.3535 | 0.0474 | <0.001 * | |
| Global thresholding | Manual | 1.0769 | 0.0360 | <0.001 * |
Abbreviation: AI, artificial intelligence; RMS, root mean square.
Denotes the significant finding, α = 0.05.
Global thresholding had significantly higher RMS values than manual segmentations for the alveolar process and remaining teeth (p < 0.001), angles of the mandible (p < 0.001), body of the mandible (p < 0.001), condyles (p < 0.001), coronoid processes (p = 0.002), entire mandible (p < 0.001), rami (p < 0.001), and symphysis (p < 0.001). AI had significantly lower RMS than global thresholding for the alveolar process and remaining teeth (p = 0.002), angles of the mandible (p < 0.001), body of the mandible (p < 0.001), condyles (p < 0.001), coronoid processes (p = 0.017), entire mandible (p < 0.001), rami (p < 0.001), and symphysis (p < 0.001).
Color maps of the surface matching differences for each group were evaluated, where green‐colored areas represented surface matching within a range of ±0.1 mm. Global thresholding predominantly revealed yellow to red areas, indicating greater overestimation and positive deviation compared to the gold standard surface scan (Figure 5a). The color mapping of manual segmentation ranged from green to light blue, suggesting minimal deviation and higher trueness compared to the surface scans (Figure 5b). AI segmentations exhibited a positive deviation, with the majority of areas ranging from green to yellow. It is important to note that some areas in the AI‐based segmentation, although AI could recognize, were shaded in dark blue, thus representing grossly negative deviations in these areas (Figure 5c). It is also important to note that AI‐based segmentation failed to recognize specific areas, such as the four condyles were omitted entirely from the segmentation (Figure 6).
FIGURE 5.

Color maps representing deviation (in mm) from all 12 mandibles. (a) Global thresholding; (b) Manual segmentation; (c) Artificial intelligence (AI)‐based segmentation.
FIGURE 6.

(a) Manual segmentation of a mandible with condyle present; (b) Artificial intelligence (AI)‐based segmentation of the same mandible. The red arrow indicates the missing condyle in the AI‐based segmentation protocol.
DISCUSSIONS
The null hypothesis was partially rejected as AI segmentation, for most areas analyzed, had significantly higher RMS values, more significant deviation than manual segmentation, and significantly lower RMS values than global thresholding.
In prosthodontics, CBCT segmentation has been utilized to create anatomical bone models to facilitate the planning of esthetic crown lengthening and the design of surgical templates. 21 When there is a need to duplicate a clinically satisfactory complete denture as a trial prosthesis or custom tray, CBCT segmentation has also been proposed to obtain a digital file of the denture. A study demonstrated that global thresholding in this duplication technique resulted in a digital denture file with an accuracy level of 0.249 ± 0.020 mm. 22 CBCT enables 3D imaging of craniofacial hard tissues but has limited FOV and contrast for soft tissues. Integrating facial scans or photos enhances CBCT segmentation, creating a photorealistic 3D virtual patient for improved diagnosis, treatment planning, and surgical simulation. 23 Additionally, CBCT segmentation allows custom articulators by replicating patient‐specific jaw movements. 24 It also enables a digital workflow in maxillofacial prosthodontics, using 3D modeling for patients with limited mouth opening when conventional impressions are impractical. 25 , 26 , 27 Additionally, segmentation accurately records maxillofacial defects, integrating with intraoral or facial scans to create precise jaw and defect replicas. 25 , 26 , 27 CBCT segmentation is widely used in implant dentistry as well, enabling the superimposition of segmented dentition onto intraoral and facial scans for designing implant placement templates. 28 , 29 , 30 In severe mandibular atrophy, it facilitates custom 3D‐printed Ti6Al4V subperiosteal implants, reducing the need for complex regenerative procedures. 31 It also preserves the natural emergence profile before extraction, aiding in custom healing abutment design. 32 Despite its broad applications, research comparing segmentation protocols is limited. In this study, the trueness of the AI‐based segmentation protocol ranged from 0.516 ± 0.570 to 1.432 ± 0.209 mm, with missing data in 4 out of 24 condyles. Variability and inaccuracies near critical structures pose risks, highlighting the need for optimized, accurate, and user‐friendly segmentation protocols to ensure patient safety.
Global thresholding is a simple and widely used segmentation technique, relying on a single threshold value for segmentation. 5 In this study, global thresholding had the greatest deviation in all locations compared to the gold standard surface scan. Choosing a single threshold value makes it difficult to delineate the borders due to the multiple types of voxels in these areas. 5 , 10 The final segmented images of all mandibles were overestimated versions of their surface scan counterparts. A single threshold struggles to delineate borders due to voxel variability, artifacts, noise, and scanner differences, causing inconsistencies. 5 , 12 Pre‐processing methods like background illumination and top‐hat filtering can enhance object‐background distinction, but global thresholding remains limited as it lacks spatial pixel distribution information and is prone to inaccuracies with artifacts. 33
Manual segmentation was the truest method in this study, consistently achieving lower RMS values than global thresholding and AI segmentation in the alveolus, body of the mandible, and the entire mandible. It involves slice‐by‐slice annotation, ensuring accuracy but making it time‐consuming and labor‐intensive. 34 , 35 , 36 This method also suffers from inter‐ and intra‐observer variability, often requiring radiologists or specialized clinicians. 13 , 14 The results of this study should also be interpreted with caution, as the manual segmentation was conducted by a trained and calibrated research team, which may reduce inter‐ and intra‐observer variability compared to what is observed in clinical outcomes. Radiodense structures like teeth and compact bone are easier to segment, whereas thinner cortical bone, particularly in the condyle and edentulous spaces, showed greater RMS values than AI‐based segmentation. These regions were more challenging to segment, requiring manual adjustments, which contributed to increased variability and segmentation difficulty in this study.
AI‐based segmentation is increasingly favored for its reproducibility and reduced human error. The system in this study employs a CNN with data‐driven algorithms for rapid, automated segmentation. 17 , 18 Small errors can be manually refined, and some metal artifacts are automatically removed. 37 AI segmentation had significantly lower RMS values than global thresholding and was truer than manual segmentation in the condyles and ramus, with no significant differences in the coronoid process, symphysis, or mandibular angle. However, it failed to recognize four condyles (Figure 6), likely due to insufficient database representation, leading to underestimation in color mapping. AI segmentation also had higher RMS values than manual segmentation in the entire mandible, alveolar process, and body of the mandible, resulting in overestimation of these structures. These variations are likely due to anatomical differences between subjects. Despite these limitations, AI segmentation has strong potential, with accuracy expected to improve as more data are incorporated. Similar to the findings from this study, past research has shown favorable results in the accuracy of AI segmentations comparable to the manual segmentations done by experienced professionals. Kargilis et al. utilized CBCT scans of mandible and mandibular dentition segmentations that have undergone model training and testing. 38 The AI segmentation achieved high quantitative accuracy, with an average dice similarity coefficient (DSC) from 0.940 to 0.945. In the blind qualitative assessment, oral and maxillofacial surgeons rated the AI‐generated segmentations as comparable in quality to manual segmentations. 38 Lo Giudice et al. compared 20 manual CBCT‐derived segmentations with CNN‐based AI segmentations and found comparable accuracy for delineating the entire mandible. 39 Verhelst et al. utilized layered 3D U‐Net architecture deep learning algorithms to perform AI segmentations with and without user refinement. 40 Manual segmentations averaged 1218.4 s while AI segmentations averaged 17 s to complete. Mean RMS values between fully automatic AI and manual segmentations were 0.2624 mm and the intersection of union of 94.6%. Research has shown favorable results for AI segmentations, both in terms of accuracy as well as timing. 40 Recent research found that a newly developed SegResNet‐based deep learning model achieved high accuracy and consistency when automatically segmenting the mandible in CBCT scans from patients with stage III to IV periodontitis. Quantitative analysis showed excellent agreement between the model's outputs and semi‐automatic segmentations. 41 In a recent systematic review, investigators evaluated segmentation algorithms using performance metrics such as DSC. 42 For mandibular segmentation, the pooled DSC was 0.94 (95% CI: 0.91–0.98), indicating excellent accuracy. Deep‐learning architectures such as CNNs and U‐Nets consistently outperformed classical machine‐learning techniques. Most published work, however, used expert‐manual segmentation as the reference gold standard. 42 By contrast, the present study adopts a surface laser scan as the ground truth, a choice that may yield a more accurate reference and, in turn, more reliable comparative results. The variability in reported metrics highlights the need for uniform evaluation and reporting protocols.
This study has several limitations. A high‐resolution voxel size of 0.15 mm was used for segmentation, but larger voxels and other scanning parameters (mA, kVp, post‐processing) could affect accuracy, requiring further investigation. 43 Only one CBCT scanner was used to limit variables, though scanner type and voxel range may influence segmentation accuracy. AI segmentation was performed with a single software, which may not represent all AI‐based tools. Due to the limitations of the AI software and its version, the remaining teeth and alveolar process were segmented as one entity for comparison in this study. Newer AI software can now segment the remaining teeth as their own entity, allowing for more refined comparisons. The study focused on the mandible, excluding the maxilla due to its complex anatomy, higher artifact incidence, and less dense cortical bone, which could complicate segmentation. 44 Additionally, mandibles were scanned without soft tissue, simplifying segmentation compared to real human scans. Future research should assess how soft tissue presence affects segmentation accuracy across different methods.
CONCLUSIONS
AI‐based segmentations produced lower RMS values than global thresholding and showed no notable differences in most areas of the mandible when evaluated against manual segmentation. Nevertheless, there remains a statistically significant difference, with manual segmentation achieving lower RMS values in the alveolus, body of the mandible, and across the entire mandible. Despite these findings, the outcomes of this study highlight the capability of AI‐based segmentation protocols to serve as effective alternatives to both global thresholding and manual segmentation.
ACKNOWLEDGMENTS
The authors did not receive any grants from funding agencies nor were they under any contractual agreement for this research. The authors express their gratitude to Doug Gillespie and Diagnocat AI (DGNCT LLC, Miami, FL) for providing the AI segmentations used in this research at no cost. The authors also extend their thanks to Dr. Andrew S. Deane, Vice Chair for Outreach, for the invaluable support in providing mandible samples from the Department of Anatomy, Cell Biology & Physiology at the Indiana University School of Medicine. Special acknowledgment is also due to Chauncey Frend, Sr. Analyst Programmer at the UITS RT Advanced Visualization Lab, part of Indiana University Information Technology Services, for the dedicated efforts in conducting the surface scans of the mandibles. In addition, the authors thank George J. Eckert and the team from the Department of Biostatistics and Health Data Science, Richard M. Fairbanks School of Public Health, for their biostatistical support.
Hernandez AKT, Dutra V, Chu T‐MG, Yang C‐C, Lin W‐S. Trueness of artificial intelligence‐based, manual, and global thresholding segmentation protocols for human mandibles. J Prosthodont. 2025;34:939–946. 10.1111/jopr.70008
REFERENCES
- 1. Rathnayaka K, Sahama T, Schuetz MA, Schmutz B. Effects of CT image segmentation methods on the accuracy of long bone 3D reconstructions. Med Eng Phys. 2011;33(2):226–233. [DOI] [PubMed] [Google Scholar]
- 2. Torosdagli N, Liberton DK, Verma P, Sincan M, Lee JS, Bagci U. Deep geodesic learning for segmentation and anatomical landmarking. IEEE Trans Med Imaging. 2019;38(4):919–931. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 3. Szymor P, Kozakiewicz M, Olszewski R. Accuracy of open‐source software segmentation and paper‐based printed three‐dimensional models. J Craniomaxillofac Surg. 2016;44(2):202–209. [DOI] [PubMed] [Google Scholar]
- 4. Dawood A, Marti Marti B, Sauret‐Jackson V, Darwood A. 3D printing in dentistry. Br Dent J. 2015;219(11):521–529. [DOI] [PubMed] [Google Scholar]
- 5. van Eijnatten M, van Dijk R, Dobbe J, Streekstra G, Koivisto J, Wolff J. CT image segmentation methods for bone used in medical additive manufacturing. Med Eng Phys. 2018;51:6–16. [DOI] [PubMed] [Google Scholar]
- 6. Drapikowski P. Surface modeling–uncertainty estimation and visualization. Comput Med Imaging Graph. 2008;32(2):134–139. [DOI] [PubMed] [Google Scholar]
- 7. Gerlach NL, Meijer GJ, Kroon DJ, Bronkhorst EM, Bergé SJ, Maal TJ. Evaluation of the potential of automatic segmentation of the mandibular canal using cone‐beam computed tomography. Br J Oral Maxillofac Surg. 2014;52(9):838–844. [DOI] [PubMed] [Google Scholar]
- 8. Liang X, Lambrichts I, Sun Y, Denis K, Hassan B, Li L, et al. A comparative evaluation of Cone Beam Computed Tomography (CBCT) and Multi‐Slice CT (MSCT). Part II: on 3D model accuracy. Eur J Radiol. 2010;75(2):270–274. [DOI] [PubMed] [Google Scholar]
- 9. Fourie Z, Damstra J, Schepers RH, Gerrits PO, Ren Y. Segmentation process significantly influences the accuracy of 3D surface models derived from cone beam computed tomography. Eur J Radiol. 2012;81(4):e524–e530. [DOI] [PubMed] [Google Scholar]
- 10. Rathnayaka K, Momot KI, Noser H, Volp A, Schuetz MA, Sahama T, et al. Quantification of the accuracy of MRI generated 3D models of long bones compared to CT generated 3D models. Med Eng Phys. 2012;34(3):357–363. [DOI] [PubMed] [Google Scholar]
- 11. Salmi M, Paloheimo KS, Tuomi J, Wolff J, Mäkitie A. Accuracy of medical models made by additive manufacturing (rapid manufacturing). J Craniomaxillofac Surg. 2013;41(7):603–609. [DOI] [PubMed] [Google Scholar]
- 12. Qiu B, Guo J, Kraeima J, Glas HH, Borra RJH, Witjes MJH, et al. Automatic segmentation of the mandible from computed tomography scans for 3D virtual surgical planning using the convolutional neural network. Phys Med Biol. 2019;64(17):175020. [DOI] [PubMed] [Google Scholar]
- 13. Egger J, Nimsky C, Chen X. Vertebral body segmentation with GrowCut: initial experience, workflow and practical application. SAGE Open Med. 2017;5:2050312117740984. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14. Wallner J, Schwaiger M, Hochegger K, Gsaxner C, Zemann W, Egger J. A review on multiplat‐form evaluations of semi‐automatic open‐source based image segmentation for cranio‐maxillofacial surgery. Comput Methods Programs Biomed. 2019;182:105102. [DOI] [PubMed] [Google Scholar]
- 15. Sharp G, Fritscher KD, Pekar V, Peroni M, Shusharina N, Veeraraghavan H, et al. Vision 20/20: perspectives on automated image segmentation for radiotherapy. Med Phys. 2014;41(5):050902. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 16. Zhu W, Huang Y, Zeng L, Chen X, Liu Y, Qian Z, et al. AnatomyNet: deep learning for fast and fully automated whole‐volume segmentation of head and neck anatomy. Med Phys. 2019;46(2):576–589. [DOI] [PubMed] [Google Scholar]
- 17. Savadjiev P, Chong J, Dohan A, Vakalopoulou M, Reinhold C, Paragios N, et al. Demystification of AI‐driven medical image interpretation: past, present and future. Eur Radiol. 2019;29(3):1616–1624. [DOI] [PubMed] [Google Scholar]
- 18. Hung K, Yeung AWK, Tanaka R, Bornstein MM. Current applications, opportunities, and limitations of AI for 3D imaging in dental research and practice. Int J Environ Res Public Health. 2020;17(12):4424. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 19. Orhan K, Bayrakdar IS, Ezhov M, Kravtsov A, Özyürek T. Evaluation of artificial intelligence for detecting periapical pathosis on cone‐beam computed tomography scans. Int Endod J. 2020;53(5):680–689. [DOI] [PubMed] [Google Scholar]
- 20. Hung K, Montalvao C, Tanaka R, Kawai T, Bornstein MM. The use and performance of artificial intelligence applications in dental and maxillofacial radiology: a systematic review. Dentomaxillofac Radiol. 2020;49(1):20190107. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 21. Kim J, Lin YC, Danielak M, Van M, Lee DH, Kim H, et al. Virtual planning and rapid 3D prototyping surgical guide for anterior crown lengthening surgery: a clinical case report. J Prosthodont. 2022;31(4):275–281. [DOI] [PubMed] [Google Scholar]
- 22. Alehaideb A, Lin WS, Levon JA, Chu TG, Yang CC. Accuracy of digital duplication scanning methods for complete dentures. J Prosthodont. 2023;34(1):42–48. 10.1111/jopr.13788 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 23. Harris BT, Yang CC, Morton D, Lin WS. Virtual facial simulation of prosthetic outcome for static computer‐aided implant surgery and CAD‐CAM prostheses. J Oral Implantol. 2022;48(1):51–58. [DOI] [PubMed] [Google Scholar]
- 24. Azer SS, Kemper E. The patient‐specific anatomical articulator. J Prosthet Dent. 2022;128(6):1158–1164. [DOI] [PubMed] [Google Scholar]
- 25. Palin CL, Huryn JM, Golden M, Booth PR, Randazzo JD. Three‐dimensional printed definitive cast for a silicone obturator prosthesis: a clinical report. J Prosthet Dent. 2019;121(2):353–357. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 26. Shahid O, Alhayek A, Ahmed ZU, Aslam N, Aldawood T, Morgano SM, et al. Maxillary interim obturator prosthesis fabrication for a patient with limited mouth opening with a digital approach: a clinical report. J Prosthodont. 2024;33(8):725–729. [DOI] [PubMed] [Google Scholar]
- 27. Michelinakis G. The use of cone beam computed tomography and three dimensional printing technology in the restoration of a maxillectomy patient using a dental implant retained obturator. J Indian Prosthodont Soc. 2017;17(4):406–411. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 28. Alfaraj A, Nagai T, AlQallaf H, Lin WS. Race to the moon or the bottom? Applications, performance, and ethical considerations of artificial intelligence in prosthodontics and implant dentistry. Dent J (Basel). 2025;13(1):13. 10.3390/dj13010013 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 29. Fang T, Gao J, Wu Q, Xie C, Zhang Y, Lu J, et al. Digital workflow of labial guides for alveolar ridge reduction during implant‐supported full‐arch rehabilitation. J Prosthodont. 2024;33(5):506–511. [DOI] [PubMed] [Google Scholar]
- 30. Amin SA, Hann S, Elsheikh AK, Boltchi F, Zandinejad A. A complete digital approach for facially generated full arch diagnostic wax up, guided surgery, and implant‐supported interim prosthesis by integrating 3D facial scanning, intraoral scan and CBCT. J Prosthodont. 2023;32(1):90–93. [DOI] [PubMed] [Google Scholar]
- 31. Fathi A, Nadian F, Ghorbani M, Razavi P, Mosharraf R, Ebadian B. Enhancing oral function: a case report on mandibular overdenture utilization with custom‐made subperiosteal implant. J Prosthodont. 2024;33(9):835–840. [DOI] [PubMed] [Google Scholar]
- 32. Pozza MB, Costa AJM, Burgoa S, Ventura D, Cortes ARG. Digital workflow for low‐cost 3D printed custom healing abutment based on emergence profile CBCT segmentation. J Prosthet Dent. 2024;132(5):863–866. [DOI] [PubMed] [Google Scholar]
- 33. Ranefall P, Wählby C. Global gray‐level thresholding based on object size. Cytometry A. 2016;89(4):385–390. [DOI] [PubMed] [Google Scholar]
- 34. Mohan G, Subashini MM. Chapter 4 – Medical imaging with intelligent systems: a review. In: Sangaiah AK, editor. Deep learning and parallel computing environment for bioengineering systems. Cambridge, MA: Academic Press; 2019. p. 53–73. [Google Scholar]
- 35. Starmans MPA, van der Voort SR, Castillo Tovar JM, Veenland JF, Klein S, Niessen WJ. Chapter 18 – Radiomics: data mining using quantitative medical image features. In: Zhou SK, Rueckert D, Fichtinger G, editors. Handbook of medical image computing and computer assisted intervention. Cambridge, MA: Academic Press; 2020. p. 429–456. [Google Scholar]
- 36. Kunapuli SS, Bhallamudi PC. Chapter 22 – A review of deep learning models for medical diagnosis. In: Kumar P, Kumar Y, Tawhid MA, editors. Machine learning, big data, and IoT for medical informatics. Cambridge, MA: Academic Press; 2021. p. 389–404. [Google Scholar]
- 37. Yang WF, Su YX. Artificial intelligence‐enabled automatic segmentation of skull CT facilitates computer‐assisted craniomaxillofacial surgery. Oral Oncol. 2021;118:105360. [DOI] [PubMed] [Google Scholar]
- 38. Kargilis DC, Xu W, Reddy S, Ramesh SSK, Wang S, Le AD, et al. Deep learning segmentation of mandible with lower dentition from cone beam CT. Oral Radiol. 2025 Jan;41(1):1–9. [DOI] [PubMed] [Google Scholar]
- 39. Lo Giudice A, Ronsivalle V, Spampinato C, Leonardi R. Fully automatic segmentation of the mandible based on convolutional neural networks (CNNs). Orthod Craniofac Res. 2021;24(Suppl 2):100–107. [DOI] [PubMed] [Google Scholar]
- 40. Verhelst PJ, Smolders A, Beznik T, Meewis J, Vandemeulebroucke A, Shaheen E, et al. Layered deep learning for automatic mandibular segmentation in cone‐beam computed tomography. J Dent. 2021;114:103786. [DOI] [PubMed] [Google Scholar]
- 41. Palkovics D, Molnar B, Pinter C, García‐Mato D, Diaz‐Pinto A, Windisch P, et al. Automatic deep learning segmentation of mandibular periodontal bone topography on cone‐beam computed tomography images. J Dent. 2025;159:105813. [DOI] [PubMed] [Google Scholar]
- 42. Alahmari M, Alahmari M, Almuaddi A, Abdelmagyd H, Rao K, Hamdoon Z, et al. Accuracy of artificial intelligence‐based segmentation in maxillofacial structures: a systematic review. BMC Oral Health. 2025;25(1):350. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 43. Alrashed S, Dutra V, Chu TG, Yang CC, Lin WS. Influence of exposure protocol, voxel size, and artifact removal algorithm on the trueness of segmentation utilizing an artificial‐intelligence‐based system. J Prosthodont. 2024;33(6):574–583. [DOI] [PubMed] [Google Scholar]
- 44. Melerowitz L, Sreenivasa S, Nachbar M, Stsefanenka A, Beck M, Senger C, et al. Design and evaluation of a deep learning‐based automatic segmentation of maxillary and mandibular substructures using a 3D U‐Net. Clin Transl Radiat Oncol. 2024;47:100780. 10.1016/j.ctro.2024.100780 [DOI] [PMC free article] [PubMed] [Google Scholar]
