Skip to main content
PeerJ Computer Science logoLink to PeerJ Computer Science
. 2021 Sep 13;7:e620. doi: 10.7717/peerj-cs.620

Descriptive analysis of dental X-ray images using various practical methods: A review

Anuj Kumar 1,, Harvendra Singh Bhadauria 1, Annapurna Singh 1
Editor: Thippa Reddy Gadekallu
PMCID: PMC8459782  PMID: 34616881

Abstract

In dentistry, practitioners interpret various dental X-ray imaging modalities to identify tooth-related problems, abnormalities, or teeth structure changes. Another aspect of dental imaging is that it can be helpful in the field of biometrics. Human dental image analysis is a challenging and time-consuming process due to the unspecified and uneven structures of various teeth, and hence the manual investigation of dental abnormalities is at par excellence. However, automation in the domain of dental image segmentation and examination is essentially the need of the hour in order to ensure error-free diagnosis and better treatment planning. In this article, we have provided a comprehensive survey of dental image segmentation and analysis by investigating more than 130 research works conducted through various dental imaging modalities, such as various modes of X-ray, CT (Computed Tomography), CBCT (Cone Beam Computed Tomography), etc. Overall state-of-the-art research works have been classified into three major categories, i.e., image processing, machine learning, and deep learning approaches, and their respective advantages and limitations are identified and discussed. The survey presents extensive details of the state-of-the-art methods, including image modalities, pre-processing applied for image enhancement, performance measures, and datasets utilized.

Keywords: Dental X-ray, Machine learning, Deep learning, Convolutional neural networks, Dental image segmentation

Introduction

Dental X-ray imaging (DXRI) has been developed as the foundation for dental professionals across the world because of the assistance provided in detecting the abnormalities present in the teeth structures (Oprea et al., 2008). For dentists, radiography imparts a significant role in assisting imaging assessment in providing a thorough clinical diagnosis and dental structures preventive examinations (Molteni, 1993). However, to analyze a dental X-ray image, researchers primarily use image processing methods to extract the relevant information. Image segmentation is the most widely used image-processing technique to analyze medical images and help improve computer-aided medical diagnosis systems (Li et al., 2006; Shah et al., 2006).

Furthermore, manual examination of a large collection of X-ray images can be time-consuming because visual inspection and tooth structure analysis have an abysmal sensitive rate; therefore, human screening may not identify a high proportion of caries (Olsen et al., 2009). In most cases, the automatic computerized tool that can help the investigation process would be highly beneficial (Abdi, Kasaei & Mehdizadeh, 2015; Jain & Chauhan, 2017). Dental image examination involved various stages consisting of image enhancement, segmentation, feature extractions, and identification of regions, which are subsequently valuable for detecting cavities, tooth fractures, cysts or tumors, root canal length, and teeth growth in children (Kutsch, 2011; Purnama et al., 2015). Also, various studies revealed that analysis of dental imaging modalities is beneficial in applications like human identification, age estimation, and biometrics (Nomir & Abdel-Mottaleb, 2007; Caruso, Silvestri & Sconfienza, 2013).

At present, deep learning (DL) and machine learning (ML) techniques have gained huge momentum in the field of DXRI analysis. Deep learning frameworks, well-known as convolutional neural networks (CNNs), are primarily employed for processing large and complex image datasets because they can obtain multiple features from obfuscated layers (Schmidhuber, 2015; Hwang et al., 2019). Many studies that used pre-trained networks like Alexnet, VGG, GoogLeNet, and Inception v3 found that they performed well in general. On the other hand, CNN networks tend to develop from shallow layer networks to broader or problem-specific self-made or complicated networks.

Recently, numerous machine learning approaches have been proposed by researchers to improve dental image segmentation and analysis performance. Deep learning and artificial intelligence techniques are remarkably successful in addressing the challenging segmentation dilemmas presented in various studies (Hatvani et al., 2018; Lee et al., 2018a; Yang et al., 2018; Hwang et al., 2019; Khanagar et al., 2021), so we can foresee a whirlwind of inventiveness and lines of findings in the coming years, based on achievements that recommend machine learning models concerning semiotic segmentation for DXRI.

In the existing surveys (Rad et al., 2013; Schwendicke et al., 2019), various techniques and methods have been discussed for DXRI. In Rad et al. (2013), segmentation techniques are divided into three classes: pixel‑based, edge‑based, and region‑based, and further classified into thresholding, clustering boundary-based, region-based, or watershed approaches. However, there is no discussion on enhancement techniques, image databases used, and modalities used for DXRI. Furthermore, after the Rad et al. (2013) survey, a large number of approaches have been introduced by researchers. Next, a review of dental image diagnosis using convolution neural network is presented by Schwendicke et al. (2019), focusing on diagnostic accuracy studies that pitted a CNN against a reference test, primarily on routine imagery data. It has been observed that in the previous surveys, a thorough investigation of traditional image processing, machine learning, and deep learning approaches is missing.

Being an emerging and promising research domain, dental X-ray imaging requires a comprehensive and detailed survey of dental image segmentation and analysis to diagnose and treat various dental diseases. In this study, we have made the following contributions that are missing in the previous surveys: First, we have imparted various studies from 2004 to 2020 covering more than 130 articles and is almost double than previous surveys given by Rad et al. (2013) and Schwendicke et al. (2019). Second, we have presented X-ray pre-processing techniques, traditional image analysis approaches, machine learning, and deep learning advancements in DXRI. Third, specific image modality (such as periapical, panoramic, bitewing and CBCT, etc.) based methods are categorized. At last, performance metrics and dataset descriptions are investigated up to a great extent. Also, specific benchmarks in the advancement of DXRI methods are represented in Fig. 1.

Figure 1. Selected benchmarks at various years for dental imaging methods.

Figure 1

A brief about dental imaging modalities

Dental imaging modalities give insights into teeth growth, bone structures, soft tissues, tooth loss, decay and also helps in root canal treatment (RCT), which is not visible during a dentist’s clinical inspection. Dental imaging modalities are mainly categorized as intra-oral and extra-oral X-rays. In dentistry, these images are frequently used for medical diagnosis (Abrahams, 2001; Caruso, Silvestri & Sconfienza, 2013). Various dental imaging modalities categorization based on intra-oral and extra-oral are presented in Fig. 2.

Figure 2. Overview of dental imaging modalities.

Figure 2

Dental radiographs can discover problems in the mouth, jaws, teeth, bone loss, fractures, cysts at an early stage. X-rays can help in finding issues that can not be visualized with an oral assessment. Identifying and diagnosing problems at the earliest stage can save you from root canal treatment and other serious issues.

Types of dental radiography

Intra-oral radiography. An X-ray film is kept in the mouth to capture the X-ray picture, which comprises all the specific details about teeth arrangement, root canal infection, and identifying caries. Categories of intra-oral X-ray images are:

  • Periapical images. It provides information of root and surrounding bone areas containing three to four teeth in the single X-ray image.

  • Bitewing images. It generally helps in detecting the information of upper and lower tooth arrangements, and an X-ray beam shows the dentist how these teeth are arranged with one another and how to spot a cavity between teeth. Bitewing X-rays may also be used to ensure that a crown is fitted correctly (a tooth-enclosing cap) or tooth restoration is done accurately. It can also detect rotting or damaged fillings.

  • Occlusal images. Occlusal X-rays provide insight into the mouth’s base, revealing the upper or lower jaw’s bite. They place a strong emphasis on children’s tooth development and placement.

Extra-oral radiography. An X-ray picture is taken from outside the mouth to capture the entire skull and jaws region. Extra-oral X-rays are classified into many types.

  • Panoramic X-rays. X-rays are full-sized and capture the overall tooth structure. Also, the pictures provide information about the skull and jaw. These images are mainly used to examine fractures, trauma, jaws diseases, pathological lesions and evaluate the impacted teeth.

  • Cephalometric X-rays. Also called ceph X-ray, it depicts the jaw’s whole part, including the head’s entire side. It is employed in both dentistry and medicine for diagnosis and clinical preparation purposes.

  • Sialogram. It uses a substance that is infused into the salivary glands to make them visible on X-ray film. Doctors may recommend this check to ensure problems with the salivary glands, such as infections or Sjogren’s syndrome signs (a symptom condition identified by sore mouth and eyes; this condition may cause tooth decay).

  • Computed tomography (CT). It is an imaging technique that gives insights into 3-D internal structures. This kind of visualization is used to identify maladies such as cysts, cancers, and fractures in the face’s bones.

  • Cone-beam computed tomography (CBCT) generates precise and high-quality pictures. Cone beam CT is an X-ray type that generates 3D visions of dental formations, soft tissues, nerves, and bones. It helps in guiding the tooth implants and finding cyst and tumefaction in the mouth. It can also find issues in the gum areas, roots, and jaws structures. Cone beam CT is identical to standard dental CT in several respects.

In this study, various articles considered in which the researchers proposed techniques that are extensively applied to periapical, bitewing, panoramic, CT, CBCT, and photographic color images. Digital X-ray imaging is currently gaining traction as a new research area with expanding applications in various fields.

Challenges faced by doctors in analyzing dental X-ray images

Dental practitioners used X-ray radiographs to examine dental anatomy and to determine the care strategy for the patient. Because of a lack of resources, X-ray interpretations rely more on the doctor’s expertise, and manual examination is complex in dentistry (Wang et al., 2016). Therefore, computer-aided systems are introduced to reduce complexity and make the identification process easy and fast. Computer-aided systems are becoming more powerful and intelligent for identifying abnormalities after processing medical images (such as X-rays, Microscopic images, Ultrasound images, and MRI images). Healthcare decision support systems were developed to provide technical guidance to clinical decision-making experts in the healthcare field (Mendonça, 2004). These systems help identify and treat the earliest symptom of demineralization of tooth caries, root canal, and periodontal diseases.

This paper explores the potential computational methods used for developing computer-aided systems, identifies the challenges faced in their implementation, and provides future directions (Amer & Aqel, 2015; Wang et al., 2016). Automatic detection of abnormalities, anomalies, and abrupt changes in teeth structures is a big challenge for researchers. In this study, some of the tooth-related problems are imparted, which is still a challenge for researchers to develop expert systems. We have worked with some of the dental practitioners to understand the common problems. These problems are significantly related to cavities (or caries), root canal treatment (RCT), cysts, teeth implants, and teeth growth. Working in collaboration with dentists helps computer science professionals and researchers to design & develop models that can solve dentist’s problems during examination.

The dental X-ray image analysis methods can be categorized in several categories: region growing techniques, edge detection methods, thresholding based, clustering techniques, level set, and active contour, etc., are presented in ‘Image processing methods for dental image analysis’ (Mahoor & Abdel-Mottaleb, 2004; Zhou & Abdel-Mottaleb, 2005; Nomir & Abdel-Mottaleb, 2005, 2007; Gao & Chae, 2008; Oprea et al., 2008; Patanachai, Covavisaruch & Sinthanayothin, 2010; Harandi & Pourghassem, 2011; Hu et al., 2014; Amer & Aqel, 2015; Zak et al., 2017; Avuçlu & Bacsçiftçi, 2020) (Rad et al., 2015; Tuan, Ngan & Son, 2016; Poonsri et al., 2016; Son & Tuan, 2016, 2017; Ali et al., 2018; Alsmadi, 2018; Obuchowicz Rafałand Nurzynska et al., 2018; Tuan et al., 2018; Fariza et al., 2019; Kumar, Bhadauria & Singh, 2020).

Conventional machine learning methods considering: back propagation neural network (BPNN), artificial neural network (ANN), support vector machine (SVM), Random forest regression-voting constrained local model (RFRV-CLM), Hybrid learning algorithms are presented in ‘Conventional machine learning algorithms for dental image analysis’ (Nassar & Ammar, 2007; Fernandez & Chang, 2012; Pushparaj et al., 2013; Prakash, Gowsika & Sathiyapriya, 2015; Bo et al., 2017; Yilmaz, Kayikcioglu & Kayipmaz, 2017; Vila-Blanco, Tomás & Carreira, 2018). Also, considering Deep learning architectures, i.e., Conventional CNN and transfer learning, GoogLeNet Inception v3, AlexNet, Mask R-CNN model, ResNet-101, six-Layer DCNN, U-net architecture, and LightNet and MatConvNet, etc., are highlighted in ‘Deep learning techniques for dental image analysis’ (Imangaliyev et al., 2016; Miki et al., 2017b, 2017a; Oktay, 2017; Prajapati, Nagaraj & Mitra, 2017; Rana et al., 2017; Srivastava et al., 2017; Chu et al., 2018; Lee et al., 2018a, 2019; Egger et al., 2018; Torosdagli et al., 2018; Yang et al., 2018; Zhang et al., 2018; Hatvani et al., 2018; Jader et al., 2018; Karimian et al., 2018; Kim et al., 2019; Murata et al., 2019; Tuzoff et al., 2019; Fukuda et al., 2019; Hiraiwa et al., 2019; Banar et al., 2020; Singh & Sehgal, 2020; Geetha, Aprameya & Hinduja, 2020).

Contribution

DXRI analysis is an evolving and prospective research field, but still, there is a lack of systematic study available except for one or two studies. In this study, we have made significant contributions as follows:

  1. A comprehensive survey consisting of more than 130 articles related to dental imaging techniques for the last 15 years is presented.

  2. Overall state-of-the-art research works have been classified into three major categories, i.e., image processing, machine learning, and deep learning approaches, and their respective advantages and limitations are identified and discussed.

  3. A comprehensive review of dental imaging methods provided in terms of various performance metrics.

  4. At last, a review of dental X-ray imaging datasets used for implementation and generation.

The rest of the review is structured as follows. The methodology is discussed in ‘Methodology’. Various performance metrics are presented in ‘Performance Measures’. DXRI datasets are given in ‘Dataset Description’. At last, the conclusion is given in ‘Conclusion’.

Methodology

In this survey, 130 research articles from 2004 to 2020 have been reviewed, as shown in Fig. 3, covering almost all research articles from different online digital libraries like Springer, Elsevier, IEEE, and Google Scholar. These articles are conferences, Book chapters, peer-reviewed and reputed journals in computer science and digital dental imaging. A total number of articles deliberating various imaging modalities: Periapical, Bitewing, Panoramic, Hybrid, CT or CBCT, Photographic color teeth images, and undefined datasets are given in Table 1. Methods are categorized as image processing techniques in ‘Image processing methods for dental image analysis’, conventional machine learning methods are given in ‘Conventional machine learning algorithms for dental image analysis’, and deep learning approaches are provided in ‘Deep learning techniques for dental image analysis’. Also, methods are characterized based on imaging modalities (Periapical X-rays, Bitewing X-rays, Panoramic X-rays, CBCT or CT images, etc.), and DXRI methods taxonomy is given in Fig. 4.

Figure 3. Number of research articles as per publication years in DXRI.

Figure 3

Table 1. Number of articles categorized based on imaging modalities.

Image modalities Number of articles published
Periapical X-ray images 30
Bitewing X-ray images 11
Panoramic X-ray images 39
CBCT or CT images 13
Photographic color images 06
Hybrid dataset 19
Image dataset not defined 07

Figure 4. Proposed taxonomy of DXRI methods.

Figure 4

The research incorporated in this comprehensive review primarily focused on medical image processing and artificial intelligence for the detection and examination of the tooth cavity, periodontal disease recognition, tooth arrangement and numbering, root canal detection, periapical lesions detection, salivary gland disease diagnosis, cyst detection, osteoporosis detection, the progress of deciduous teeth, analysis of cephalometric landmarks and fracture identification, etc.

Image processing methods for dental image analysis

The research adopts various image processing strategies for dental imaging to investigate the structures of teeth, caries, and abnormalities to help dental practitioners for the appropriate diagnosis. It involves various pre-processing, segmentation, and classification approaches to make an automatic dental identification system that makes doctor’s work more accessible, unambiguous, and faster. A simple traditional model used for dental image processing is given in Fig. 5.

Figure 5. Traditional model used for dental image segmentation and classification.

Figure 5

Pre-processing techniques

Dental imaging consists of different image modalities, where X-rays are the most common medical imaging method used to classify bone and hard tissues. In dentistry, imaging modalities help identify fractures, teeth structures, jaws alignment, cyst, and bone loss, which has become tremendously popular in dental imaging (Goyal, Agrawal & Sohi, 2018). Noise level, artifacts, and image contrast are vital values that control an image's overall quality. The image quality obtained depends on varying factors such as the dynamic range of the sensors, the lighting conditions, distortion, and the artifact examined (Sarage & Jambhorkar, 2012). Interpretation of a low-resolution image is often a complex and time-consuming process. Pre-processing techniques enhance the quality of low-resolution images, which corrects the spatial resolution and local adjustment to improve the input image’s overall quality (Hossain, Alsharif & Yamashita, 2010). Moreover, enhancement and filtering methods improve the overall image quality parameters before further processing. In Table 2, pre-processing techniques are addressed to recuperate the quality of dental images.

Table 2. Pre-processing methods used for dental imaging modality.
Author & Year Enhancement/Noise removal technique
Methods used for Bitewing X-ray
(Lai & Lin, 2008) Adaptive local contrast stretching is used to make the tooth region smoother after that, adaptive morphological enhancement is applied to improve the texture values.
(Prajapati, Desai & Modi, 2012) A median filter is used to eradicate picture impulse noise.
(Mahoor & Abdel-Mottaleb, 2004; Zhou & Abdel-Mottaleb, 2005; Huang et al., 2012) Top hat and bottom hat filters are applied where the teeth become brightened, and the bone and shadow regions obscured.
(Pushparaj, Gurunathan & Arumugam, 2013) Butterworth high pass filter used with a homomorphic filter. In which homomorphic filter compensates the effect of non-uniform illumination.
Methods used for Periapical X-ray
(Harandi & Pourghassem, 2011) Histogram equalization and noise reduction using wavelets, and also make use of spatial filters like Laplacian filter.
(Lin, Huang & Huang, 2012) Average filter with 25 * 25 mask then histogram equalization is used.
(Nuansanong, Kiattisin & Leelasantitham, 2014) Gaussian spatial filter with kernel size 5 * 5 and sigma value 1.4 is fixed.
(Lin et al., 2014) Enhancement is done by combining adaptive power law transformation, local singularity, and bilateral filter.
(Rad et al., 2015) Median filtering is applied to enhance the images
(Purnama et al., 2015) Contrast stretching used to improve the X-ray quality so that it can be easily interpreted and examined correctly
(Jain & Chauhan, 2017) Gaussian filtering employed to make a more smoothed gradient nearby the edges also helps in reducing noise.
(Obuchowicz Rafałand Nurzynska et al., 2018) Histogram equalization (HEQ) and a statistical dominance algorithm (SDA) are initiated.
(Singh & Agarwal, 2018) Median filtering is used to lower noise, and an unsharp marking filter is used to enhance the high-frequency component.
(Datta, Chaki & Modak, 2019) Local averaging is used to eliminate noisy features.
(Kumar, Bhadauria & Singh, 2020) The guided filter is applied with a window size of 3 * 3 and is cast-off towards calculating output pixel size.
Methods used for Panoramic X-rays
(Frejlichowski & Wanat, 2011) Some basic filters are added to select pyramid layers, including sharpening filter and contrast adjustment before image recomposition.
(Vijayakumari et al., 2012) Block analysis and contrast stretching applied.
(Pushparaj et al., 2013) A combination of the Butterworth bandpass filter and the homomorphic filter is used to enhance the edges and illumination.
(Razali et al., 2014) Canny edge detection is applied, where the gaussian filter is used to eliminate the noise.
(Banu et al., 2014) Image inverse and contrast stretching procedures have been used to identify the region of interest.
(Amer & Aqel, 2015) Contrast enhancement with intensity transformations is used to improve the segmentation procedure.
(Poonsri et al., 2016) Image enhancement using adaptive thresholding (Bradley & Roth, 2007).
(Veena Divya, Jatti & Revan Joshi, 2016) The image contrast is balanced to enhance the picture’s appearance and to visualize the cyst or tumor.
(Zak et al., 2017) A combination of top hat/bottom hat filter and adaptive power-law transformation(APLT) is used to enhance images.
(Alsmadi, 2018) Speckle noise is reduced by using a median filter.
(Divya et al., 2019) Negative transformation applied and caries identified by using the difference of contrast improved Image and image negative.
(Banday & Mir, 2019) Adaptive histogram equalization and median filtering are combinedly applied.
(Fariza et al., 2019) Dental X-ray image is processed using CLAHE, and gamma correction is done to improve the contrast.
(Avuçlu & Bacsçiftçi, 2020) Median softening filter applied after contrast stretching.
Methods used for hybrid dataset
(Said et al., 2006) Internal noise is reduced by closing top-hat transformation, which is described by subtracting the picture from its morphological closure.
(Tuan, Ngan & Son, 2016) Background noise is minimized using a Gaussian filter; then, a Gaussian(DoG) filter is used to measure the gradient along the x and y-axis.
Methods used for color images
(Ghaedi et al., 2014) A contrast enhancement focused on the histogram is introduced to the gray-level Image.
(Datta & Chaki, 2015a) Denoising is done by using a wiener filter.
(Datta & Chaki, 2015b) A Wiener filter is applied to eliminate the blurring effect and additive noise.
(Berdouses et al., 2015) Gray level transformation performed.
Methods used for CBCT & CT
(Benyó et al., 2009) Image with high-frequency noise are enhanced by applying a median filter
(Ji, Ong & Foong, 2014) Initially, the intensity range was adjusted, followed by Gaussian filtering with a standard deviation to suppress noise.

Contrast stretching, Grayscale stretching, Log transformation, Gamma correction, Image negative, and Histogram equalization methods are standard enhancement methods to improve the quality of medical images. X-rays are typically grayscale pictures with high noise rates and low resolution. Thus, the image contrast and boundary representation are relatively weak and small (Ramani, Vanitha & Valarmathy, 2013). Extracting features from these X-rays is quite a difficult task with very minimal details and a low-quality image. By adding specific contrast enhancement techniques significantly improves image quality. So that segmentation and extraction of features from such images can be performed more accurately and conveniently (Kushol et al., 2019). Therefore, a contrast stretching approach has been widely used to enhance digital X-rays quality (Lai & Lin, 2008; Vijayakumari et al., 2012; Berdouses et al., 2015; Purnama et al., 2015; Avuçlu & Bacsçiftçi, 2020). Adaptive local contrast stretching makes use of local homogeneity to solve the problem of over and under enhancement. One of the prominent methods to refine the contrast of the image is histogram equalization (HE) (Harandi & Pourghassem, 2011; Menon & Rajeshwari, 2016; Obuchowicz Rafałand Nurzynska et al., 2018; Banday & Mir, 2019). HE is the way of extending the dynamic range of an image histogram and it also causes unrealistic impacts in images; however, it is very effective for scientific pictures i.e., satellite images, computed tomography, or X-rays. A downside of the approach is its indiscriminate existence. This can increase ambient noise contrast while reducing the useful quality features of an image.

On the other hand, filtering methods applied to medical images help to eradicate the noise up to some extent. Gaussian, Poisson, and Quantum noise are different types of noise artifacts usually found in X-Rays & CTs, particularly when the image is captured (Razifar et al., 2005; Goyal, Agrawal & Sohi, 2018). The noise-free images achieve the efficiency to get the best result and improve the test’s precision. If we try to minimize one class of noise, it may disrupt the other. Various filters have been used to achieve the best potential outcome for the irregularities present in dental images like Average filter, Bilateral filter, Laplacian filter, Homomorphic filter, and Butterworth filter, Median Gaussian filter, and Weiner filter. In recent studies, various filtering techniques used by researchers but widely used filtering methods are Gaussian filter and the median filter, which shows the best result (Benyó et al., 2009; Prajapati, Desai & Modi, 2012; Nuansanong, Kiattisin & Leelasantitham, 2014; Razali et al., 2014; Datta & Chaki, 2015a,b; Rad et al., 2015; Tuan, Ngan & Son, 2016; Jain & Chauhan, 2017; Alsmadi, 2018). However, the drawback of the median filter is that it degrades the boundary details. Whereas the Gaussian filter performs best in peak detection, the limitation is that it reduces the picture’s information.

Dental image segmentation approaches used for different imaging modalities

DXRI segmentation is an essential step to extract valuable information from various imaging modalities. In dentistry, segmentation faces more difficulties than other medical imaging modalities, making the segmentation process more complicated or challenging. Here, the problems faced by researchers in analyzing dental X-ray images and the purpose of segmentation are given in Fig. 6. The segmentation process refers to the localization of artifacts or the boundary tracing, analysis of structure, etc. Human eyes quickly distinguish objects of interest and remove them from the background tissues, but it is a great challenge in developing algorithms.

Figure 6. Purpose of segmentation & problems in dental imaging.

Figure 6

Furthermore, image segmentation has applications distinct from computer vision; it is often used to extract or exclude different portions of an image. General dental image segmentation methods are categorized as thresholding-based, contour or snake models, level set methods, clustering, and region growing (Rad et al., 2013). Moreover, there has been a significant number of surveys presented by various authors (Rad et al., 2013; Sharma, Rana & Kundra, 2015). However, none of them categorized the methods based on dental imaging modalities. Various segmentation and classification techniques are discussed and reviewed in this article, considering multiple dental imaging modalities. In the field of dental imaging, the choice of selecting a correct algorithm for the particular image dataset is most important. This study explores image processing techniques explicitly applied for dental imaging modalities, as given in Table 3.

Table 3. Review findings of the image processing techniques using different imaging modalities.
Author & Year Relevant review findings Total images Detection/Identification
Imaging modality: Bitewing X-rays
(Mahoor & Abdel-Mottaleb, 2004) For segmentation, adaptive thresholding methods is being used, then features are extracted, and teeth numbering is done using the Bayesian classification technique. 50 Teeth numbering
(Zhou & Abdel-Mottaleb, 2005) Proposed segmentation using a window-based adaptive thresholding scheme and minimum Hausdorff distance used for matching purposes. Training =102 images
Testing = 40 images
Human identification
(Nomir & Abdel-Mottaleb, 2005) Results are improved by using a signature vector in conjunction with adaptive and iterative thresholding. 117 Human identification
(Nomir & Abdel-Mottaleb, 2007) Iterative followed by adaptive thresholding used for the segmentation and features extracted using fourier descriptors after forcefield transformation then matching is done by using euclidian distance 162 Human identification
(Lai & Lin, 2008) The B-spline curve is used to extract intensity and texture characteristics for K-means clustering to locate the bones and teeth contour. N.A Teeth detection
(Nomir & Abdel-Mottaleb, 2008) The procedure starts with an iterative process guided by adaptive thresholding. Finally, the Bayesian framework is employed for tooth matching. 187 Human identification
(Harandi, Pourghassem & Mahmoodian, 2011) An active geodesic contour is employed for upper and lower jaws segmentation. 14 Jaw identification
(Huang et al., 2012) An adaptive windowing scheme with isolation-curve verification is used to detect missing tooth regions. 60 Missing teeth detection
(Prajapati, Desai & Modi, 2012) A region growing technique is applied to the X-rays to extract the tooth; then, the content-based image retrieval (CBIR) technique is used for matching purposes. 30 Human identification
(Pushparaj, Gurunathan & Arumugam, 2013) The tooth area's shape is extracted using contour-based connected component labeling, and the Mahalanobis distance (MD) is measured for matching. 50 Person identification
Imaging modality: Periapical X-rays
(Huang & Hsu, 2008) Binary image transformations, thresholding, quartering, characterization, and labeling were all used as part of the process. 420 Teeth detection
(Oprea et al., 2008) Simple thresholding technique applied for segmentation of caries. N.A Caries detection
(Harandi & Pourghassem, 2011) Otsu thresholding method with canny edge detection is used to segment the root canal area. 43 Root canal detection
(Lin, Huang & Huang, 2012) The lesion is detected using a variational level set method after applying otsu’s method. 6 Lesion detection
(Sattar & Karray, 2012) Phase congruency based approach is used to provide a framework for local image structure + edge detection N.A Teeth detection
(Niroshika, Meegama & Fernando, 2013) Deformation and re-parameterize are added to the contour to detect the tooth comer points. N.A Teeth detection
(Ayuningtiyas et al., 2013) Dentin and pulp are separated using active contour, and qualitative analysis is conducted using the dentist’s visual inspection, while quantitative testing is done by measuring different statistic parameters. N.A Tooth detection
(Nuansanong, Kiattisin & Leelasantitham, 2014) Canny edge detection was initially used, followed by an active contour model with data mining (J48 tree) and integration with the competence path. Approx. 50 Tooth detection
(Lin et al., 2014) The otsu’s threshold and connected component analysis are used to precisely segment the teeth from alveolar bones and remove false teeth areas. 28 Teeth detection
(Purnama et al., 2015) For root canal segmentation, an active shape model and thinning (using a hit-and-miss transform) were used. 7 Root canal detection
(Rad et al., 2015) The segmentation is initially done using K-means clustering. Then, using a gray-level co-occurrence matrix, characteristics were extracted from the X-rays. 32 Caries detection
(Jain & Chauhan, 2017) First, all parameter values defined in the snake model then initial contour points initializes, and at last canny edge detection extract the affected part. N.A Cyst detection
(Singh & Agarwal, 2018) The color to mark the carious lesion is provided by the contrast limited adaptive histogram (CLAHE) technique combined with masking. 23 Caries detection
(Rad et al., 2018) The level set segmentation process (LS) is used in two stages. The first stage is the initial contour creation to create the most appropriate IC, and the second stage is the artificial neural network-based smart level approach. 120 Caries detection
(Obuchowicz Rafałand Nurzynska et al., 2018) K-means clustering applied considering intensity values and first-order features (FOF) to detect the caries spots. 10 Caries detection
(Devi, Banumathi & Ulaganathan, 2019) The hybrid algorithm is applied using isophote curvature and the fast marching method (FMM) to extract the cyst. 3 Cyst detection
(Datta, Chaki & Modak, 2019) The geodesic active contour method is applied to identify the dental caries lesion. 120 Caries detection
(Osterloh & Viriri, 2019) It uses unsupervised model to extract the caries region. Jaws partition is done using thresholding and an integral projection algorithm. The top and bottom hats, as well as active contours, were used to detect caries borders. N.A Caries detection
(Kumar, Bhadauria & Singh, 2020) The various dental structures were separated using the fuzzy C-means algorithm and the hyperbolic tangent gaussian kernel function. 152 Dental structures
(Datta, Chaki & Modak, 2020) This method converts the X-ray image data into its neutrosophic analog domain. A custom feature called 'weight' is used for neutrosophication. Contrary to popular belief, this feature is determined by merging other features. 120 Caries detection
Imaging Modality: Panoramic X-rays
(Patanachai, Covavisaruch & Sinthanayothin, 2010) The wavelet transform, thresholding segmentation, and adaptive thresholding segmentation are all compared. Where, the results of wavelet transform show better accuracy as compare to others. N.A Teeth detection
(Frejlichowski & Wanat, 2011) An automatic human identification system applies a horizontal integral projection to segment the individual tooth in this approach. 218 Human identification
(Vijayakumari et al., 2012) A gray level co-occurrence matrix is used to detect the cyst. 3 Cyst detection
(Pushparaj et al., 2013) Horizontal integral projection with a B-spline curve is employed to separate maxilla and mandible N.A Teeth numbering
(Lira et al., 2014) Supervised learning used for segmentation and feature extraction is carried out through computing moments and statistical characteristics. At last, the bayesian classifier is used to identify different classes. 1 Teeth detection
(Banu et al., 2014) The gray level co-occurrence matrix is used to compute texture characteristics (GLCM) and classification results obtained in the feature space, focusing on the centroid and K-mean classifier. 23 Cyst detection
(Razali et al., 2014) This study aims to compare the edge segmentation methods: Canny and Sobel on X-ray images. N.A Teeth detection
(Amer & Aqel, 2015) The segmentation process uses the global Otsu’s thresholding technique with linked component labeling. The ROI extraction and post-processing are completed at the end. 1 Wisdom teeth detection
(Abdi, Kasaei & Mehdizadeh, 2015) Four stages used for segmentation: Gap valley extraction, canny edge with morphological operators, contour tracing, and template matching. 95 Mandible detection
(Veena Divya, Jatti & Revan Joshi, 2016) Active contour or snake model used to detect the cyst boundary. 10 Cyst detection
(Poonsri et al., 2016) Teeth identification, template matching using correlation, and area segmentation using K-means clustering are used. 25 Teeth detection
(Zak et al., 2017) Individual arc teeth segmentation (IATS) with adaptive thresholding is applied to find the palatal bone. 94 Teeth detection
(Alsmadi, 2018) In panoramic X-ray images that can help in diagnosing jaw lesions, the fuzzy C-means concept and the neutrosophic technique are combinedly used to segment jaw pictures and locate the jaw lesion region. 60 Lesion detection
(Dibeh, Hilal & Charara, 2018) The methods use a shape-free layout fitted into a 9-degree polynomial curve to segment the area between the maxillary and mandibular jaws. 62 Jaw separation + teeth detection
(Mahdi & Kobashi, 2018) Quantum Particle Swarm Optimization (QPSO) is employed for multilevel thresholding. 12 Teeth detection
(Ali et al., 2018) A new clustering method based on the neutrosophic orthogonal matrix is presented to help in the extraction of teeth and jaws areas from panoramic X-rays. 66 Teeth detection
(Divya et al., 2019) Textural details extracted using GLCM to classify the cyst and caries. 10 Dental caries & cyst extraction
(Banday & Mir, 2019) Edge detection method for the segmentation then, the Autoregression(AR) model is adopted, and AR coefficients are derived from the feature vector. At last, matching is performed using euclidean distance. 210 Human identification
(Fariza et al., 2019) For tooth segmentation, the Gaussian kernel-based conditional spatial fuzzy c-means (GK-csFCM) clustering algorithm is used. 10 Teeth detection
(Aliaga et al., 2020) The region of interest is extracted from the entire X-ray image, and segmentation is performed using k-means clustering. 370 Osteoporosis detection, mandible detection
(Avuçlu & Bacsçiftçi, 2020) The Image is converted to binary using Otsu's thresholding, and then a canny edge detector is used to find the object of interest. 1,315 Determination of age and gender
Imaging modality: Hybrid dataset images
(Said et al., 2006) Thresholding with mathematical morphology is performed for the segmentation. A total of 500 Bitewing & 130 Periapical images. Teeth detection
(Li et al., 2006) The fast and accurate segmentation approach used strongly focused on mathematical morphology and shape analysis. A total of 500 (Bitewing and Periapical images) Person identification
(Al-sherif, Guo & Ammar, 2012) A two-phase threshold processing is used, starting with an iterative threshold followed by an adaptive threshold to binarize teeth images after separating the individual tooth using the seam carving method. A total of 500 Bitewing & 130 Periapical images Teeth detection
(Ali, Ejbali & Zaied, 2015) The Chan-vese model and an active contour without edges are used to divide an image into two regions with piece-constant intensities. N.A Teeth detection
(Son & Tuan, 2016) The otsu threshold procedure, fuzzy C-means, and semi-supervised fuzzy clustering are all part of a collaborative framework (eSFCM). A total of eight & 56 Image dataset (Bitewing + Panoramic) Teeth structures
(Tuan, Ngan & Son, 2016) It uses a semi-supervised fuzzy clustering algorithm – SSFC-FS based on the Interactive Fuzzy Satisficing method. A total of 56 (Periapical & Panoramic) Teeth structures
(Son & Tuan, 2017) Semi-supervised fuzzy clustering algorithm combined with spatial constraints (SSFC-SC) for dental image segmentation. A total of 56 (Periapical & panoramic images) Teeth structures
(Tuan et al., 2018) Graph-based clustering algorithm called enhanced affinity propagation clustering (APC) used for classification process and fuzzy aggregation operators used for disease detection. A total of 87 (Periapical & Panoramic) Disease detection
Imaging modality: Photographic color images
(Ghaedi et al., 2014) Segmentation functions in two ways. In the first step, the tooth surface is partitioned using a region-widening approach and the Circular Hough Transform (CHT). The second stage uses morphology operators to quantify texture to define the abnormal areas of the tooth's boundaries. Finally, a random forest classifies the various classes. 88 Caries detection
(Datta & Chaki, 2015a) The method uses a biometrics dental technique using RGB images. Segment individual teeth with water Shed and Snake’s help, then afterward incisors teeth features are obtained to identify the human. A total of 270 images dataset Person identification
(Datta & Chaki, 2015b) The proposed method introduces a method for filtering optical teeth images and extracting caries lesions followed by cluster-based Segmentation. 45 Caries detection
(Berdouses et al., 2015) The proposed scheme included two processes: (a) identification, in which regions of interest (pre-cavitated and cavitated occlusal lesions) were partitioned, and (b) classification, in which the identified zones were categorized into one of the seven ICDAS classes. 103 Caries detection
Imaging modality: CT & CBCT
(Gao & Chae, 2008) The multi-step procedure using thresholding, dilation, connected component labeling, upper-lower jaw separation, and last arch curve fitting was used to find the tooth region. N.A Teeth detection
(Hosntalab et al., 2010) Otsu thresholding, morphological operations, and panoramic re-sampling, and variational level set were used. Following that, feature extraction with a wavelet-Fourier descriptor (WFD) and a centroid distance signature is accomplished. Finally, multilayer perceptron (MLP), Kohonen self-organizing network, and hybrid structure are used for Classification. 30 Multislice CT image (MSCT) dataset consists of 804 teeth Teeth detection and Classification
(Gao & Chae, 2010) An adaptive active contour tracking algorithm is used. In which the root is tracked using a single level set technique. In addition, the variational level was increased in several ways. A total of 18 CT images Teeth detection
(Mortaheb, Rezaeian & Soltanian-Zadeh, 2013) Mean shift algorithm is used for CBCT segmentation with new feature space and is compared to thresholding, watershed, level set, and active contour techniques. A total of two CBCT images Teeth detection
(Gao & Li, 2013) The volume data are initially divided into homogeneous blocks and then iteratively merged to produce the initial labeled and unlabeled instances for semi-supervised study. N.A Teeth detection
(Ji, Ong & Foong, 2014) The study adds a new level set procedure for extracting the contour of the anterior teeth. Additionally, the proposed method integrates the objective functions of existing level set methods with a twofold intensity model. A total of ten CBCT images Teeth structure
(Hu et al., 2014) Otsu and mean thresholding technique combinedly used to improve the segmentation. Image dataset consists of 300 layers Teeth detection

Bitewing X-rays are widely used by researchers for the application of human identification and biometrics. Human identification is achieved by applying adaptive thresholding, iterative thresholding, and region-growing approaches. Afterwards, image features are extracted to archive and retrieve dental images used for human identification (Mahoor & Abdel-Mottaleb, 2004, 2005; Nomir & Abdel-Mottaleb, 2005, 2007, 2008; Zhou & Abdel-Mottaleb, 2005). In Huang et al. (2012), missing tooth locations were detected with an adaptive windowing scheme combined with the isolation curve method, which shows the accuracy rate higher than (Nomir & Abdel-Mottaleb, 2005). In Pushparaj, Gurunathan & Arumugam (2013), primarily aimed at estimating the shape of the entire tooth. In which segmentation is performed by applying horizontal and vertical integral projection. In addition, teeth boundary was estimated using the fast connected component labeling algorithm, and lastly, Mahalanobis distance is measured for the matching.

Periapical X-rays help in clinical diagnosis considering dental caries and root canal regions by applying various image processing techniques (Oprea et al., 2008). Many times dentists use periapical X-ray images to spot caries lesions from dental X-rays. Regardless of human brain vision, it is often hard to correctly identify caries by manually examining the X-ray image. Caries detection methods for periapical X-rays have been used iteratively to isolate the initially suspected areas. Then, separated regions are subsequently analyzed. In Rad et al. (2015), automatic caries was identified by applying segmentation using k-means clustering and feature detection using GLCM. However, it shows image quality issues in some cases, and because of these issues, tooth detection may give a false result. On the other hand, (Singh & Agarwal, 2018) applied color masking techniques to mark the curios lesions to find the percentage value of the affected area.

Another approach is given by (Osterloh & Viriri, 2019) mainly focused on upper and lower jaws separation with the help of thresholding and integral projection, and the learning model is employed to extract caries. This model shows better accuracy than (Dykstra, 2008; Tracy et al., 2011; Valizadeh et al., 2015). In Obuchowicz Rafałand Nurzynska et al. (2018), k-means clustering (CLU) and first-order features (FOF) were used to show the best performance for the identification of caries. However, this approach was applied to the dataset of 10 patients with confirmed caries. A geodesic contour technique (Datta, Chaki & Modak, 2019) shows better computational time results than multilevel thresholding, watershed, and level set. The limitation of this approach is that it does not work well for poor-quality pictures, which leads to inappropriate feature extraction. In Datta, Chaki & Modak (2020), a method reduced the computational efforts and caries region identified in optimum time. The X-ray image is processed in the neutrosophic domain to identify the suspicious part, and an active contour method is employed to detect the outer line of the carious part. The benefit of this method is that it prevents recursive iterations using neutrosophication during suspicious area detection.

The semi-automatic method for root canal length detection is proposed by Harandi & Pourghassem (2011) and Purnama et al. (2015) to help dental practitioners properly treat root canal treatment (RCT). In some studies, periapical X-rays are also used for the automatic segmentation of cysts or abscesses. Devi, Banumathi & Ulaganathan (2019) proposed a fully automated hybrid method that combined feature-base isophote curvature and model-based fast marching (FMM). It shows good accuracy and optimum results as compared to Jain & Chauhan (2017). Furthermore, various approaches were used to automatically detect teeth structures (Huang & Hsu, 2008; Sattar & Karray, 2012; Niroshika, Meegama & Fernando, 2013; Nuansanong, Kiattisin & Leelasantitham, 2014; Kumar, Bhadauria & Singh, 2020).

Panoramic X-rays help identify jaw fractures, the structure of jaws, and deciduous teeth. These X-rays are less detailed as compared to periapical and bitewing. It has been observed that the segmentation of panoramic X-rays using wavelet transformation shows better results than adaptive and iterative thresholding (Patanachai, Covavisaruch & Sinthanayothin, 2010). Another fully automatic segmentation of the teeth using the template matching technique introduced by Poonsri et al. (2016) shows 50% matching accuracy results. In Razali et al. (2014) analyzed X-rays for the age estimations by comparing edge detection approaches. Amer & Aqel (2015) have suggested a method used to extract wisdom teeth using the Otsu’s threshold combined with morphological dilation. Then, jaws and teeth regions are extracted using connected component labeling.

In Mahdi & Kobashi (2018), it sets a multi-threshold by applying quantum particle swarm optimization to improve the accuracy. Fariza et al. (2019) employed a method to extract dentin, enamel, pulp, and other surrounding dental structures using conditional spatial fuzzy C-means clustering. Subsequently, the performance improved as compared to inherently used FCM approaches. Dibeh, Hilal & Charara (2018) separates maxillary and mandibular jaws using N-degree polynomial regression. In Abdi, Kasaei & Mehdizadeh (2015), a four-step method is proposed: gap valley extraction, modified canny edge detector, guided iterative contour tracing, and template matching. However, estimating the overall performance of automated segmentation with individual results, all of which were estimated to be above 98%, clearly demonstrates that the computerized process can still be improved to meet the gold standard more precisely.

In Veena Divya, Jatti & Revan Joshi (2016), active contour-based segmentation is proposed for cystic lesion segmentation and extraction to analyze cyst development behavior. The segmentation method has positive results for nonlinear background, poor contrast, and noisy image. Divya et al. (2019) has compared the level set method and watershed segmentation to detect cysts and lesions. The study reveals that the level set segmentation produces more predicted results for cyst/Lesion. An approach used to identify age & gender by analyzing dental images is very useful in biometrics (Avuçlu & Bacsçiftçi, 2020). Several other image processing techniques are used on dental images to achieve the best biometric results.

Hybrid-dataset is the image dataset combining different dental imaging modalities used for the analysis. Said et al. (2006) have used periapical & bitewing X-rays for the teeth segmentation. In this approach, the background area is discarded using an appropriate threshold, then mathematical morphology and connected component labeling are applied for the teeth extraction. This approach finds difficulty in extracting images having low contrast between teeth and bones, blurred images, etc. Another approach introduced by Tuan, Ngan & Son (2016), Son & Tuan (2017), Tuan et al. (2018) the semi-supervised fuzzy clustering method with some modification to find the various teeth and bone structures. Ali, Ejbali & Zaied (2015) compared CPU & GPU results after applying the Chan-Vese model with active contour without edge. It shows that GPU model implementation is several times faster than the CPU version.

Photographic color images are the RGB images of occlusal surfaces that are mainly useful for detecting caries and human identification (Datta & Chaki, 2015a, 2015b). Teeth segmentation is performed by integrating watershed and snake-based techniques on dental RGB images. Subsequently, incisors tooth features extracted for the recognition of a person. This method can segment individual teeth, lesions from caries and track the development of lesion size. This research’s primary objective is to identify the caries lesions of the tooth surfaces, which benefits to improve the diagnosis. In Ghaedi et al. (2014), caries segmentation was employed using the region-widening method and circular hough transform (CHT), then morphological operations applied to locate the unstable regions around the tooth boundaries. Another fully automatic approach for the caries classification is given by Berdouses et al. (2015), where segmentation separates caries lesion then after area features are extracted to assign the region to a particular class. It can be a valuable method to support the dentist in making more reliable and accurate detection and analysis of occlusal caries.

CT & CBCT Images provide 3D visualization of teeth and assist dental practitioners in orthodontic surgery, dental implants, and cosmetic surgeries. Hosntalab et al. (2010) recommended a multi-step procedure for labeling and classification in CT images. However, teeth segmentation is performed by employing global thresholding, morphological operations, region growing, and variational level sets. Another approach, a multi-step procedure, was introduced by Mortaheb, Rezaeian & Soltanian-Zadeh (2013) based on the mean shift algorithm for CT image segmentation of the tooth area, which results best as compare with watershed, thresholding, active contour. Another technique that does not depend on mean shift is suggested by Gao & Li (2013), which uses an iterative scheme to label events for the segmentation. Furthermore, segmentation methods are improved by applying active contour tracking algorithms and level set methods (Gao & Chae, 2010). It shows higher accuracy and visualization of tooth regions as compared to other methods.

Conventional machine learning algorithms for dental image analysis

Development in the field of Machine Learning (ML) and Artificial Intelligence (AI) is growing over the last few years. ML and AI methods have made a meaningful contribution to the field of dental imaging, such as computer-aided diagnosis & treatment, X-ray image interpretation, image-guided treatment, infected area detection, and information representation adequately and efficiently. The ML and AI make it easier and help doctors diagnose and presume disease risk accurately and more quickly in time. Conventional machine learning algorithms for image perception rely exclusively on expertly designed features, i.e., identifying dental caries involves extracting texture features—an overview of various machine learning algorithms is given in Fig. 7.

Figure 7. Overview of machine learning algorithms.

Figure 7

ML datasets are generally composed of exclusive training, validation, and test sets. It determines system characteristics by evaluating and testing the dataset then validates the features acquired from the input data. Using the test dataset, one might finally verify ML’s precision and extract valuable features to formulate a powerful training model. Table 4 reveals the conventional machine-learning algorithms used for dental X-ray imaging.

Table 4. The table shows relevant review findings of conventional machine learning algorithms for different imaging modalities.

Author & Year Relevant review findings Images Feature classifier Detection
Imaging modality: Periapical X-rays
(Li et al., 2005, 2007) To segment the dental Image into normal, abnormal, and potentially abnormal areas, the variational level set function is used. 60 X-rays Trained SVM is used to characterize the normal and abnormal regions after Segmentation. Bone loss & root decay
Imaging modality: Panoramic X-rays
(Pushparaj et al., 2013) The geometrical features are used to classify both premolar and molar teeth, while for tooth numbering, the matching templates method is used effectively. N.A Feature extraction (Projected principal edge distribution (PPED) + Geometric properties + Region descriptors) + SVM Teeth numbering and Classification are used to help Forensic odontologists.
(Sornam & Prabhakaran, 2017) The Linearly Adaptive Particle Swarm algorithm is developed and implemented to improve the accuracy rate of the neural system classifier. N.A Back Propagation Neural Network (BPNN) and Linearly Adaptive Particle Swarm Optimization (LA-PSO) Caries detection
(Bo et al., 2017) A two-stage SVM model was proposed for the Classification of osteoporosis. Dataset consists of 40 images HOG (histogram of oriented gradients + SVM Osteoporosis detection
(Vila-Blanco, Tomás & Carreira, 2018) Segmentation of mandibular teeth carried out by applying Random forest regression-voting constrained local model (RFRV-CLM) in two steps: The 1st step gives an estimate of individual teeth and mandible regions used to initialize search for the tooth. In the second step, the investigation is carried out separately for each tooth. Training images: 261
Testing images: 85
(RFRV-CLMs) Adult age teeth detection or a missing tooth for person identification.
Imaging Modality: Photographic color images
(Fernandez & Chang, 2012) Teeth segmentation and classification of teeth palate using ANN gives better results as compared to SVM. It shows that ANN is seven-times faster than SVM in terms of time N.A ANN + Multilayer perceptrons trained with the error back-propagation algorithm. Oral infecto-contagious diseases,
(Prakash, Gowsika & Sathiyapriya, 2015) The prognosticating faults method includes the following stages: pre-processing, Segmentation, features extraction, SVM classification, and prediction of diseases. N.A Adaptive threshold + Unsupervised SVM classifier Dental defect prediction
Imaging modality: CBCT or CT
(Yilmaz, Kayikcioglu & Kayipmaz, 2017) Classifier efficiency improved by using the forward feature selection algorithm to reduce the size of the feature vector. The SVM classifier performs best in classifying periapical cyst and keratocystic odontogenic tumor (KCOT) lesions. A total of 50 CBCT 3D scans Order statistics (median, standard deviation, skewness, kurtosis, entropy) and 3D Haralick Features + SVM Periapical cyst and keratocystic odontogenic tumor
Imaging modality: Hybrid dataset images
(Nassar & Ammar, 2007) A hybrid learning algorithm is used to evaluate the binary bayesian classification filters’ metrics and the class-conditional intensities. Bitewing & Periapical films Feature extraction + Bayesian classification. Teeth are matching for forensic odontology.
(Avuçlu & Bacsçiftçi, 2019) First, image pre-processing and segmentation are applied to extract the features and quantitative information obtained from the feature extraction from teeth images. Subsequently, features are taken as input to the multilayer perceptron neural network. A total of 1,315 Dental X-ray images,162 different age groups Otsu thresholding + Feature extraction (average absolute deviation) + Multilayer perceptron neural network Age and gender classification

Deep learning techniques for dental image analysis

Artificial intelligence, machine learning, and deep learning approaches assist medical imaging technicians in spotting abnormalities and diagnosing disorders in a fraction of the time required earlier (and with more accurate tests generally). Deep learning (DL) is an improvement of artificial neural networks (ANN), which has more layers and allows for more accurate data predictions (LeCun, Bengio & Hinton, 2015; Schmidhuber, 2015). Deep learning is associated with developing self-learning back-propagation techniques that incrementally optimize data outcomes and increase computing power. Deep learning is a rapidly developing field with numerous applications in the healthcare sector. The number of available, high-quality datasets in ML and DL applications plays a significant role in evaluating the outcome accuracy. Also, information fusion assists in integrating multiple datasets and their use of DL models to enhance accuracy parameters. The predictive performance of deep learning algorithms in the medical imaging field exceeds human skill levels, transforming the role of computer-assisted diagnosis into a more interactive one (Burt et al., 2018; Park & Park, 2018).

Health diagnostic computer-aided software is used in the medical field as a secondary tool, but developing traditional CAD systems tend to be very strenuous. Recently, there have been introducing deep learning approaches to CAD, with accurate outcomes for different clinical applications (Cheng et al., 2016). The research study mostly used a convolution neural network model to analyze other dental imaging modalities. CNN’s are a typical form of deep neural network feed-forward architectures, and they are usually used for computer vision and image object identification tasks. CNN's were initially released about two decades back; however, in 2012, AlexNet’s architecture outpaced added ImageNet large-scale competition challenges (Krizhevsky, Sutskever & Hinton, 2012). Machine vision came in as the deep learning revolution, and since then, CNNs have been rapidly evolving. Feature learning methods have taken a massive turn since the CNN model has come into the picture. Fully convolution neural network Alexnet architecture is used to categorize teeth, including molar, premolar, canine, and incisor, by training cone-beam CT images (Miki et al., 2017a; Oktay, 2017). Tuzoff et al. (2019) applied the Faster R-CNN model, which interprets pipeline and optimizes computation to detect the tooth (Ren et al., 2017) and VGG-16 convolutional architecture for classification (Simonyan & Zisserman, 2014). These methods are beneficial in practical applications and further investigation of computerized dental X-ray image analysis.

In DXRI, CNNs have been extensively used to detect tooth fractures, bone loss, caries detection, periapical lesions, or also used for the analysis of different dental structures (Lee et al., 2018b; Schwendicke et al., 2019). Neural networks need to be equipped and refined, and X-ray dataset repositories are necessary (Lee et al., 2018a). In Lee et al. (2019), the mask R-CNN model is applied based on a CNN that can identify, classify, and mask artifacts in an image. A mask R-CNN mask operated in two steps. In the first step, the Region of interest (ROIs) selection procedure was performed. Next, the R-CNN mask includes a binary mask similarity to the classification and bounding box foresight for each ROI (Romera-Paredes & Torr, 2016; He et al., 2017).

Dental structures (enamel, dentin, and pulp) identified using U-net architecture show the best outcome (Ronneberger, Fischer & Brox, 2015). CNN is a standard technique for multi-class identification and characterization, but it requires extensive training to achieve a successful result if used explicitly. In the medical sphere, the lack of public data is a general problem because of privacy. To address this issue, (Zhang et al., 2018) suggested a technique that uses a label tree to assign multiple labels to each tooth and decompose a task that can manage data shortages. Table 5 presents various studies considering deep learning-based techniques in the field of dentistry.

Table 5. The table shows relevant review findings of deep learning algorithms for different imaging modalities.

Authors Deep learning architectures Detection/Application Metrics
Imaging modality: Periapical X-rays
(Prajapati, Nagaraj & Mitra, 2017) CNN and transfer learning Dental caries, periapical infection, and periodontitis Accuracy:- 0.8846
(Yang et al., 2018) Conventional CNN Automated clinical diagnosis F1 score 0.749
(Zhang et al., 2018) CNN (label tree with cascade network structure) Teeth detection & classification Precision:- 0.958, Recall:- 0.961
F-score :- 0.959
(Choi, Eun & Kim, 2018) Conventional CNN Caries detection F1max:- 0.74 with FPs:- 0.88
(Lee et al., 2018b) GoogLeNet Inception v3 CNN network Caries and Non-caries Premolar accuracy (premolar):- 0.89, Accuracy (molar):- 0.88, and Accuracy:- 0.82, AUC (premolar):- 0.917, AUC (molar):- 0.890, and an AUC (Both premolar and molar):- 0.845
(Lee et al., 2018a) CNN (VGG-19) Periodontally compromised teeth (PCT) For premolars, the total diagnostic Accuracy(premolars):- 0.810, Accuracy(molars):- 76.7%
(Geetha, Aprameya & Hinduja, 2020) Back‑propagation neural network Caries detection Accuracy:- 0.971, FPR:- 0.028, ROC :- 0.987, PRC :- 0.987 with learning rate:- 0.4, momentum:- 0.2
Imaging modality: Panoramic X-rays
(Oktay, 2017) AlexNet Teeth detection and classification Accuracy (tooth detection):- 0.90
Classification accuracy:
Molar :-0.9432, Premolar:- 0.9174, Canine & Incissor:- 0.9247
(Chu et al., 2018) Deep octuplet Siamese network (OSN) Osteoporosis analysis Accuracy:- 0.898
(Wirtz, Mirashi & Wesarg, 2018) Coupled shape model + neural network Teeth detection Precision:- 0.790, Recall:- 0.827
Dice coefficient:- 0.744
(Jader et al., 2018) Mask R-CNN model Teeth detection Accuracy:- 0.98, F1-score:- 0.88, precision:- 0.94, Recall:- 0.84, and Specificity:- 0.99
(Lee et al., 2019) Mask R-CNN model Teeth segmentation for diagnosis and forensic identification F1 score:- 0.875, Precision:- 0.858, Recall:- 0.893, Mean‘IoU’:- 0.877
(Kim et al., 2019) DeNTNet (deep neural transfer Network) Bone loss detection F1 score:- 0.75,
Accuracy:- 0.69.
(Tuzoff et al., 2019) R-CNN Teeth detection and numbering Tooth detection (Precision:- 0.9945
Sensitivity:- 0.9941)
Tooth numbering (Specificity:- 0.9994, Sensitivity = 0.9800)
(Fukuda et al., 2019) DetectNet with DIGITS version 5.0 Vertical root fracture Recall:- 0.75, Precision:- 0.93
F-measure:- 0.83
(Murata et al., 2019) AlexNet Maxillary sinusitis Accuracy:- 0.875, Sensitivity:- 0.867, Specificity:- 0.883, and AUC:- 0.875.
(Kats et al., 2019) ResNet-101 Plaque detection Sensitivity:- 0.75, Specificity:- 0.80,
Accuracy:- 0.83, AUC:- 0.5
(Singh & Sehgal, 2020) 6-Layer DCNN Classification of molar, premolar, canine and incisor Accuracy (augmented database):- 0.95, Accuracy (original database):- 0.92
(Muramatsu et al., 2020) CNN (Resnet 50) Teeth detection and classification Tooth detection sensitivity:- 0.964
Average classification accuracy (single model):- 0.872, (multisized models):- 0.932
(Banar et al., 2020) Conventional CNN Teeth detection Dice score:- 0.93, accuracy:- 0.54, a MAE:- 0.69, and a linear weighted Cohen’s kappa coefficient:- 0.79.
Imaging modality: Bitewing X-rays
(Srivastava et al., 2017) Fully convolutional neural network FCNN Detection of dental caries Recall:- 80.5, Precision:- 61.5,
F-score:- 70.0
Imaging modality: CT & CBCT
(Miki et al., 2017a) AlexNet architecture A total of seven-Tooth-type classification (canine, molar, premolar, etc.) Accuracy:- 0.91
(Miki et al., 2017b) AlexNet Teeth detection and classification Detection accuracy:- 0.774,
Classification accuracy:- 0.771
(Hatvani et al., 2018) Subpixel network + U-Net architecture Teeth resolution enhancement Mean of difference (area mm2):- 0.0327
Mean of difference(micrometer):- 114.26
Dice coefficient:- 0.9101
(Torosdagli et al., 2018) CNN (a long short-term memory (LSTM) network) Anatomical Landmarking DSC:- 0.9382
(Egger et al., 2018) CNN (VGG16, FCN) Mandible detection Accuracy:- 0.9877, Dice coefficient:- 0.8964 and Standard deviation:- 0.0169
(Hiraiwa et al., 2019) AlexNet and GoogleNet Classification of root morphology (Single or extra) Diagnostic accuracy:- 0.869
Imaging modality: Hybrid dataset
(Wang et al., 2016) U-net architecture (Ronneberger, Fischer & Brox, 2015) Landmark detection in cephalometric radiographs and Dental structure in bitewing radiographs. F-score => 0.7
(Lee, Park & Kim, 2017) LightNet and MatConvNet Landmark detection N.A
(Karimian et al., 2018) Conventional CNN Caries detection Sensitivity:- 97.93%~99.85%
Specificity:- 100%
Imaging modality: Color images/Oral images
(Rana et al., 2017) Conventional CNN Detection of inflamed and healthy gingiva precision:- 0.347, Recall: 0.621, AUC:- 0.746
Image type not defined
(Imangaliyev et al., 2016) Conventional CNN Dental plaque F1-score:- 0.75

Challenges and future directions

After reviewing various works focusing on traditional image processing techniques, it has been perceived that researchers faced multiple challenges in the field of DXRI segmentation and analysis, such as intensity variation in the X-ray images, poor image quality due to noise, irregular shape of an object, limitations of capturing devices, proper selection of methodology and lack of availability of datasets. Also, experience severe challenges in automatically detecting abnormalities, root canal infection, and sudden changes in the oral cavity. Since there are different varieties of dental X-ray images, it is hard to find a particular segmentation approach; it all depends on the precise condition of the X-rays. Some articles have used pre-processed digital X-rays that were manually cropped to include the area of interest. Because of inconsistencies in the manual method, it is hard to accurately interpret and compare outcomes (Lee, Park & Kim, 2017).

Moreover, convolutional neural networks (and their derivatives) are performing outstandingly in dental X-ray image analysis. One notable conclusion is that many researchers use almost the same architectures, the same kind of network, but have very different outcomes. Deep neural networks are most successful when dealing with a large training dataset, but large datasets are not publically available in the DXRI and are not annotated. If vast publicly accessible dental X-ray image datasets were constructed, our research community would undoubtedly benefit exceedingly.

For the future perspective, the dental X-ray image public repository needs to be developed, and data uniformity is required for deep learning applications in dentistry. Also, DXRI aims to create a classifier that can classify multiple anomalies, caries classes, types of jaw lesions, cyst, root canal infection, etc., in dental images using features derived from the segmentation results. There is also a need to build machine learning-based investigative methods and rigorously validate them with a large number of dental professionals. The participation of specialists in this process will increase the likelihood of growth and development. Currently, there exists no universally acceptable software or tool for dental image analysis. However, such a tool is essentially needed to improve the performance of CAD systems and better treatment planning.

Performance measures

In general, if the algorithm’s efficiency is more significant than other algorithms, one algorithm is prioritized over another. Evaluating the effectiveness of a methodology requires the use of a universally accessible and valid measure. Various performance metrics have been used to compare algorithms or machine learning approaches depending on the domain or study area. It comprises accuracy, Jaccard index, sensitivity, precision, recall, DSC, F-measure, AUC, MSE, error rate, etc. Here, we include a thorough analysis of the success metrics employed in dental image analysis.

Performance metrics used for dental image processing

Calculating performance metrics used for dental segmentation is performed by authenticating pixel by pixel and analyzing the segmentation results with the gold standard. Manual annotation of X-ray images done by a radiologist is considered to be the gold standard. Pixel-based metrics are measured using precision, dice coefficient, accuracy, specificity, and F-score widely used in segmentation analysis. Some of the problems in analyzing image segmentation are metric selection, the use of multiple meanings for some metrics in the literature, and inefficient metric measurement implementations that lead to significant large volume dataset difficulties. Poorly described metrics can result in imprecision conclusions on state-of-the-art algorithms, which affects the system’s overall growth. Table 6 presents an overview of performance metrics widely used by researchers for dental image segmentation and analysis.

Table 6. Performance metrics used by various researchers for the dental image analysis.

Metrics Symbol Author’s
True positive rate (sensitivity, recall) TPR (Hosntalab et al., 2010; Mortaheb, Rezaeian & Soltanian-Zadeh, 2013; Pushparaj et al., 2013; Ghaedi et al., 2014; Abdi, Kasaei & Mehdizadeh, 2015; Berdouses et al., 2015; Datta & Chaki, 2015b; Alsmadi, 2018; Datta, Chaki & Modak, 2019, 2020)
True negative rate (specificity) TNR (Hosntalab et al., 2010; Mortaheb, Rezaeian & Soltanian-Zadeh, 2013; Ghaedi et al., 2014; Abdi, Kasaei & Mehdizadeh, 2015; Berdouses et al., 2015; Datta & Chaki, 2015b; Alsmadi, 2018; Datta, Chaki & Modak, 2019)
Positive predictive value (precision) PPV (Hosntalab et al., 2010; Mortaheb, Rezaeian & Soltanian-Zadeh, 2013; Pushparaj et al., 2013; Berdouses et al., 2015; Datta, Chaki & Modak, 2020)
Jaccard index JAC (Ji, Ong & Foong, 2014)
Dice coefficient DSC (Ji, Ong & Foong, 2014; Abdi, Kasaei & Mehdizadeh, 2015; Datta, Chaki & Modak, 2019; Devi, Banumathi & Ulaganathan, 2019)
F-Measure (F1 Measure = Dice) FMS (Berdouses et al., 2015; Datta, Chaki & Modak, 2020)
Accuracy ACC (Huang & Hsu, 2008; Olsen et al., 2009; Banu et al., 2014; Nuansanong, Kiattisin & Leelasantitham, 2014; Ghaedi et al., 2014; Lin et al., 2014; Datta & Chaki, 2015a,b; Poonsri et al., 2016; Rad et al., 2018; Osterloh & Viriri, 2019; Datta, Chaki & Modak, 2019, 2020; Devi, Banumathi & Ulaganathan, 2019; Kumar, Bhadauria & Singh, 2020)
Mahalanobis distance MHD (Pushparaj, Gurunathan & Arumugam, 2013)
Hausdorff distance HD (Abdi, Kasaei & Mehdizadeh, 2015)
Distance vector DV (Prajapati, Desai & Modi, 2012)
Similarity measure SM (Pushparaj, Gurunathan & Arumugam, 2013; Alsmadi, 2018; Singh & Agarwal, 2018)
The area under ROC curve AUC (Nuansanong, Kiattisin & Leelasantitham, 2014)
Cohens kappa coefficient KAP (Berdouses et al., 2015)
Mean absolute error MAE (Vijayakumari et al., 2012; Amer & Aqel, 2015; Tuan et al., 2018; Kumar, Bhadauria & Singh, 2020)
Mean square error MSE (Vijayakumari et al., 2012; Singh & Agarwal, 2018; Tuan et al., 2018)
Error rate ERR (Zhou & Abdel-Mottaleb, 2005; Nomir & Abdel-Mottaleb, 2008; Hosntalab et al., 2010; Harandi & Pourghassem, 2011; Lira et al., 2014; Datta & Chaki, 2015b; Purnama et al., 2015; Tuan et al., 2018; Banday & Mir, 2019)
Failure rate FR (Said et al., 2006; Al-sherif, Guo & Ammar, 2012)

The significance of accuracy and assurance is essential in the medical imaging field. Also, the validation of segmentation achieves the result and dramatically increases the precision, accuracy, conviction, and computational speed of segmentation. Segmentation methods are especially helpful in computer-aided medical diagnostic applications where the interpretation of objects that are hard to differentiate by human vision is a significant component.

Confusion matrix

The confusion matrix is used to estimate the performance of medical image segmentation and classification. The confusion matrix helps identify the relationship between the outcomes of the predictive algorithm and the actual ones. Some of the terms used for the confusion matrix are given in Table 7: True positive (TP): correctly identified or detected; False positive (FP): evaluated or observed incorrectly; False negative (FN): wrongly rejected; True Negative (TN): correctly rejected. In the approach (Mahoor & Abdel-Mottaleb, 2005), experimental outcomes proved that molar classification is relatively easy compared to premolars, and for teeth classification, centroid distance is less effective than a coordinate signature. Various metrics such as the signature vector, force field (FF), and Fourier descriptor (FD) were used to test the efficiency of the approach given by Nomir & Abdel-Mottaleb (2007), and for matching euclidean distance and absolute distance, FF & FD give small values, suggesting that they performed better than the others. Here, FF & FD give small values for matching Euclidean distance and absolute distance, indicating that the performance is better than the other two methods. In another approach (Prajapati, Desai & Modi, 2012), feature vectors are evaluated and used to find the image distance vector (Dn) using formula: Dn=|TnFVFVQ|, where feature vector (TnFV) is used for database image and (FVQ) is used for the query image. The minimum value of the distance vector indicates the best match of the image with the database image.

Table 7. Confusion matrix.

True positive rate (Recall/Sensitivity): It implies how the caries lesion is accurately detected when it is present there. Sensitivity is expressed as TPTP+FN
True negative rate (Specificity): That is the amount of negative caries lesion examination when there's no affected lesion. Specificity is measured as TNTN+FP
Dice Coefficient: This metric measures between two samples. It is defined as (2|AB|)(|A|+|B|), where |A| and |B| are the number of elements in the sample.
Accuracy: It can be defined as the percentage of correctly classified instances. It is calculated as TP+TNTP+TN+FN+FP100.
Precision: It explains the pureness of our positive detections efficiently compared to the ground truth. It is the positive predictive value defined as TPTP+FP
F-Score: The F-score is a process of combining the model's precision and recall and the harmonic mean of the model's precision and recall. It is expressed as 2×Precision×RecallPrecision+Recall

The study (Huang et al., 2012) shows better isolation precision accuracy for the segmentation of jaws as compared with Nomir and Abdel–Mottaleb. Another method evaluated the complete length of the tooth and capered with the dentist's manual estimation (Harandi & Pourghassem, 2011). Here, measurement error (ME) is evaluated for root canals applying the formula: ME=MesuredlengthActuallength and evaluated ME is lowest for one canal compared to two and three canals.

Niroshika, Meegama & Fernando (2013) traced the tooth boundaries using active contour and distance parameters are compared with the Kass algorithm. The value of the standard distance parameter was found to be lower than that of the Kass algorithm, implying that the proposed method is more efficient for tracing the tooth boundary than the Kass algorithm. Another approach used for counting molar and premolar teeth is considering precision and sensitivity (Pushparaj et al., 2013). Here performance is using metric η is given by:η=(mn)n100. Where ‘m’ represents the total number of teeth counted, and ‘n’ represents the incorrectly numbered teeth. The counting of molar and premolar teeth is more than 90% accurate using this method.

In Abdi, Kasaei & Mehdizadeh (2015), mandible segmentation and Hausdorff distance parameters were compared to the manually annotated gold standard. The algorithm results appear to be very close to the manually segmented gold standard in terms of sensitivity, accuracy, and dice similarity coefficient (DSC). In Amer & Aqel (2015), a wisdom tooth is extracted, and the mean absolute error (MAE) is used to equate the procedure with the other two methods. As compared to other approaches, the lower MAE value showed better segmentation.

In Poonsri et al. (2016), precision is calculated for single-rooted and double-rooted teeth using template matching. According to their study, segmentation accuracy is greater than 40%. Son & Tuan (2016, 2017) used the following cluster validity metrics: PBM, Simplified Silhouette Width Criterion (SSWC), Davis-Bouldin (DB), BH, VCR, BR, and TRA, and the measures of these parameters indicate the best performance as compared with the results of current algorithms.

PBM: The maximum value of this index is said to be the PBM index, across the hierarchy provides the best partitioning.

Simplified Silhouette Width Criterion (SSWC): The silhouette analysis tests how well the observation is clustered and calculates the average distance between clusters. The silhouette plot shows how similar each point in a cluster is to the neighboring clusters’ points.

Davies-Bouldin index (DB): This index determines the average ‘similarity’ amongst clusters, in which the resemblance is a metric that measures the distance between clusters with the size of clusters themselves. The lower Davies-Bouldin index refers to a model with a greater detachment of clusters.

Ball and Hall index (BH): It is used to determine the distance within a group, with a higher value showing better results.

Calinski-Harabasz index, also called Variance Ratio Criterion (VCR): It can be applied to evaluate the partition data by variance, and its higher value indicates good results.

Banfeld-Raftery index (BR): It is evaluated using a variance-covariance matrix for each cluster.

Difference-like index (TRA): It calculates the cluster difference, and a higher value gives the best results.

Comparison of various performance metrics used in dental X-ray imaging considering deep learning methods are given in Fig. 8.

Figure 8. Performances measure comparisons used for deep learning methods.

Figure 8

Dataset description

The researcher in the dental imaging field has used various types of databases. In which some of the databases are available online, while some records are not present. The most prominent dilemma is finding out which investigation has given valid results because everyone has shown promising results on their datasets. All the dental imaging databases that have been used so far are given in Table 8.

Table 8. Dental X-ray image dataset description used for deep learning methods.

Authors & Year Dataset description
(Eun & Kim, 2016) Periapical X-rays: 500 periapical images used for training where each Image is containing five teeth and 100 images used for testing with corresponding ground truth.
(Wang et al., 2016) Total number of patients: 400 (100 additional patients)
Cephalometric radiographs: 400 images .tiff format dimension of 1,935 × 2,400 pixels, 120 bitewing radiographs (new) (Age group 6 to 60 yrs)
Software used: Soredex CRANEXr Excel Ceph machine (Tuusula, Finland) and Soredex SorCom software (3.1.5, version 2.0)
(Prajapati, Nagaraj & Mitra, 2017) Periapical RadioVisio Graphy (RVG) X-ray: 251 images (labeled dataset) where 180 used for training, 26 images for testing & 45 images validation.
(Rana et al., 2017) Color images: Training and validation data consist of 258 images & 147 images.
(Lee, Park & Kim, 2017) A total of 300 Dental X-ray images with resolution 1,935 × 2,400 pixels and 150 images used for training, and 150 images used for testing.
(Srivastava et al., 2017) Bitewing images: More than 3,000 images
(Miki et al., 2017a) CBCT dataset taken from Asahi University Hospital, Gifu, Japan.
Two dental units: Veraviewepocs 3D (J.Morita Mfg, Corp., Kyoto, Japan) and Alphard VEGA (Asahi Roentgen Ind. Co., Ltd., Kyoto, Japan
(Miki et al., 2017b) CT data: 52 images, Training group: 42 images, testing group: 10 images
(Oktay, 2017) Panoramic Images: Dataset taken from 3-different X-ray machines have image dimensions 2871 × 1577, 1435 × 791, or 2612 × 1244 pixels. Testing and validation are done using images of 100 different people.
(Yang et al., 2018) A small dataset of 196 periapical images used, and also augmentation is performed.
(Zhang et al., 2018) Periapical Images: Initially for training, 800 images and 200 used for testing.
and data is annotated with the help of bounding box labels in 32 teeth position.
(Wirtz, Mirashi & Wesarg, 2018) Panoramic X-rays: 14 test images used.
Image augmentation is used to increase training images up to 4000.
(Choi, Eun & Kim, 2018) Periapical X-rays: 475 images dimension of 300 × 413 from 688 × 944 or 1200 × 1650.
(Jader et al., 2018) Panoramic X-ray images:
A total of 193 images used for training containing 6987 teeth and 83 images for validation containing 3,040.
(Lee et al., 2018b) Periapical Images: 3,000 images .jpeg format dimension resized to 299 × 299 pixels The training and validation dataset was 2,400 and a test dataset of 600. The training and validation dataset consisted of 1,200 dental caries and 1,200 non-dental caries, and the test dataset consisted of 300 dental caries and 300 non-dental caries. Augmentation is done up to ten times for training.
(Hatvani et al., 2018) Micro CT images: a training set consists of 5,680 slices and a test set of 1,824 slices was used.
(Torosdagli et al., 2018) CBCT dataset of 50 patients and qualitative visual inspection were done for 250 patients with high variability.
(Karimian et al., 2018) Training is performed using different batches containing ten optical coherence tomography (OCT) images per batch.
(Lee et al., 2018a) Periapical X-ray images resized to 224 × 224 pixels (from the original 1,440 × 1,920 pixels) in .png format : For training (n = 1,044), validation (n = 348), and test (n = 348) datasets.
(Egger et al., 2018) CT dataset containing 45 images as DICOM files with dimension 512 × 512 from a department of craniomaxillofacial surgery in Austria. 1st Image set containing 1,680 slices, 2nd one with induced noise images 6720 slices, 3rd after transformation 13,440 slices, and 4th covered augmentation 18,480 slices
(Chu et al., 2018) Panoramic X-ray: 108 images.
(Hiraiwa et al., 2019) CBCT images and panoramic radiographs used for 760 mandibular first molars (400 patients)
(Lee et al., 2019) Panoramic X-rays: Dimensions of 2,988 × 1,369 pixels.
Total 846 annotated tooth images.
Training group: 30 radiographs, Validation & testing: 20 images.
Augmentation technique used to reduce overfitting ( obtained 1,024 training samples from 846 original data points )
(Kim et al., 2019) Panoramic Images:12,179 images (annotated by experts)
Trained, validated, and tested using 11, 189, 190, and 800.
(Tuzoff et al., 2019) Panoramic radiographs: 1,352 images
Training group: 1,352 images, Testing group: 222 images
(Murata et al., 2019) Panoramic X-rays: Total patients: 100 (50 men and 50 women), Training data for 400 healthy and 400 inflamed maxillary sinuses and data augmentation used to make 6,000 samples
(Kats et al., 2019) Panoramic X-ray:65 images and augmentation performed.
(Fukuda et al., 2019) Panoramic X-ray: 300 images with 900 × 900 pixels.
Training set: 240 images, Testing set: 60 images
(Singh & Sehgal, 2020) Panoramic X-rays: Total 400 images. Training group: 240 images, Testing group: 160 images. Also, augmentation is done by using transformation.
(Muramatsu et al., 2020) Panoramic X-rays: 100 images dimension of 3,000 × 1,500 pixels used for testing and training both.
(Geetha, Aprameya & Hinduja, 2020) Periapical X-rays: 105 images saved as in .bmp format dimension resized to 256 × 256, where caries identified 49 images
Training, validation, and testing consists of 49 caries and 56 sound dental X-ray images.
(Banar et al., 2020) Panoramic (OPGs) image dataset of 400 images used.

Conclusion

Dental X-ray image analysis is a challenging area, and it receives significantly less attention from the community of researchers. There is, however, no systematic review that addresses the state-of-the-art approaches of DXRI. This paper has performed a thorough analysis of more than 130 techniques suggested by different researchers over the last few decades. This study presented a survey of various segmentation and classification techniques widely used for dental X-ray imaging. Methods are characterized as image processing, conventional machine learning, and deep learning. Furthermore, a novel taxonomy mainly focusing on the imaging modalities-based categorization such as bitewing, periapical, panoramic, CBCT/CT, hybrid datasets, and color pictures. Various studies have found that opting for one type of segmentation technique is very difficult in conventional image-processing methods because of image dataset variability. The primary barrier in the growth of a high-performance classification model is the requirement of an annotated datasets, as pointed by various researchers mentioned in this study. Dental X-ray imaging data is not the same as other medical images because of the different image characteristics. This difference has an impact on the deep learning model’s adaptability during image classification. It is also challenging to validate and verify the algorithm’s correctness because of the inadequate datasets available for the hypothesis.

Now we would like to bring the researcher’s attention towards future directions in DXRI. Since most dental X-ray image analysis methods result in decreased efficiency, more sophisticated segmentation techniques should be designed to improve clinical treatment. Further, it is being observed that limited work is employed in the recent studies to detect caries classes such as classes I–VI, and root canal infection. Researchers should therefore focus on implementing new methodologies for caries classification and detection. Recently, deep learning has improved DXRI segmentation and classification performance and requires large annotated image datasets for training, but large annotated X-ray datasets are not publicly accessible. Further, a public repository for dental X-ray images needs to be developed. It is still an open problem so that we can expect new findings and research outcomes in the coming years.

Funding Statement

The authors received no funding for this work.

Additional Information and Declarations

Competing Interests

The authors declare that they have no competing interests.

Author Contributions

Anuj Kumar conceived and designed the experiments, performed the experiments, analyzed the data, performed the computation work, prepared figures and/or tables, authored or reviewed drafts of the paper, and approved the final draft.

Harvendra Singh Bhadauria conceived and designed the experiments, performed the experiments, analyzed the data, performed the computation work, prepared figures and/or tables, authored or reviewed drafts of the paper, and approved the final draft.

Annapurna Singh conceived and designed the experiments, performed the experiments, analyzed the data, performed the computation work, prepared figures and/or tables, authored or reviewed drafts of the paper, and approved the final draft.

Data Availability

The following information was supplied regarding data availability:

This is a Review Article; there are no data or code files.

References

  • Abdi, Kasaei & Mehdizadeh (2015).Abdi AH, Kasaei S, Mehdizadeh M. Automatic segmentation of mandible in panoramic x-ray. Journal of Medical Imaging. 2015;2(4):44003. doi: 10.1117/1.JMI.2.4.044003. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • Abrahams (2001).Abrahams JJ. Dental CT imaging: a look at the jaw. Radiology. 2001;219(2):334–345. doi: 10.1148/radiology.219.2.r01ma33334. [DOI] [PubMed] [Google Scholar]
  • Al-sherif, Guo & Ammar (2012).Al-sherif N, Guo G, Ammar HH. A new approach to teeth segmentation. 2012 IEEE International Symposium on Multimedia; Piscataway: IEEE; 2012. pp. 145–148. [Google Scholar]
  • Ali, Ejbali & Zaied (2015).Ali RB, Ejbali R, Zaied M. GPU-based segmentation of dental x-ray images using active contours without edges. 2015 15th International Conference on Intelligent Systems Design and Applications (ISDA); 2015. pp. 505–510. [Google Scholar]
  • Ali et al. (2018).Ali M, Son LH, Khan M, Tung NT. Segmentation of dental x-ray images in medical imaging using neutrosophic orthogonal matrices. Expert Systems with Applications. 2018;91(6):434–441. doi: 10.1016/j.eswa.2017.09.027. [DOI] [Google Scholar]
  • Aliaga et al. (2020).Aliaga I, Vera V, Vera M, García E, Pedrera M, Pajares G. Automatic computation of mandibular indices in dental panoramic radiographs for early osteoporosis detection. Artificial Intelligence in Medicine. 2020;103:101816. doi: 10.1016/j.artmed.2020.101816. [DOI] [PubMed] [Google Scholar]
  • Alsmadi (2018).Alsmadi MK. A hybrid Fuzzy C-Means and Neutrosophic for jaw lesions segmentation. Ain Shams Engineering Journal. 2018;9(4):697–706. doi: 10.1016/j.asej.2016.03.016. [DOI] [Google Scholar]
  • Amer & Aqel (2015).Amer YY, Aqel MJ. An efficient segmentation algorithm for panoramic dental images. Procedia Computer Science. 2015;65:718–725. doi: 10.1016/j.procs.2015.09.016. [DOI] [Google Scholar]
  • Avuçlu & Bacsçiftçi (2019).Avuçlu E, Bacsçiftçi F. Novel approaches to determine age and gender from dental x-ray images by using multiplayer perceptron neural networks and image processing techniques. Chaos, Solitons & Fractals. 2019;120:127–138. [Google Scholar]
  • Avuçlu & Bacsçiftçi (2020).Avuçlu E, Bacsçiftçi F. The determination of age and gender by implementing new image processing methods and measurements to dental x-ray images. Measurement. 2020;149(2):106985. doi: 10.1016/j.measurement.2019.106985. [DOI] [Google Scholar]
  • Ayuningtiyas et al. (2013).Ayuningtiyas A, Putra NK, Juliastuti E, Epsilawati L. Quantitative image analysis of periapical dental radiography for dental condition diagnosis. 2013 3rd International Conference on Instrumentation, Communications, Information Technology and Biomedical Engineering; 2013. pp. (ICICI–BME). 363–366. [Google Scholar]
  • Banar et al. (2020).Banar N, Bertels J, Laurent F, Boedi RM, De Tobel J, Thevissen P, Vandermeulen D. Towards fully automated third molar development staging in panoramic radiographs. International Journal of Legal Medicine. 2020;134(5):1–11. doi: 10.1007/s00414-020-02283-3. [DOI] [PubMed] [Google Scholar]
  • Banday & Mir (2019).Banday M, Mir AH. Dental biometric identification system using AR model. TENCON 2019–2019 IEEE Region 10 Conference (TENCON); Piscataway: IEEE; 2019. pp. 2363–2369. [Google Scholar]
  • Banu et al. (2014).Banu AFS, Kayalvizhi M, Arumugam B, Gurunathan U. Texture based classification of dental cysts. 2014 International Conference on Control, Instrumentation, Communication and Computational Technologies (ICCICCT); 2014. pp. 1248–1253. [Google Scholar]
  • Benyó et al. (2009).Benyó B, Szilágyi L, Haidegger T, Kovács L, Nagy-Dobó C. Detection of the root canal’s centerline from dental micro-CT records. 2009 Annual International Conference of the IEEE Engineering in Medicine and Biology Society; Piscataway: IEEE; 2009. pp. 3517–3520. [DOI] [PubMed] [Google Scholar]
  • Berdouses et al. (2015).Berdouses ED, Koutsouri GD, Tripoliti EE, Matsopoulos GK, Oulis CJ, Fotiadis DI. A computer-aided automated methodology for the detection and classification of occlusal caries from photographic color images. Computers in Biology and Medicine. 2015;62(9555):119–135. doi: 10.1016/j.compbiomed.2015.04.016. [DOI] [PubMed] [Google Scholar]
  • Bo et al. (2017).Bo C, Liang X, Chu P, Xu J, Wang D, Yang J, Megalooikonomou V, Ling H. Osteoporosis prescreening using dental panoramic radiographs feature analysis. 2017 IEEE 14th International Symposium on Biomedical Imaging (ISBI 2017); Piscataway: IEEE; 2017. pp. 188–191. [Google Scholar]
  • Bradley & Roth (2007).Bradley D, Roth G. Adaptive thresholding using the integral image. Journal of Graphics Tools. 2007;12:13–21. [Google Scholar]
  • Burt et al. (2018).Burt JR, Torosdagli N, Khosravan N, RaviPrakash H, Mortazi A, Tissavirasingham F, Hussein S, Bagci U. Deep learning beyond cats and dogs: recent advances in diagnosing breast cancer with deep neural networks. The British Journal of Radiology. 2018;91:20170545. doi: 10.1259/bjr.20170545. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • Caruso, Silvestri & Sconfienza (2013).Caruso P, Silvestri E, Sconfienza LM. Cone beam CT and 3D imaging. Berlin: Springer Science & Business Media; 2013. [Google Scholar]
  • Cheng et al. (2016).Cheng J-Z, Ni D, Chou Y-H, Qin J, Tiu C-M, Chang Y-C, Huang C-S, Shen D, Chen C-M. Computer-aided diagnosis with deep learning architecture: applications to breast lesions in US images and pulmonary nodules in CT scans. Scientific Reports. 2016;6(1):1–13. doi: 10.1038/srep24454. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • Choi, Eun & Kim (2018).Choi J, Eun H, Kim C. Boosting proximal dental caries detection via combination of variational methods and convolutional neural network. Journal of Signal Processing Systems. 2018;90:87–97. [Google Scholar]
  • Chu et al. (2018).Chu P, Bo C, Liang X, Yang J, Megalooikonomou V, Yang F, Huang B, Li X, Ling H. Using octuplet siamese network for osteoporosis analysis on dental panoramic radiographs. 2018 40th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC); Piscataway: IEEE; 2018. pp. 2579–2582. [DOI] [PubMed] [Google Scholar]
  • Datta & Chaki (2015a).Datta S, Chaki N. Person identification technique using RGB based dental images. IFIP International Conference on Computer Information Systems and Industrial Management; 2015a. pp. 169–180. [Google Scholar]
  • Datta & Chaki (2015b).Datta S, Chaki N. Detection of dental caries lesion at early stage based on image analysis technique. 2015 IEEE International Conference on Computer Graphics, Vision and Information Security (CGVIS); Piscataway: IEEE; 2015b. pp. 89–93. [Google Scholar]
  • Datta, Chaki & Modak (2019).Datta S, Chaki N, Modak B. A novel technique to detect caries lesion using isophote concepts. IRBM. 2019;40(3):174–182. doi: 10.1016/j.irbm.2019.04.001. [DOI] [Google Scholar]
  • Datta, Chaki & Modak (2020).Datta S, Chaki N, Modak B. Neutrosophic set-based caries lesion detection method to avoid perception error. SN Computer Science. 2020;1(1):63. doi: 10.1007/s42979-020-0066-0. [DOI] [Google Scholar]
  • Devi, Banumathi & Ulaganathan (2019).Devi RK, Banumathi A, Ulaganathan G. An automated and hybrid method for cyst segmentation in dental x-ray images. Cluster Computing. 2019;22(S5):12179–12191. doi: 10.1007/s10586-017-1580-2. [DOI] [Google Scholar]
  • Dibeh, Hilal & Charara (2018).Dibeh G, Hilal A, Charara J. A novel approach for dental panoramic radiograph segmentation. 2018 IEEE International Multidisciplinary Conference on Engineering Technology (IMCET); Piscataway: IEEE; 2018. pp. 1–6. [Google Scholar]
  • Divya et al. (2019).Divya KV, Jatti A, Joshi PR, Krishna SD. Progress in Advanced Computing and Intelligent Engineering. Berlin: Springer; 2019. A correlative study of contrary image segmentation methods appending dental panoramic x-ray images to detect jawbone disorders; pp. 25–35. [Google Scholar]
  • Dykstra (2008).Dykstra B. Interproximal caries detection: how good are we? Dentistry Today. 2008;27:144–146. [PubMed] [Google Scholar]
  • Egger et al. (2018).Egger J, Pfarrkirchner B, Gsaxner C, Lindner L, Schmalstieg D, Wallner J. Fully convolutional mandible segmentation on a valid ground-truth dataset. 2018 40th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC); Piscataway: IEEE; 2018. pp. 656–660. [DOI] [PubMed] [Google Scholar]
  • Eun & Kim (2016).Eun H, Kim C. Oriented tooth localization for periapical dental X-ray images via convolutional neural network. 2016 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA); 2016. pp. 1–7. [Google Scholar]
  • Fariza et al. (2019).Fariza A, Arifin AZ, Astuti ER, Kurita T. Segmenting tooth components in dental x-ray images using Gaussian kernel-based conditional spatial Fuzzy C-Means clustering algorithm. International Journal of Intelligent Engineering and Systems. 2019;12(3):108–117. doi: 10.22266/ijies2019.0630.12. [DOI] [Google Scholar]
  • Fernandez & Chang (2012).Fernandez K, Chang C. Teeth/palate and interdental segmentation using artificial neural networks. IAPR Workshop on Artificial Neural Networks in Pattern Recognition; 2012. pp. 175–185. [Google Scholar]
  • Frejlichowski & Wanat (2011).Frejlichowski D, Wanat R. Automatic segmentation of digital orthopantomograms for forensic human identification. International Conference on Image Analysis and Processing; 2011. pp. 294–302. [Google Scholar]
  • Fukuda et al. (2019).Fukuda M, Inamoto K, Shibata N, Ariji Y, Yanashita Y, Kutsuna S, Nakata K, Katsumata A, Fujita H, Ariji E. Evaluation of an artificial intelligence system for detecting vertical root fracture on panoramic radiography. Oral Radiology. 2019;36(4):1–7. doi: 10.1007/s11282-019-00409-x. [DOI] [PubMed] [Google Scholar]
  • Gao & Chae (2008).Gao H, Chae O. Automatic tooth region separation for dental CT images. 2008 Third International Conference on Convergence and Hybrid Information Technology; 2008. pp. 897–901. [Google Scholar]
  • Gao & Chae (2010).Gao H, Chae O. Individual tooth segmentation from CT images using level set method with shape and intensity prior. Pattern Recognition. 2010;43(7):2406–2417. doi: 10.1016/j.patcog.2010.01.010. [DOI] [Google Scholar]
  • Gao & Li (2013).Gao Y, Li X. Teeth segmentation via semi-supervised learning. 2013 6th International Conference on Biomedical Engineering and Informatics; 2013. pp. 558–563. [Google Scholar]
  • Geetha, Aprameya & Hinduja (2020).Geetha V, Aprameya KS, Hinduja DM. Dental caries diagnosis in digital radiographs using back-propagation neural network. Health Information Science and Systems. 2020;8(1):1–14. doi: 10.1007/s13755-019-0096-y. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • Ghaedi et al. (2014).Ghaedi L, Gottlieb R, Sarrett DC, Ismail A, Belle A, Najarian K, Hargraves RH. An automated dental caries detection and scoring system for optical images of tooth occlusal surface. 2014 36th Annual International Conference of the IEEE Engineering in Medicine and Biology Society; Piscataway: IEEE; 2014. pp. 1925–1928. [DOI] [PubMed] [Google Scholar]
  • Goyal, Agrawal & Sohi (2018).Goyal B, Agrawal S, Sohi BS. Noise issues prevailing in various types of medical images. Biomedical & Pharmacology Journal. 2018;11(3):1227–1237. doi: 10.13005/bpj/1484. [DOI] [Google Scholar]
  • Harandi & Pourghassem (2011).Harandi AA, Pourghassem H. A semi automatic algorithm based on morphology features for measuring of root canal length. 2011 IEEE 3rd International Conference on Communication Software and Networks; Piscataway: IEEE; 2011. pp. 260–264. [Google Scholar]
  • Harandi, Pourghassem & Mahmoodian (2011).Harandi AA, Pourghassem H, Mahmoodian H. Upper and lower jaw segmentation in dental X-ray image using modified active contour. 2011 International Conference on Intelligent Computation and Bio-Medical Instrumentation; 2011. pp. 124–127. [Google Scholar]
  • Hatvani et al. (2018).Hatvani J, Horváth A, Michetti J, Basarab A, Kouamé D, Gyöngy M. Deep learning-based super-resolution applied to dental computed tomography. IEEE Transactions on Radiation and Plasma Medical Sciences. 2018;3(2):120–128. doi: 10.1109/TRPMS.2018.2827239. [DOI] [Google Scholar]
  • He et al. (2017).He K, Gkioxari G, Dollár P, Girshick R. Mask r-cnn. Proceedings of the IEEE International Conference on Computer Vision; 2017. pp. 2961–2969. [Google Scholar]
  • Hiraiwa et al. (2019).Hiraiwa T, Ariji Y, Fukuda M, Kise Y, Nakata K, Katsumata A, Fujita H, Ariji E. A deep-learning artificial intelligence system for assessment of root morphology of the mandibular first molar on panoramic radiography. Dentomaxillofacial Radiology. 2019;48(3):20180218. doi: 10.1259/dmfr.20180218. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • Hosntalab et al. (2010).Hosntalab M, Zoroofi RA, Tehrani-Fard AA, Shirani G. Classification and numbering of teeth in multi-slice CT images using wavelet-Fourier descriptor. International Journal of Computer Assisted Radiology and Surgery. 2010;5(3):237–249. doi: 10.1007/s11548-009-0389-8. [DOI] [PubMed] [Google Scholar]
  • Hossain, Alsharif & Yamashita (2010).Hossain MF, Alsharif MR, Yamashita K. Medical image enhancement based on nonlinear technique and logarithmic transform coefficient histogram matching. IEEE/ICME International Conference on Complex Medical Engineering; Piscataway: IEEE; 2010. pp. 58–62. [Google Scholar]
  • Hu et al. (2014).Hu Z, Wu PZ, Gui J, Chen Y, Zheng H. Teeth segmentation using dental CT data. 2014 7th International Conference on Biomedical Engineering and Informatics; 2014. pp. 81–84. [Google Scholar]
  • Huang & Hsu (2008).Huang C-H, Hsu C-Y. Computer-assisted orientation of dental periapical radiographs to the occlusal plane. Oral Surgery, Oral Medicine, Oral Pathology, Oral Radiology, and Endodontology. 2008;105(5):649–653. doi: 10.1016/j.tripleo.2007.08.019. [DOI] [PubMed] [Google Scholar]
  • Huang et al. (2012).Huang P-W, Lin P-L, Kuo C-H, Cho YS. An effective tooth isolation method for bitewing dental x-ray images. 2012 International Conference on Machine Learning and Cybernetics; 2012. pp. 1814–1820. [Google Scholar]
  • Hwang et al. (2019).Hwang J-J, Jung Y-H, Cho B-H, Heo M-S. An overview of deep learning in the field of dentistry. Imaging Science in Dentistry. 2019;49(1):1. doi: 10.5624/isd.2019.49.1.1. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • Imangaliyev et al. (2016).Imangaliyev S, van der Veen MH, Volgenant CMC, Keijser BJF, Crielaard W, Levin E. Deep learning for classification of dental plaque images. International Workshop on Machine Learning, Optimization, and Big Data; 2016. pp. 407–410. [Google Scholar]
  • Jader et al. (2018).Jader G, Fontineli J, Ruiz M, Abdalla K, Pithon M, Oliveira L. Deep instance segmentation of teeth in panoramic x-ray images. 2018 31st SIBGRAPI Conference on Graphics, Patterns and Images (SIBGRAPI); 2018. pp. 400–407. [Google Scholar]
  • Jain & Chauhan (2017).Jain KR, Chauhan NC. An automatic segmentation approach towards the objectification of cyst diagnosis in periapical dental radiograph. International Conference on Information and Communication Technology for Intelligent Systems; 2017. pp. 164–172. [Google Scholar]
  • Ji, Ong & Foong (2014).Ji DX, Ong SH, Foong KWC. A level-set based approach for anterior teeth segmentation in cone beam computed tomography images. Computers in Biology and Medicine. 2014;50:116–128. doi: 10.1016/j.compbiomed.2014.04.006. [DOI] [PubMed] [Google Scholar]
  • Karimian et al. (2018).Karimian N, Salehi HS, Mahdian M, Alnajjar H, Tadinada A. Deep learning classifier with optical coherence tomography images for early dental caries detection. Lasers in Dentistry XXIV; 2018. p. 1047304. [Google Scholar]
  • Kats et al. (2019).Kats L, Vered M, Zlotogorski-Hurvitz A, Harpaz I. Atherosclerotic carotid plaque on panoramic radiographs: neural network detection. International Journal of Computerized Dentistry. 2019;22:163–169. [PubMed] [Google Scholar]
  • Khanagar et al. (2021).Khanagar SB, Al-ehaideb A, Maganur PC, Vishwanathaiah S, Patil S, Baeshen HA, Sarode SC, Bhandi S. Developments, application, and performance of artificial intelligence in dentistry–a systematic review. Journal of Dental Sciences. 2021;16(1):508–522. doi: 10.1016/j.jds.2020.06.019. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • Kim et al. (2019).Kim J, Lee H-S, Song I-S, Jung K-H. DeNTNet: deep neural transfer network for the detection of periodontal bone loss using panoramic dental radiographs. Scientific Reports. 2019;9(1):1–9. doi: 10.1038/s41598-019-53758-2. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • Krizhevsky, Sutskever & Hinton (2012).Krizhevsky A, Sutskever I, Hinton GE. Imagenet classification with deep convolutional neural networks. Advances in Neural Information Processing Systems; 2012. pp. 1097–1105. [Google Scholar]
  • Kumar, Bhadauria & Singh (2020).Kumar A, Bhadauria HS, Singh A. Semi-supervised OTSU based hyperbolic tangent Gaussian kernel fuzzy C-mean clustering for dental radiographs segmentation. Multimedia Tools and Applications. 2020;79(3–4):2745–2768. doi: 10.1007/s11042-019-08268-8. [DOI] [Google Scholar]
  • Kushol et al. (2019).Kushol R, Raihan M, Salekin MS, Rahman ABM. Contrast enhancement of medical x-ray image using morphological operators with optimal structuring element. 2019. https://arxiv.org/abs/1905.08545 https://arxiv.org/abs/1905.08545 ArXiv preprint.
  • Kutsch (2011).Kutsch VK. Caries detection, inside dentistry. Dallas, TX: AEGIS Communications; 2011. [Google Scholar]
  • Lai & Lin (2008).Lai YH, Lin PL. Effective segmentation for dental x-ray images using texture-based fuzzy inference system. International Conference on Advanced Concepts for Intelligent Vision Systems; 2008. pp. 936–947. [Google Scholar]
  • LeCun, Bengio & Hinton (2015).LeCun Y, Bengio Y, Hinton G. Deep learning. Nature. 2015;521(7553):436–444. doi: 10.1038/nature14539. [DOI] [PubMed] [Google Scholar]
  • Lee et al. (2019).Lee J-H, Han S-S, Kim YH, Lee C, Kim I. Application of a fully deep convolutional neural network to the automation of tooth segmentation on panoramic radiographs. Oral Surgery, Oral Medicine, Oral Pathology and Oral Radiology. 2019;129(6):635–642. doi: 10.1016/j.oooo.2019.11.007. [DOI] [PubMed] [Google Scholar]
  • Lee et al. (2018a).Lee J-H, Kim D-H, Jeong S-N, Choi S-H. Detection and diagnosis of dental caries using a deep learning-based convolutional neural network algorithm. Journal of Dentistry. 2018a;77(Suppl. 1):106–111. doi: 10.1016/j.jdent.2018.07.015. [DOI] [PubMed] [Google Scholar]
  • Lee et al. (2018b).Lee J-H, Kim D, Jeong S-N, Choi S-H. Diagnosis and prediction of periodontally compromised teeth using a deep learning-based convolutional neural network algorithm. Journal of Periodontal & Implant Science. 2018b;48(2):114–123. doi: 10.5051/jpis.2018.48.2.114. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • Lee, Park & Kim (2017).Lee H, Park M, Kim J. Cephalometric landmark detection in dental x-ray images using convolutional neural networks. Medical Imaging 2017: Computer-Aided Diagnosis; 2017. p. 101341W. [Google Scholar]
  • Li et al. (2005).Li S, Fevens T, Krzyżak A, Jin C, Li S. Toward automatic computer aided dental X-ray analysis using level set method. International Conference on Medical Image Computing and Computer-Assisted Intervention; 2005. pp. 670–678. [DOI] [PubMed] [Google Scholar]
  • Li et al. (2006).Li S, Fevens T, Krzyżak A, Li S. Automatic clinical image segmentation using pathological modeling, PCA and SVM. Engineering Applications of Artificial Intelligence. 2006;19(4):403–410. doi: 10.1016/j.engappai.2006.01.011. [DOI] [Google Scholar]
  • Li et al. (2007).Li S, Fevens T, Krzyżak A, Jin C, Li S. Semi-automatic computer aided lesion detection in dental X-rays using variational level set. Pattern Recognition. 2007;40:2861–2873. [Google Scholar]
  • Lin, Huang & Huang (2012).Lin P-L, Huang P-Y, Huang P-W. An automatic lesion detection method for dental X-ray images by segmentation using variational level set. 2012 International Conference on Machine Learning and Cybernetics; 2012. pp. 1821–1825. [Google Scholar]
  • Lin et al. (2014).Lin PL, Huang PY, Huang PW, Hsu HC, Chen CC. Teeth segmentation of dental periapical radiographs based on local singularity analysis. Computer Methods and Programs in Biomedicine. 2014;113:433–445. doi: 10.1016/j.cmpb.2013.10.015. [DOI] [PubMed] [Google Scholar]
  • Lira et al. (2014).Lira PHM, Giraldi GA, Neves LAP, Feijoo RA. Dental r-ray image segmentation using texture recognition. IEEE Latin America Transactions. 2014;12:694–698. [Google Scholar]
  • Mahdi & Kobashi (2018).Mahdi FP, Kobashi S. Quantum particle swarm optimization for multilevel thresholding-based image segmentation on dental x-ray images. 2018 Joint 10th International Conference on Soft Computing and Intelligent Systems (SCIS) and 19th International Symposium on Advanced Intelligent Systems (ISIS); 2018. pp. 1148–1153. [Google Scholar]
  • Mahoor & Abdel-Mottaleb (2004).Mahoor MH, Abdel-Mottaleb M. Automatic classification of teeth in bitewing dental images. 2004 International Conference on Image Processing. ICIP’04; 2004. pp. 3475–3478. [Google Scholar]
  • Mahoor & Abdel-Mottaleb (2005).Mahoor MH, Abdel-Mottaleb M. Classification and numbering of teeth in dental bitewing images. Pattern Recognition. 2005;38(4):577–586. doi: 10.1016/j.patcog.2004.08.012. [DOI] [Google Scholar]
  • Mendonça (2004).Mendonça EA. Clinical decision support systems: perspectives in dentistry. Journal of Dental Education. 2004;68:589–597. doi: 10.1002/(ISSN)1930-7837. [DOI] [PubMed] [Google Scholar]
  • Menon & Rajeshwari (2016).Menon HP, Rajeshwari B. Enhancement of dental digital x-ray images based on the image quality. The International Symposium on Intelligent Systems Technologies and Applications; 2016. pp. 33–45. [Google Scholar]
  • Miki et al. (2017a).Miki Y, Muramatsu C, Hayashi T, Zhou X, Hara T, Katsumata A, Fujita H. Classification of teeth in cone-beam CT using deep convolutional neural network. Computers in Biology and Medicine. 2017a;80:24–29. doi: 10.1016/j.compbiomed.2016.11.003. [DOI] [PubMed] [Google Scholar]
  • Miki et al. (2017b).Miki Y, Muramatsu C, Hayashi T, Zhou X, Hara T, Katsumata A, Fujita H. Tooth labeling in cone-beam CT using deep convolutional neural network for forensic identification. Medical Imaging 2017: Computer-Aided Diagnosis; 2017b. p. 101343E. [DOI] [PubMed] [Google Scholar]
  • Molteni (1993).Molteni R. Direct digital dental x-ray imaging with Visualix/VIXA. Oral surgery, Oral Medicine, Oral Pathology. 1993;76(2):235–243. doi: 10.1016/0030-4220(93)90211-L. [DOI] [PubMed] [Google Scholar]
  • Mortaheb, Rezaeian & Soltanian-Zadeh (2013).Mortaheb P, Rezaeian M, Soltanian-Zadeh H. Automatic dental CT image segmentation using mean shift algorithm. 2013 8th Iranian Conference on Machine Vision and Image Processing (MVIP); 2013. pp. 121–126. [Google Scholar]
  • Murata et al. (2019).Murata M, Ariji Y, Ohashi Y, Kawai T, Fukuda M, Funakoshi T, Kise Y, Nozawa M, Katsumata A, Fujita H, Ariji E. Deep-learning classification using convolutional neural network for evaluation of maxillary sinusitis on panoramic radiography. Oral Radiology. 2019;35(3):301–307. doi: 10.1007/s11282-018-0363-7. [DOI] [PubMed] [Google Scholar]
  • Muramatsu et al. (2020).Muramatsu C, Morishita T, Takahashi R, Hayashi T, Nishiyama W, Ariji Y, Zhou X, Hara T, Katsumata A, Ariji E, Fujita H. Tooth detection and classification on panoramic radiographs for automatic dental chart filing: improved classification by multi-sized input data. Oral Radiology. 2020:1–7. doi: 10.1007/s11282-019-00418-w. [DOI] [PubMed] [Google Scholar]
  • Nassar & Ammar (2007).Nassar DEM, Ammar HH. A neural network system for matching dental radiographs. Pattern Recognition. 2007;40(1):65–79. doi: 10.1016/j.patcog.2006.04.046. [DOI] [Google Scholar]
  • Niroshika, Meegama & Fernando (2013).Niroshika UAA, Meegama RGN, Fernando TGI. Active contour model to extract boundaries of teeth in dental x-ray images. 2013 8th International Conference on Computer Science & Education; 2013. pp. 396–401. [Google Scholar]
  • Nomir & Abdel-Mottaleb (2005).Nomir O, Abdel-Mottaleb M. A system for human identification from x-ray dental radiographs. Pattern Recognition. 2005;38(8):1295–1305. doi: 10.1016/j.patcog.2004.12.010. [DOI] [Google Scholar]
  • Nomir & Abdel-Mottaleb (2007).Nomir O, Abdel-Mottaleb M. Human identification from dental x-ray images based on the shape and appearance of the teeth. IEEE Transactions on Information Forensics and Security. 2007;2(2):188–197. doi: 10.1109/TIFS.2007.897245. [DOI] [Google Scholar]
  • Nomir & Abdel-Mottaleb (2008).Nomir O, Abdel-Mottaleb M. Fusion of matching algorithms for human identification using dental X-ray radiographs. IEEE Transactions on Information Forensics and Security. 2008;3(2):223–233. doi: 10.1109/TIFS.2008.919343. [DOI] [Google Scholar]
  • Nuansanong, Kiattisin & Leelasantitham (2014).Nuansanong J, Kiattisin S, Leelasantitham A. Diagnosis and interpretation of dental x-ray in case of deciduous tooth extraction decision in children using active contour model and J48 tree. 2014 International Electrical Engineering Congress (iEECON); 2014. pp. 1–4. [Google Scholar]
  • Obuchowicz Rafałand Nurzynska et al. (2018).Obuchowicz Rafałand Nurzynska K, Obuchowicz B, Urbanik A, Piórkowski A. Caries detection enhancement using texture feature maps of intraoral radiographs. Oral Radiology. 2018;36(3):275–287. doi: 10.1007/s11282-018-0354-8. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • Oktay (2017).Oktay AB. Tooth detection with convolutional neural networks. 2017 Medical Technologies National Congress (TIPTEKNO); 2017. pp. 1–4. [Google Scholar]
  • Olsen et al. (2009).Olsen GF, Brilliant SS, Primeaux D, Najarian K. An image-processing enabled dental caries detection system. 2009 ICME International Conference on Complex Medical Engineering; 2009. pp. 1–8. [Google Scholar]
  • Oprea et al. (2008).Oprea S, Marinescu C, Lita I, Jurianu M, Visan DA, Cioc IB. Image processing techniques used for dental x-ray image analysis. 2008 31st International Spring Seminar on Electronics Technology; 2008. pp. 125–129. [Google Scholar]
  • Osterloh & Viriri (2019).Osterloh D, Viriri S. Computer Aided Intervention and Diagnostics in Clinical and Medical Images. Berlin: Springer; 2019. Caries detection in non-standardized periapical dental x-rays; pp. 143–152. [Google Scholar]
  • Park & Park (2018).Park WJ, Park J-B. History and application of artificial neural networks in dentistry. European Journal of Dentistry. 2018;12(04):594–601. doi: 10.4103/ejd.ejd_325_18. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • Patanachai, Covavisaruch & Sinthanayothin (2010).Patanachai N, Covavisaruch N, Sinthanayothin C. Wavelet transformation for dental x-ray radiographs segmentation technique. 2010 Eighth International Conference on ICT and Knowledge Engineering; 2010. pp. 103–106. [Google Scholar]
  • Poonsri et al. (2016).Poonsri A, Aimjirakul N, Charoenpong T, Sukjamsri C. Teeth segmentation from dental x-ray image by template matching. 2016 9th Biomedical Engineering International Conference (BMEiCON); 2016. pp. 1–4. [Google Scholar]
  • Prajapati, Desai & Modi (2012).Prajapati DB, Desai NP, Modi CK. A simple and novel CBIR technique for features extraction using AM dental radiographs. 2012 International Conference on Communication Systems and Network Technologies; 2012. pp. 198–202. [Google Scholar]
  • Prajapati, Nagaraj & Mitra (2017).Prajapati SA, Nagaraj R, Mitra S. Classification of dental diseases using CNN and transfer learning. 2017 5th International Symposium on Computational and Business Intelligence (ISCBI); 2017. pp. 70–74. [Google Scholar]
  • Prakash, Gowsika & Sathiyapriya (2015).Prakash M, Gowsika U, Sathiyapriya S. Emerging Research in Computing, Information, Communication and Applications. Berlin: Springer; 2015. An identification of abnormalities in dental with support vector machine using image processing; pp. 29–40. [Google Scholar]
  • Purnama et al. (2015).Purnama IKE, Kurniastuti I, Rinastiti M, Purnomo MH. Semi-automatic determination of root canal length in dental x-ray image. 2015 4th International Conference on Instrumentation, Communications, Information Technology, and Biomedical Engineering (ICICI-BME); 2015. pp. 49–53. [Google Scholar]
  • Pushparaj, Gurunathan & Arumugam (2013).Pushparaj V, Gurunathan U, Arumugam B. An effective dental shape extraction algorithm using contour information and matching by mahalanobis distance. Journal of Digital Imaging. 2013;26(2):259–268. doi: 10.1007/s10278-012-9492-4. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • Pushparaj et al. (2013).Pushparaj V, Gurunathan U, Arumugam B, Baskaran A, Valliappan A. An effective numbering and classification system for dental panoramic radiographs. 2013 Fourth International Conference on Computing, Communications and Networking Technologies (ICCCNT); 2013. pp. 1–8. [Google Scholar]
  • Rad et al. (2015).Rad AE, Amin IBM, Rahim MSM, Kolivand H. Computational Intelligence in Information Systems. Berlin: Springer; 2015. Computer-aided dental caries detection system from x-ray images; pp. 233–243. [Google Scholar]
  • Rad et al. (2013).Rad AE, Mohd Rahim MS, Rehman A, Altameem A, Saba T. Evaluation of current dental radiographs segmentation approaches in computer-aided applications. IETE Technical Review. 2013;30(3):210–222. doi: 10.4103/0256-4602.113498. [DOI] [Google Scholar]
  • Rad et al. (2018).Rad AE, Rahim MSM, Kolivand H, Norouzi A. Automatic computer-aided caries detection from dental x-ray images using intelligent level set. Multimedia Tools and Applications. 2018;77:28843–28862. [Google Scholar]
  • Ramani, Vanitha & Valarmathy (2013).Ramani R, Vanitha NS, Valarmathy S. The pre-processing techniques for breast cancer detection in mammography images. International Journal of Image, Graphics and Signal Processing. 2013;5(5):47–54. doi: 10.5815/ijigsp.2013.05.06. [DOI] [Google Scholar]
  • Rana et al. (2017).Rana A, Yauney G, Wong LC, Gupta O, Muftu A, Shah P. Automated segmentation of gingival diseases from oral images. 2017 IEEE Healthcare Innovations and Point of Care Technologies (HI-POCT); Piscataway: IEEE; 2017. pp. 144–147. [Google Scholar]
  • Razali et al. (2014).Razali MRM, Ahmad NS, Hassan R, Zaki ZM, Ismail W. Sobel and canny edges segmentations for the dental age assessment. 2014 International Conference on Computer Assisted System in Health; 2014. pp. 62–66. [Google Scholar]
  • Razifar et al. (2005).Razifar P, Sandström M, Schnieder H, Långström B, Maripuu E, Bengtsson E, Bergström M. Noise correlation in PET, CT, SPECT and PET/CT data evaluated using autocorrelation function: a phantom study on data, reconstructed using FBP and OSEM. BMC Medical Imaging. 2005;5(1):5. doi: 10.1186/1471-2342-5-5. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • Ren et al. (2017).Ren S, He K, Girshick R, Sun J. Faster r-cnn: towards real-time object detection with region proposal networks. IEEE Transactions on Pattern Analysis and Machine Intelligence; Los Alamitos. Piscataway: IEEE; 2017. pp. 1137–1149. [DOI] [PubMed] [Google Scholar]
  • Romera-Paredes & Torr (2016).Romera-Paredes B, Torr PHS. Recurrent instance segmentation. European Conference on Computer Vision; 2016. pp. 312–329. [Google Scholar]
  • Ronneberger, Fischer & Brox (2015).Ronneberger O, Fischer P, Brox T. U-net: Convolutional networks for biomedical image segmentation. International Conference on Medical Image Computing and Computer-Assisted Intervention; 2015. pp. 234–241. [Google Scholar]
  • Said et al. (2006).Said EH, Nassar DEM, Fahmy G, Ammar HH. Teeth segmentation in digitized dental x-ray films using mathematical morphology. IEEE Transactions on Information Forensics and Security. 2006;1(2):178–189. doi: 10.1109/TIFS.2006.873606. [DOI] [Google Scholar]
  • Sarage & Jambhorkar (2012).Sarage GN, Jambhorkar S. Enhancement of chest x-ray images using filtering techniques. International Journal of Advanced Research in Computer Science and Software Engineering. 2012;2:308–312. [Google Scholar]
  • Sattar & Karray (2012).Sattar F, Karray FO. Dental x-ray image segmentation and object detection based on phase congruency. International Conference Image Analysis and Recognition; 2012. pp. 172–179. [Google Scholar]
  • Schmidhuber (2015).Schmidhuber J. Deep learning in neural networks: an overview. Neural Networks. 2015;61(3):85–117. doi: 10.1016/j.neunet.2014.09.003. [DOI] [PubMed] [Google Scholar]
  • Schwendicke et al. (2019).Schwendicke F, Golla T, Dreher M, Krois J. Convolutional neural networks for dental image diagnostics: a scoping review. Journal of Dentistry. 2019;91(7639):103226. doi: 10.1016/j.jdent.2019.103226. [DOI] [PubMed] [Google Scholar]
  • Shah et al. (2006).Shah S, Abaza A, Ross A, Ammar H. Automatic tooth segmentation using active contour without edges. 2006 Biometrics Symposium: Special Session on Research at the Biometric Consortium Conference; 2006. pp. 1–6. [Google Scholar]
  • Sharma, Rana & Kundra (2015).Sharma M, Rana NK, Kundra H. A review on the existing image segmentation techniques for the dental x-ray images. 2015;3:15–19. [Google Scholar]
  • Simonyan & Zisserman (2014).Simonyan K, Zisserman A. Very deep convolutional networks for large-scale image recognition. 2014. https://arxiv.org/abs/1409.1556 https://arxiv.org/abs/1409.1556 arXiv preprint arXiv:1409.1556.
  • Singh & Agarwal (2018).Singh HV, Agarwal R. Computational Vision and Bio Inspired Computing. Berlin: Springer; 2018. Diagnosis of carious legions using digital processing of dental radiographs; pp. 864–882. [Google Scholar]
  • Singh & Sehgal (2020).Singh P, Sehgal P. Numbering and classification of panoramic dental images using 6-layer convolutional neural network. Pattern Recognition and Image Analysis. 2020;30(1):125–133. doi: 10.1134/S1054661820010149. [DOI] [Google Scholar]
  • Son & Tuan (2016).Son LH, Tuan TM. A cooperative semi-supervised fuzzy clustering framework for dental x-ray image segmentation. Expert Systems with Applications. 2016;46(4):380–393. doi: 10.1016/j.eswa.2015.11.001. [DOI] [Google Scholar]
  • Son & Tuan (2017).Son LH, Tuan TM. Dental segmentation from x-ray images using semi-supervised fuzzy clustering with spatial constraints. Engineering Applications of Artificial Intelligence. 2017;59(12):186–195. doi: 10.1016/j.engappai.2017.01.003. [DOI] [Google Scholar]
  • Sornam & Prabhakaran (2017).Sornam M, Prabhakaran M. A new linear adaptive swarm intelligence approach using back propagation neural network for dental caries classification. 2017 IEEE International Conference on Power, Control, Signals and Instrumentation Engineering (ICPCSI); Piscataway: IEEE; 2017. pp. 2698–2703. [Google Scholar]
  • Srivastava et al. (2017).Srivastava MM, Kumar P, Pradhan L, Varadarajan S. Detection of tooth caries in bitewing radiographs using deep learning. 2017. https://arxiv.org/abs/1711.07312 https://arxiv.org/abs/1711.07312 arXiv preprint.
  • Torosdagli et al. (2018).Torosdagli N, Liberton DK, Verma P, Sincan M, Lee JS, Bagci U. Deep geodesic learning for segmentation and anatomical landmarking. IEEE Transactions on Medical Imaging. 2018;38(4):919–931. doi: 10.1109/TMI.2018.2875814. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • Tracy et al. (2011).Tracy KD, Dykstra BA, Gakenheimer DC, Scheetz JP, Lacina S, Scarfe WC, Farman AG. Utility and effectiveness of computer-aided diagnosis of dental caries. General Dentistry. 2011;59:136–144. [PubMed] [Google Scholar]
  • Tuan et al. (2018).Tuan TM, Fujita H, Dey N, Ashour AS, Ngoc VTN, Chu D-T. Dental diagnosis from x-ray images: an expert system based on fuzzy computing. Biomedical Signal Processing and Control. 2018;39(12):64–73. doi: 10.1016/j.bspc.2017.07.005. [DOI] [Google Scholar]
  • Tuan, Ngan & Son (2016).Tuan TM, Ngan TT, Son LH. A novel semi-supervised fuzzy clustering method based on interactive fuzzy satisficing for dental x-ray image segmentation. Applied Intelligence. 2016;45(2):402–428. doi: 10.1007/s10489-016-0763-5. [DOI] [Google Scholar]
  • Tuzoff et al. (2019).Tuzoff DV, Tuzova LN, Bornstein MM, Krasnov AS, Kharchenko MA, Nikolenko SI, Sveshnikov MM, Bednenko GB. Tooth detection and numbering in panoramic radiographs using convolutional neural networks. Dentomaxillofacial Radiology. 2019;48(4):20180051. doi: 10.1259/dmfr.20180051. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • Valizadeh et al. (2015).Valizadeh S, Goodini M, Ehsani S, Mohseni H, Azimi F, Bakhshandeh H. Designing of a computer software for detection of approximal caries in posterior teeth. Iranian Journal of Radiology. 2015;12(4):65. doi: 10.5812/iranjradiol.12(2)2015.16242. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • Veena Divya, Jatti & Revan Joshi (2016).Veena Divya K, Jatti A, Revan Joshi P. Appending active contour model on digital panoramic dental x-rays images for segmentation of maxillofacial region. 2016 IEEE EMBS Conference on Biomedical Engineering and Sciences (IECBES); Kaula Lumpur, Malaysia. Piscataway: IEEE; 2016. pp. 4–8. [Google Scholar]
  • Vijayakumari et al. (2012).Vijayakumari B, Ulaganathan G, Banumathi A, Banu AFS, Kayalvizhi M. Dental cyst diagnosis using texture analysis. 2012 International Conference on Machine Vision and Image Processing (MVIP); 2012. pp. 117–120. [Google Scholar]
  • Vila-Blanco, Tomás & Carreira (2018).Vila-Blanco N, Tomás I, Carreira MJ. Fully automatic teeth segmentation in adult OPG images. Multidisciplinary Digital Publishing Institute Proceedings. 2018;2(18):1199. doi: 10.3390/proceedings2181199. [DOI] [Google Scholar]
  • Wang et al. (2016).Wang C-W, Huang C-T, Lee J-H, Li C-H, Chang S-W, Siao M-J, Lai T-M, Ibragimov B, Vrtovec T, Ronneberger O, Fischer P, Cootes TF, Lindner C. A benchmark for comparison of dental radiography analysis algorithms. Medical Image Analysis. 2016;31(10):63–76. doi: 10.1016/j.media.2016.02.004. [DOI] [PubMed] [Google Scholar]
  • Wirtz, Mirashi & Wesarg (2018).Wirtz A, Mirashi SG, Wesarg S. Automatic teeth segmentation in panoramic x-ray images using a coupled shape model in combination with a neural network. International Conference on Medical Image Computing and Computer-Assisted Intervention; 2018. pp. 712–719. [Google Scholar]
  • Yang et al. (2018).Yang J, Xie Y, Liu L, Xia B, Cao Z, Guo C. Automated dental image analysis by deep learning on small dataset. 2018 IEEE 42nd Annual Computer Software and Applications Conference (COMPSAC); Piscataway: IEEE; 2018. pp. 492–497. [Google Scholar]
  • Yilmaz, Kayikcioglu & Kayipmaz (2017).Yilmaz E, Kayikcioglu T, Kayipmaz S. Computer-aided diagnosis of periapical cyst and keratocystic odontogenic tumor on cone beam computed tomography. Computer Methods and Programs in Biomedicine. 2017;146:91–100. doi: 10.1016/j.cmpb.2017.05.012. [DOI] [PubMed] [Google Scholar]
  • Zak et al. (2017).Zak J, Korzynska A, Roszkowiak L, Siemion K, Walerzak S, Walerzak M, Walerzak K. The method of teeth region detection in panoramic dental radiographs. International Conference on Computer Recognition Systems; 2017. pp. 298–307. [Google Scholar]
  • Zhang et al. (2018).Zhang K, Wu J, Chen H, Lyu P. An effective teeth recognition method using label tree with cascade network structure. Computerized Medical Imaging and Graphics. 2018;68(5):61–70. doi: 10.1016/j.compmedimag.2018.07.001. [DOI] [PubMed] [Google Scholar]
  • Zhou & Abdel-Mottaleb (2005).Zhou J, Abdel-Mottaleb M. A content-based system for human identification based on bitewing dental x-ray images. Pattern Recognition. 2005;38(11):2132–2142. doi: 10.1016/j.patcog.2005.01.011. [DOI] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Data Availability Statement

The following information was supplied regarding data availability:

This is a Review Article; there are no data or code files.


Articles from PeerJ Computer Science are provided here courtesy of PeerJ, Inc

RESOURCES