Abstract
Objectives
This study aimed to develop a robust and accurate deep learning network for detecting the posterior superior alveolar artery (PSAA) in dental cone-beam CT (CBCT) images, focusing on the precise localization of the centre pixel as a critical centreline pixel.
Methods
PSAA locations were manually labelled on dental CBCT data from 150 subjects. The left maxillary sinus images were horizontally flipped. In total, 300 datasets were created. Six different deep learning networks were trained, including 3D U-Net, deeply supervised 3D U-Net (3D U-Net DS), multi-scale deeply supervised 3D U-Net (3D U-Net MSDS), 3D Attention U-Net, 3D V-Net, and 3D Dense U-Net. The performance evaluation involved predicting the centre pixel of the PSAA. This was assessed using mean absolute error (MAE), mean radial error (MRE), and successful detection rate (SDR).
Results
The 3D U-Net MSDS achieved the best prediction performance among the tested networks, with an MAE measurement of 0.696 ± 1.552 mm and MRE of 1.101 ± 2.270 mm. In comparison, the 3D U-Net showed the lowest performance. The 3D U-Net MSDS demonstrated a SDR of 95% within a 2 mm MAE. This was a significantly higher result than other networks that achieved a detection rate of over 80%.
Conclusions
This study presents a robust deep learning network for accurate PSAA detection in dental CBCT images, emphasizing precise centre pixel localization. The method achieves high accuracy in locating small vessels, such as the PSAA, and has the potential to enhance detection accuracy and efficiency, thus impacting oral and maxillofacial surgery planning and decision-making.
Keywords: arteries, cone-beam computed tomography, deep learning, maxillary sinus
Introduction
Maxillary sinuses, the largest paranasal sinuses, are intrabony air-filled spaces within the maxillary bones. Anatomically, the maxillary sinus has a pyramidal structure and consists of 6 walls and floors.1–3 The blood supply to the maxillary sinus is provided by branches of the maxillary artery, namely, the posterior superior alveolar artery (PSAA), the inferior orbital artery (IOA), and the posterior lateral nasal artery.4,5 The PSAA and IOA give rise to extraosseous and intraosseous branches that form an arterial arcade around the maxillary sinus, supplying its lateral wall and, partially, the alveolar process.2,4,6,7
Preoperative assessment for a maxillary sinus bone graft involves identifying any inflammation or lesions that may predispose the sinus membrane to perforation.2 Accurate localization of the PSAA is important to minimize the risk of arterial damage during the opening of the bony window.2,8,9 While damage to the vessel is not typically life-threatening, it can lead to significant bleeding, necessitating haemostasis.9 Complications stemming from arterial bleeding can have serious implications, obstructing the surgical field and impacting the insertion and maintenance of bone graft material.2,8,10 Moreover, numerous studies have established a clear correlation between Schneiderian membrane perforation during surgical procedures and heightened rates of implant failure. This increased failure rate is often attributed to arterial bleeding originating from the PSAA, which can disrupt the surgeon’s field of vision.11,12 Additional reported complications include post-surgical bleeding, nasal bleeding, congestion, haematoma, empyema, and sinus infection.3 Thus, possessing knowledge about the location and course of the PSAA is necessary consideration in surgical planning.
With advances in imaging devices, such as the cone-beam CT (CBCT), it has become easier to detect the PSAA compared to the past. A 2015 study conducted by Varela-Centelles et al13 found that the detection rate of PSAA in the maxillary sinus reached 74.5%. In 2017, Tehranchi et al14 reported a detection rate of 87%, while in 2021, Fayek et al15 determined the rate to be 92.0%. Over time, the improved detection has enhanced the effectiveness of preoperative PSAA evaluation during surgical planning.4,8,16 The detection and preservation of the PSAA during oral and maxillofacial surgeries, such as maxillary sinus bone grafting,2,9,10 fracture repair,17,18 Caldwell-luc surgery,18 maxillary osteotomy,18 and tumour resection,19 help in reducing unnecessary bleeding and preventing associated complications. However, evaluating the PSAA using CBCT is often time-consuming, cumbersome, and prone to human error.
Recent advancements in deep learning-based techniques have demonstrated great potential in medical and dental imaging applications.20–22 Notably, digital architecture, such as 3D U-Net,23,24 3D V-Net,25 3D Attention U-Net,26,27 and 3D Dense U-Net28 have demonstrated excellent performance in medical imaging, even with limited training data.24 While structural properties can facilitate extraction of the medium to large vessels, accurately detecting small blood vessels remains challenging.29 The reliability of small vessel detection is hindered by variations in image quality, which renders manual labelling time-consuming and introduces errors in segmentation. Although many studies have been conducted on anatomical characteristics, such as the prevalence, diameter, and location of the PSAA,2–4,6,7,14–16,30,31 research on deep learning-based PSAA extraction has not yet been conducted. Automatic detection of certain processes through deep learning may have clinical utility, as it can alleviate the burden of cumbersome and potentially forgettable processes and minimize human errors.
This study aimed to propose a method for locating the PSAA in CBCT images using deep learning and identify a robust deep learning network for this task. Automatic detection of the PSAA before surgery has the potential to reduce complications by preventing rupture of the artery during surgery. The findings of this study may contribute to the development of a reliable and efficient method for PSAA detection using CBCT, thus, benefiting both clinicians and patients.
Methods
Patient selection and data preparation
This study was conducted retrospectively from data obtained for clinical purposes. It was approved by the Institutional Review Board of Seoul National University Dental Hospital, Seoul National University (No. ERI23007). The ethics committee approved the waiver for the informed consent because this was a retrospective study. The study was performed according to the Declaration of Helsinki.
Dental CBCT images of 300 patients who visited the Seoul National University Dental Hospital for clinical reasons between 2016 and 2020 were collected for this study. Patients’ data were obtained using the CS9300 Cone Beam Computed Tomography system manufactured by Carestream Health in New York, United States, operating at 80 kVp, 8 mA, and with a 24 s exposure time. The CBCT dimensions were 841 × 841 × 289 pixels, with voxel sizes of 0.2 × 0.2 × 0.2 mm3, and 16-bit depth.
All personal information including the patient’s name and chart number was anonymized and encrypted. The maxillary sinus sizes, shapes, and pathological conditions varied among the patients. Low-quality images with severe artefacts, haziness, or distortion, images in which the PSAA is undetectable, and abnormal defects, such as fracture lines, and benign and malignant lesions in the maxilla, were excluded. Finally, the dental CBCT images from 150 patients were included in this study.
The PSAA was manually annotated by an oral and maxillofacial surgeon with 16 years of experience using the imaging software (3D Slicer for Windows 10, Version 5.0.2; MIT, Cambridge, MA, United States). For each dental CBCT image, a sphere was drawn along the centre of the PSAA seen in each slice of the coronal plane, the sphere was overlapped, and interpolation was performed on the connection part. Labelling was performed using Gaussian and median smoothing functions while checking whether the size and shape of blood vessels were reflected in the cross-section (Figure 1). To ensure the accuracy of ground truth labelling and minimize systematic errors, we performed intra-observer validation on 20% of the cases under the same conditions at 3-month interval. The mean intra-observer variability on the test dataset was 1.731 ± 2.323 mm mean absolute error (MAE).
Figure 1.

The posterior superior alveolar artery (PSAA) captured using dental cone beam computed tomography (CBCT). (A) The axial plane of CBCT. The annotated PSAA is shown. (B) The spheres centred on the centre of the PSAA. (C) The coronal plane of CBCT. (D) The sagittal plane of CBCT.
We constructed a dataset of 300 right-sided CBCT volumes using CBCT images obtained from 150 patients. These were created by horizontally flipping the original images along the centreline, effectively generating mirrored versions of the left side. Subsequently, we divided this dataset into training, validation, and test sets, following a 6:2:2 ratio. To avoid including redundant left-right pairs from the same patient, we implemented careful measures during the initial dataset division. Specifically, we allocated individuals to the training and test sets right from the outset, ensuring a balanced distribution of 150 individuals into training, validation, and test sets, with a ratio of 90, 30, and 30 individuals, respectively. The left and right maxillary sinus areas were cropped to a region of interest (ROI), and ROIs were resized to. Zero padding was performed to maintain the input volume of the same length for all patients with regions of different lengths.22 As part of the preprocessing phase, we utilized an algorithm designed to automatically crop the sinus area from dental CBCT images.
To detect the centre of the PSAA using CBCT, a heatmap-based detection method was adopted to transfer each centreline of the PSAA into a heatmap.32 Coordinates of the centreline pixel is represented by a Gaussian heatmap , which is considered as a probability of a centreline pixel in the range [0, 1] as shown in Figure 2. The probability of a pixel is 1.0 at the centre of a heatmap. The probability of a pixel decreases below 1.0 as it moves further from the centre. The Gaussian distribution is defined as:
where is the scale factor to define the size of a heatmap, empirically set as 5. The is a standard deviation, which is the hyper-parameter to determine the sharpness of the Gaussian distribution. Centreline pixels with a lower show a much sharper distribution than a higher at the centre of a heatmap. CBCT images were used as input volumes and stacked heatmaps of centreline pixels were used as ground truth volumes for network training.
Figure 2.
A coronal slice of the posterior superior alveolar artery (PSAA) is shown in (A) CBCT images. The ground truth segmentation mask of (B) the PSAA, (C) converted to heatmap train data with label. The segmented label is converted into a heatmap where the centre pixel is 1, and decreases as it moves further away from the centre. Abbreviation: CBCT = computed tomography.
Deeply supervised multi-scale 3D network
The proposed deeply supervised multi-scale 3D network (3D U-Net MSDS) had a 3D encoder and decoder architecture consisting of 3D convolution blocks, multi-scale inputs, and deep supervisions.33
In the proposed 3D U-Net MSDS, 3D convolution blocks were comprised of a 3 × 3 × 3 convolution, batch normalization, and rectified linear unit at the encoder part. After repeating the 3D convolution block twice, a 2 × 2 × 2 max-pooling operation for extracting down-sampled feature maps. To alleviate volumetric information loss due to max-pooling operations, the multi-scale inputs down-sampled by 2 × 2 × 2 average pooling from the original input were merged at each level of layer in the encoder parts. The number of feature maps of 3D convolution blocks gradually increased from 32 to 64, 128, and 256 at the encoder part of 3D U-Net MSDS. Similar to the encoder structure, 2 3D convolution blocks, and a 3D transposed convolution were used for extracting feature maps and up sampled at each level of the decoder. The 3D transposed convolution had a 2 × 2 × 2 convolution with a stride of 2. Skip connections were used to concatenate feature maps of each level of layers in the encoder with the corresponding up-sampled feature maps. To mitigate the vanishing problem for encouraging backpropagation of the gradient flow,33 a deep supervision layer was embedded, which concatenated feature maps generated from output maps at every level of the layers in the decoder. The output activation was a linear function for regressing a heatmap. The number of feature maps of 3D convolution blocks gradually decreased from 256 to 128, 64, and 32 at the decoder part of the 3D U-Net MSDS (Figure 3).
Figure 3.
The 3D U-Net architecture with multiscale deep supervision. It combines the benefits of the 3D U-Net architecture, and the use of multiscale input and deep supervision, to enhance network performance.
The proposed networks were trained using RMSProp optimizer and mean squared error loss function. An initial learning rate was set to start at 10−3 and decrease to the minimum learning rate of 10−6 by reducing the learning rate by half when there is no performance improvement for 5 epochs in the validation dataset. The maximum training epoch was set to 300 with a batch size of 1, and training was terminated when there was no performance improvement in the validation dataset during 45 epochs. The proposed networks were implemented using Python3 and TensorFlow on Ubuntu 16.04 and an NVIDIA RTX A6000 graphic card. Data augmentation was performed to increase the variation of the actual training dataset, using rotation, Gaussian noise, and brightness. By increasing the variation of the training dataset in this way, overfitting can be reduced, and the model can perform better in practice.21,34
Performance evaluation
The performance of the centre pixel of the blood vessel was compared using 3D U-Net MSDS in conjunction with other networks, such as 3D V-Net,25 3D Attention U-Net,26 and 3D Dense U-Net.28
The detection performance was assessed based on the centre pixel. The MAE was utilized to quantify the discrepancy between the ground truth and predicted values on a point-to-point basis. The obtained MAE (mm) values were accompanied by the standard deviation (SD) errors, which were computed across all test data using the following formula:
and represent the coordinate for the ground truth and predicted centreline, respectively. denote the total number of data points in a test data.
The mean radial error (MRE) was also evaluated to quantify network performance. MRE (mm) was defined as the absolute distance between the ground truth and predicted value using the following equation:
and represent the coordinate for the ground truth and predicted centreline, respectively. denotes the total number of data points in the test data.
A successful detection rate (SDR) for each network was obtained. This SDR (%) represents the probability of successful detection when the MAE and MRE results for the prediction and the ground truth were <1, 1.5, 2, 2.5, and 3 mm, respectively. The SDR with a precision of less than a threshold value (t) mm can be expressed as:
represents the MAE or MRE score of test data , and is for the threshold value of a range of measurement values for evaluation, such as 1, 1.5, 2, 2.5, and 3 mm. , and n is the total number of the test data.
We used the one-way Kruskal-Wallis test tests to compare performances among 3D U-Net MSDS, 3D U-Net, 3D U-Net DS, 3D V-Net,25 3D Attention U-Net,26 and 3D Dense U-Net28 networks using PASW Statistics for Windows 10 (Version 18.0.0, IBM, Armonk, NY, United States) since the data did not follow a normal distribution. The statistical significance level was set to 0.05.
Results
The results in Table 1 show the detection performance of MAE and MRE by the 3D U-Net MSDS, 3D U-Net DS, 3D U-Net, 3D Dense U-Net, 3D V-Net, and 3D Attention U-Net networks. The performances were compared to a dataset with 60 CBCT volumes not used for training. The 3D U-Net MSDS network showed the lowest values of MAE and MRE for the centreline pixel detection compared to other 3D networks (P < .05).
Table 1.
The mean absolute error (MAE) and mean radial error (MRE) values from different networks (P < 0.05).
| Network | MAE (mm) | MRE (mm) |
|---|---|---|
| 3D Attention U-Net | 1.012 ± 1.248 | 1.559 ± 1.819 |
| 3D V-Net | 1.340 ± 1.594 | 2.102 ± 2.380 |
| 3D Dense U-Net | 1.064 ± 1.472 | 1.700 ± 2.286 |
| 3D U-Net | 1.681 ± 2.759 | 2.524 ± 3.987 |
| 3D U-Net DS | 0.804 ± 1.068 | 1.292 ± 1.626 |
| 3D U-Net MSDS | 0.696 ± 1.552 | 1.101 ± 2.270 |
The detection performance of the 6 networks was evaluated. The results showed that the 3D U-Net MSDS demonstrated the best performance, while the conventional 3D U-Net exhibited the worst performance. Notably, the detection performance of MAE and MRE achieved 0.696 ± 1.552 and 1.101 ± 2.270 mm, respectively, thus indicating that 3D U-Net MSDS modifications significantly enhanced the detection performance of 3D U-Net.
Tables 2 and 3 show the SDR of MAE and MRE by networks for PSAA detection, respectively. All 6 deep learning networks achieved an SDR of over 80% within 2 mm. The 3D U-Net MSDS demonstrated an accuracy of up to 86.67% within 1 mm, and an SDR of 96.67% within 2.5 mm. As shown in Table 3, the 3D U-Net MSDS demonstrated an accuracy of up to 80.00% within 1 mm, and 95.00% of SDR within 2.5 mm. The results were consistent with the MAE and MRE findings. In both the SDR of MAE and MRE, the 3D U-Net MSDS demonstrated better performance compared to the other models.
Table 2.
A successful detection rate (SDR) from different networks by mean absolute error.
| SDR (%) | <1 mm | <1.5 mm | <2 mm | <2.5 mm | <3 mm |
|---|---|---|---|---|---|
| 3D Attention U-Net | 71.67 | 80.00 | 85.00 | 88.33 | 90.00 |
| 3D V-Net | 63.33 | 68.33 | 80.00 | 83.33 | 86.67 |
| 3D Dense U-Net | 73.33 | 80.00 | 85.00 | 88.33 | 90.00 |
| 3D U-Net | 65.00 | 75.00 | 80.00 | 83.33 | 83.33 |
| 3D U-Net DS | 78.33 | 85.00 | 91.67 | 93.33 | 93.33 |
| 3D U-Net MSDS | 86.67 | 93.33 | 95.00 | 96.67 | 96.67 |
Table 3.
A successful detection rate (SDR) from different networks by mean radial error.
| SDR (%) | <1 mm | <1.5 mm | <2 mm | <2.5 mm | <3 mm |
|---|---|---|---|---|---|
| 3D Attention U-Net | 60.00 | 71.67 | 75.00 | 80.00 | 86.67 |
| 3D V-Net | 48.33 | 60.00 | 66.67 | 73.33 | 78.33 |
| 3D Dense U-Net | 61.67 | 73.33 | 78.33 | 81.67 | 85.00 |
| 3D U-Net | 56.67 | 65.00 | 70.00 | 78.33 | 80.00 |
| 3D U-Net DS | 68.33 | 76.67 | 81.67 | 85.00 | 91.67 |
| 3D U-Net MSDS | 80.00 | 86.67 | 91.67 | 95.00 | 95.00 |
Figure 4 illustrates the 3D reconstruction of PSAA positions predicted by the network, offering multiple examples. In the 3D visualizations and comparison with the ground truth, areas with disconnections or multiple detections were indicated by black arrows. As a result, 3D U-Net MSDS consistently displayed lower levels of disconnection and false positives when compared to other networks.
Figure 4.
Three-dimensional (3D) reconstruction of the posterior superior alveolar artery (PSAA) for various distinct cases offering multiple examples. (A) The ground truth using the (B) 3D U-Net MSDS, (C) 3D U-Net, (D) 3D U-Net DS, (E) 3D V-Net, (F) 3D Dense U-Net, and (G) 3D Attention U-Net. The 3D U-Net MSDS shows less PSAA disconnection and fewer false positives, compared to other networks, as indicated by the black arrow.
Figure 5 is intended to provide examples of the performance of the deep learning network in various cases, including scenarios such as the presence of impacted teeth, sinusitis. Each column represents a set of these diverse cases, illustrating the performance of each network. The prediction for the PSAA by the 3D U-Net MSDS alongside the ground truth was superimposed on the CBCT images. To assess the clinical significance of the network’s performance, we employed a visualization method involving the drawing of a rectangle with a height and a width of 2 pixels. This rectangle was centred around the pixel coordinate predicted by the deep learning network and was subsequently superimposed on the CBCT images. Predicted areas are highlighted in red, ground truth areas in blue, and overlapping areas in yellow.
Figure 5.
Multiple examples of the posterior superior alveolar artery (PSAA) detection results are displayed. In (A), the original CBCT image is presented, while (B) shows the results obtained with 3D U-Net MSDS. The subsequent columns, (C) through (G), correspond to the results achieved by the following networks: (C) 3D U-Net, (D) 3D U-Net DS, (E) 3D V-Net, (F) 3D Dense U-Net, and (G) 3D Attention U-Net, respectively. Each column represents a set of diverse cases, illustrating the performance of each network. To assess the clinical significance of these performances, a rectangle with a height and width of 2 pixels is placed around the centre pixel of the PSAA for visualization. Predicted areas are highlighted in red, ground truth in blue, and overlapping areas in yellow.
The MAE and MRE values for the test dataset were plotted from the posterior to the anterior regions of the image (Figure 6). Across most networks, there was a noticeable increase in the distance between the predictions and ground truth in both the posterior and anterior regions. However, the increase in error was more pronounced in the anterior region compared to the posterior. In the midsection of the vessels, there was an overall higher level of agreement in terms of distance. Among them, the 3D U-Net MSDS demonstrated the least variation in distance between the predicted values and ground truth when compared to other networks, indicating higher performance. This pattern can be attributed to the continuous nature of blood vessel anatomy, which makes it difficult to precisely distinguish the beginning and end points of the vessels.
Figure 6.
The line plots show the (A) mean absolute error (mm) and (B) mean radial error (mm) from the posterior slice to the anterior slice of the posterior superior alveolar artery (PSAA) for 3D U-Net MSDS, 3D U-Net, 3D U-Net DS, 3D V-Net, 3D Dense U-Net, and 3D Attention U-Net.
Discussion
In the field of medical imaging, deep learning technology has continuously evolved to detect and segment anatomical structures or abnormalities.20,21,34 Deep learning technology has been extensively employed in oral and maxillofacial imaging to ensure accurate segmentation and classification of anatomical structures, such as the mandibular canal segmentation (in CBCT),33 tooth segmentation (in CBCT),35 cyst and tumour detection (in CBCT and panorama),20,21 as well as facilitated precise age and gender estimations (in cephalogram).36 Deep learning algorithms have also been utilized for the detection of tooth caries (in periapical radiograph),37 apical lesions (in periapical radiograph),37,38 and landmark identifications (in cephalogram).39,40 Research on the use of deep learning in various aspects of maxillary sinus-related studies is ongoing. These include automatic segmentation of the sinuses,41 the diagnosis of sinusitis,42 and cyst detection.41 Importantly, to date, no deep learning studies have been conducted specifically on the PSAA.
Vascular structures in the maxillofacial region are often surrounded by cortical bone, such as the mandibular canal. Their characteristics are reflected in the image as density differences, which are useful for structural deep learning. Segmentation methods, commonly used for learning and detecting the 3D shape of anatomical structures, have shown success in improving the imaging of anatomical structures.43 This traditional approach is based on filtering-based algorithms that extract features using the optical and spatial characteristics of blood vessels.44 This method has been effective in detecting larger blood vessels and major anatomical structures. However, using this method to accurately identify small vessels, such as the PSAA, is challenging. The small diameter of the PSAA and its complex pathway all contribute to the difficulty. The insufficient contrast and poor continuity between the PSAA and the maxillary sinus wall further contribute to the difficulty of automatic segmentation using deep learning methods.43–45 As such, this study employed a centreline pixel detection method to accurately locate the PSAA using CBCT.
Centreline pixel detection is a method developed to accurately identify intricate vascular structures. The method originated from the idea of centreline prediction, which was first introduced by Tetteh et al.40 Centreline prediction involves post-processing blood vessel segmentation to locate the branching points and the skeleton of the blood vessels. For small-sized blood vessels, such as the PSAA, positional information is considered more crucial than other morphological features. In the centreline pixel detection method, the centre of the blood vessel is designated as 1 and serves as a central reference point. Deep learning algorithms are employed to locate these pixels accurately. To address the 3D nature of the PSAA and ensure accurate detection and localization, a 3D context-aware network was employed, ultimately resulting in favourable outcomes.
In this study, several widely utilized 3D networks in the field of deep learning for medical imaging were compared to the 3D U-Net MSDS. The centerline pixel detection method yielded the highest network performance (in both MAE and MRE) with the 3D U-Net MSDS, followed in decreasing order by the 3D U-Net DS, 3D Attention U-Net, 3D Dense U-Net, 3D V-Net, and conventional 3D U-Net. In particular, the 3D U-Net MSDS demonstrated the highest accuracy in the centreline pixel detection, as evidenced by a >90% SDR for both MAE and MRE within a detection range of 2 mm or less. This can be interpreted as successfully detecting the position of the vessel in over 90% of cases within a distance that is close to its diameter, which is typically around 1-2 mm for the PSAA.15,17,30 Our results demonstrated a significant performance improvement, compared to the 3D U-Net, which showed an SDR of 80% for MAE and 70% for MRE within the same range.
One of the main advantages of the 3D U-Net MSDS is its use of a deep supervision strategy, where intermediate predictions are made at different levels of the network and are used to provide feedback to earlier layers during training. This helps to prevent the vanishing gradient problem, which can occur in deep neural networks, leading to difficulties in training.46,47 With the deep supervision strategy, gradients are propagated back to earlier layers, allowing the network to learn more efficiently, converge faster, and ultimately, reduce false positives. Our findings indicate that the 3D U-Net DS exhibits improved performance compared to the 3D U-Net (MAE: 0.804 ± 1.068 vs 1.681 ± 2.759 mm and MRE: 1.292 ± 1.626 vs 2.524 ± 3.987 mm).
Another advantage of the 3D U-Net MSDS is its ability to effectively learn and incorporate features at multiple scales. The path of the PSAA varies for each individual, and the surrounding anatomical structures do not appear in a consistent pattern. Therefore, multiscale learning is necessary. The network consists of multiple encoder-decoder pathways, each with a different receptive field, allowing the network to capture information details at various levels.48,49 This enables the network to identify small and large structures in the image, and is particularly useful in medical image analysis. Our findings confirmed that the 3D U-Net DS (MAE: 0.804 ± 1.068 mm and MRE: 1.292 ± 1.626 mm) improved in performance when the multiscale strategy was applied (MAE: 0.696 ± 1.552 mm and MRE: 1.101 ± 2.270 mm). Overall, the 3D U-Net MSDS is a powerful and effective deep learning architecture for the volumetric segmentation of medical images, with the ability to capture features at multiple scales and efficiently train deep neural networks.
Our study found that the 3D U-Net MSDS outperformed all other tested networks, although all tested networks demonstrated clinically acceptable performance. Our findings affirm the efficacy of employing various 3D networks in medical imaging. These networks excel at learning and incorporating the inherent continuity present in medical image data, allowing them to probabilistically exclude non-continuous predictions. The primary difference from 2D networks lies in their ability to capture and reflect the 3D continuity of anatomical structures. This effect is particularly evident when applying the centreline pixel detection method. The choice of architecture depends on the specific requirements of the task. The specific benefits depend on the task and the dataset being used. Hence, careful experimentation and validation are necessary to determine the optimal configurations for any given application.
This study has some limitations. First, like other continuous anatomical structures, unclear structural boundaries of the PSAA resulted in a high number of false positives at the start and end points, which lowered the accuracy (Figure 6). Second, to reduce the memory requirements for handling large amounts of data, we used resized and cropped images rather than original images. This approach necessitated additional image preprocessing. Third, the generalizability of the results may be limited since the data were collected from a single type of CBCT.
To enhance the accuracy of the extended model, it would be beneficial to collect annotated data from multiple types of CBCT and different CBCT machines. Furthermore, inter-observer variability could not be assessed in this study as the annotation was performed by a single oral-maxillofacial surgeon. To further validate the generalizability of the proposed deep learning-based centre detection method, additional datasets annotated by different radiologists using diverse sources and in different clinical settings should be considered in future studies. In the future, studies to assess the performance and robustness of the method across different imaging modalities, patient populations, and pathological conditions, to ensure its reliability and applicability in real-world clinical scenarios, are warranted.
Conclusion
This study aimed to determine the most robust method and network for detecting the PSAA, an important anatomical structure that crosses the maxillary sinus. To overcome the ambiguous clarity in the cortical bone and issues associated with the low contrast of CBCT images, we introduced the centreline pixel detection method to identify the central point of the blood vessel on each coronal slice. Instead of segmentation, we compared the performance of this detection method among several 3D networks. To derive the most robust network, the 3D U-Net was customized. Our results demonstrated the superiority of the 3D U-Net MSDS compared to other networks. By combining the centreline pixel detection method and the 3D U-Net MSDS, the accuracy and efficiency of PSAA detection are enhanced. This automatic detection of PSAA could be beneficial in reducing bleeding and preventing complications in the maxillary sinus area. Further validation using larger datasets and better-designed clinical trials to validate the clinical utility of this approach is paramount.
Contributor Information
Jae-An Park, Department of Oral and Maxillofacial Radiology and Dental Research Institute, School of Dentistry, Seoul National University, 101 Daehak-ro, Jongno-gu, Seoul, 03080, South Korea.
DaEl Kim, Interdisciplinary Program in Bioengineering, Graduate School of Engineering, Seoul National University, 1 Gwanak-ro, Gwanak-gu, Seoul, 08826, South Korea.
Su Yang, Department of Applied Bioengineering, Graduate School of Convergence Science and Technology, Seoul National University, 1 Gwanak-ro, Gwanak-gu, Seoul, 08826, South Korea.
Ju-Hee Kang, Department of Oral and Maxillofacial Radiology, Seoul National University Dental Hospital, 101 Daehak-ro, Jongno-gu, Seoul, 03080, South Korea.
Jo-Eun Kim, Department of Oral and Maxillofacial Radiology and Dental Research Institute, School of Dentistry, Seoul National University, 101 Daehak-ro, Jongno-gu, Seoul, 03080, South Korea.
Kyung-Hoe Huh, Department of Oral and Maxillofacial Radiology and Dental Research Institute, School of Dentistry, Seoul National University, 101 Daehak-ro, Jongno-gu, Seoul, 03080, South Korea.
Sam-Sun Lee, Department of Oral and Maxillofacial Radiology and Dental Research Institute, School of Dentistry, Seoul National University, 101 Daehak-ro, Jongno-gu, Seoul, 03080, South Korea.
Won-Jin Yi, Department of Oral and Maxillofacial Radiology and Dental Research Institute, School of Dentistry, Seoul National University, 101 Daehak-ro, Jongno-gu, Seoul, 03080, South Korea.
Min-Suk Heo, Department of Oral and Maxillofacial Radiology and Dental Research Institute, School of Dentistry, Seoul National University, 101 Daehak-ro, Jongno-gu, Seoul, 03080, South Korea.
Author contributions
J.A. Park and D. Kim contributed equally and are considered co-senior authors of this work. In addition, they had full access to all the data in the study and took responsibility for the integrity of data and accuracy of the data analysis.
Funding
This work was supported by a National Research Foundation of Korea (NRF) grant, funded by the Korean Government (Ministry of Science and Information Communication Technology) (Number: 2023R1A2C200532611).
This work was also supported by a Korea Medical Device Development Fund grant, funded by the Korean Government (Ministry of Science and Information Communication Technology; Ministry of Trade, Industry, and Energy; Ministry of Health & Welfare; Ministry of Food and Drug Safety) (Project Number: 1711194231, KMDF_PR_20200901_0011), (Project Number:1711174552, KMDF_PR_20200901_0147).
Conflicts of interest
None declared.
References
- 1. Whyte A, Boeddinghaus R.. The maxillary sinus: physiology, development and imaging anatomy. Dentomaxillofac Radiol. 2019;48(8):20190205. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 2. Park WH, Choi SY, Kim CS.. Study on the position of the posterior superior alveolar artery in relation to the performance of the maxillary sinus bone graft procedure in a Korean population. J Korean Assoc Oral Maxillofac Surg. 2012;38(2):71-77. [Google Scholar]
- 3. Tofangchiha M, Hematzadeh S, Vali ME, et al. Anatomical localization of posterior superior alveolar artery: a retrospective study by cone-beam computed tomography. Dent Med Probl. 2022;59(3):407-412. [DOI] [PubMed] [Google Scholar]
- 4. Solar P, Geyerhofer U, Traxler H, Windisch A, Ulm C, Watzek G.. Blood supply to the maxillary sinus relevant to sinus floor elevation procedures. Clin Oral Implants Res. 1999;10(1):34-44. [DOI] [PubMed] [Google Scholar]
- 5. Iwanaga J, Wilson C, Lachkar S, Tomaszewski KA, Walocha JA, Tubbs RS.. Clinical anatomy of the maxillary sinus: application to sinus floor augmentation. Anat Cell Biol. 2019;52(1):17-24. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 6. Kqiku L, Biblekaj R, Weiglein AH, Kqiku X, Städtler P.. Arterial blood architecture of the maxillary sinus in dentate specimens. Croat Med J. 2013;54(2):180-184. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 7. Yang SM, Kye SB.. Location of maxillary intraosseous vascular anastomosis based on the tooth position and height of the residual alveolar bone: computed tomographic analysis. J Periodontal Implant Sci. 2014;44(2):50-56. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8. Wolf MK, Rostetter C, Stadlinger B, Locher M, Damerau G.. Preoperative 3D imaging in maxillary sinus: brief review of the literature and case report. Quintessence Int. 2015;46:627-631. [DOI] [PubMed] [Google Scholar]
- 9. Testori T, Rosano G, Taschieri S, Del Fabbro M.. Ligation of an unusually large vessel during maxillary sinus floor augmentation. A case report. Eur J Oral Implantol. 2010;3:255-258. [PubMed] [Google Scholar]
- 10. Flanagan D. Arterial supply of maxillary sinus and potential for bleeding complication during lateral approach sinus elevation. Implant Dent. 2005;14(4):336-338. [DOI] [PubMed] [Google Scholar]
- 11. Jamcoski VH, Faot F, Marcello-Machado RM, Melo ACM, Fontão F.. 15-year retrospective study on the success rate of maxillary sinus augmentation and implants: influence of bone substitute type, presurgical bone height, and membrane perforation during sinus lift. Biomed Res Int. 2023;2023:9144661. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 12. Al-Moraissi E, Elsharkawy A, Abotaleb B, Alkebsi K, Al-Motwakel H.. Does intraoperative perforation of Schneiderian membrane during sinus lift surgery causes an increased the risk of implants failure?: a systematic review and meta regression analysis. Clin Implant Dent Relat Res. 2018;20(5):882-889. [DOI] [PubMed] [Google Scholar]
- 13. Varela-Centelles P, Loira-Gago M, Seoane-Romero J, Takkouche B, Monteiro L, Seoane J.. Detection of the posterior superior alveolar artery in the lateral sinus wall using computed tomography/cone beam computed tomography: a prevalence meta-analysis study and systematic review. Int J Oral Maxillofac Surg. 2015;44(11):1405-1410. [DOI] [PubMed] [Google Scholar]
- 14. Tehranchi M, Taleghani F, Shahab S, Nouri A.. Prevalence and location of the posterior superior alveolar artery using cone-beam computed tomography. Imaging Sci Dent. 2017;47(1):39-44. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15. Fayek MM, Amer ME, Bakry AM.. Evaluation of the posterior superior alveolar artery canal by cone-beam computed tomography in a sample of the Egyptian population. Imaging Sci Dent. 2021;51(1):35-40. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 16. Pandharbale AA, Gadgil RM, Bhoosreddy AR, et al. Evaluation of the posterior superior alveolar artery using cone beam computed tomography. Pol J Radiol. 2016;81:606-610. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 17. Hwang K, Choi HG.. Bleeding from posterior superior alveolar artery in Le Fort I fracture. J Craniofac Surg. 2009;20(5):1610-1612. [DOI] [PubMed] [Google Scholar]
- 18. Rahpeyma A, Khajehahmadi S.. Alveolar antral artery: review of surgical techniques involving this anatomic structure. Iran J Otorhinolaryngol. 2014;26:73-78. [PMC free article] [PubMed] [Google Scholar]
- 19. Kurozumi M, Takumi Y, Uehara T, et al. Preoperative arterial embolization to avoid intraoperative bleeding during endoscopic sinus surgery for organized hematoma of the maxillary sinus: a case series and literature review. Interv Radiol (Higashimatsuyama). 2021;6(3):102-107. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 20. Schwendicke F, Samek W, Krois J.. Artificial intelligence in dentistry: chances and challenges. J Dent Res. 2020;99(7):769-774. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 21. Heo MS, Kim JE, Hwang JJ, et al. Artificial intelligence in oral and maxillofacial radiology: what is currently possible? Dentomaxillofac Radiol. 2021;50:20200375. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 22. Yamashita R, Nishio M, Do RKG, Togashi K.. Convolutional neural networks: an overview and application in radiology. Insights Imaging. 2018;9(4):611-629. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 23. Çiçek Ö, Abdulkadir A, Lienkamp SS, Brox T, Ronneberger O.. 3D U-Net: learning dense volumetric segmentation from sparse annotation. MICCAI. 2016;9901:424-432. [Google Scholar]
- 24. Ronneberger O, Fischer P, Brox T.. U-Net: convolutional networks for biomedical image segmentation. MICCAI. 2015;9351:234-241. [Google Scholar]
- 25. Milletarì F, Navab N, Ahmadi S. V-Net: Fully Convolutional Neural Networks for Volumetric Medical Image Segmentation. 2016 Fourth International Conference on 3D Vision, Stanford, CA, USA; 2016:565-571. doi: 10.1109/3DV.2016.79
- 26. Oktay O, Schlemper J, Folgoc L, Lee M, Heinrich M, Misawa K, Mori K, et al. Attention U-Net: learning where to look for the pancreas. Medical Imaging with Deep Learning, (MIDL) 2018.
- 27. Li S, Dong M, Du G, Mu X.. Attention Dense-U-Net for automatic breast mass segmentation in digital mammogram. IEEE Access. 2019;7:59037-59047. [Google Scholar]
- 28. Huang G, Liu Z, Van Der Maaten L, Weinberger K. Densely Connected Convolutional Networks, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA; 2017:2261-2269. doi: 10.1109/CVPR.2017.243
- 29. Jerman T, PernušF, Likar B, Špiclin Ž. Beyond Frangi: an improved multiscale vesselness filter. Proc. SPIE 9413, Medical Imaging 2015: Image Processing, 94132A.
- 30. Ibrahim AA, Al Nakib LH.. Location and diameter of the posterior superior alveolar artery among Iraqi subjects using computed tomography. IOSR. 2016;15(09):80-84. [Google Scholar]
- 31. Kqiku L, Biblekaj R, Weiglein AH.. Location of the extraosseous and intraosseous arterial anastomosis of the maxillary sinus in edentulous specimens. Clin Oral Investig. 2016;20(8):2311-2314. [DOI] [PubMed] [Google Scholar]
- 32. Payer C, Štern D, Bischof H, Urschler M.. Integrating spatial configuration into heatmap regression based CNNs for landmark localization. Med Image Anal. 2019;54:207-219. [DOI] [PubMed] [Google Scholar]
- 33. Jeoun BS, Yang S, Lee SJ, et al. Canal-Net for automatic and robust 3D segmentation of mandibular canals in CBCT images using a continuity-aware contextual network. Sci Rep. 2022;12(1):13460. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 34. Nagi R, Aravinda K, Rakesh N, Gupta R, Pal A, Mann AK.. Clinical applications and performance of intelligent systems in dental and maxillofacial radiology: a review. Imaging Sci Dent. 2020;50(2):81-92. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 35. Cui Z, Fang Y, Mei L, et al. A fully automatic AI system for tooth and alveolar bone segmentation from cone-beam CT images. Nat Commun. 2022;13(1):2096. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 36. Khazaei M, Mollabashi V, Khotanlou H, Farhadian M.. Sex determination from lateral cephalometric radiographs using an automated deep learning convolutional neural network. Imaging Sci Dent. 2022;52(3):239-244. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 37. Musri N, Christie B, Ichwan SJA, Cahyanto A.. Deep learning convolutional neural network algorithms for the early detection and diagnosis of dental caries on periapical radiographs: a systematic review. Imaging Sci Dent. 2021;51(3):237-242. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 38. Song IS, Shin HK, Kang JH, et al. Deep learning-based apical lesion segmentation from panoramic radiographs. Imaging Sci Dent. 2022;52(4):351-357. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 39. Lee JH, Yu HJ, Kim MJ, Kim JW, Choi J.. Automated cephalometric landmark detection with confidence regions using Bayesian convolutional neural networks. BMC Oral Health. 2020;20(1):270. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 40. Kim YH, Lee C, Ha EG, Choi YJ, Han SS.. A fully deep learning model for the automatic identification of cephalometric landmarks. Imaging Sci Dent. 2021;51(3):299-306. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 41. Choi H, Jeon KJ, Kim YH, Ha EG, Lee C, Han SS.. Deep learning-based fully automatic segmentation of the maxillary sinus on cone-beam computed tomographic images. Sci Rep. 2022;12(1):14009. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 42. Serindere G, Bilgili E, Yesil C, Ozveren N.. Evaluation of maxillary sinusitis from panoramic radiographs and cone-beam computed tomographic images using a convolutional neural network. Imaging Sci Dent. 2022;52(2):187-195. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 43. Tetteh G, Efremov V, Forkert ND, et al. DeepVesselNet: vessel segmentation, centerline prediction, and bifurcation detection in 3-D angiographic volumes. Front Neurosci. 2020;14:592352. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 44. Xian Z, Wang X, Yan S, Yang D, Chen J, Peng C.. Main coronary vessel segmentation using deep learning in smart medical. Math Probl Eng. 2020;2020:8858344. [Google Scholar]
- 45. Li Y, Liu X, Jia X, et al. A high-performance deep-learning-based pipeline for whole-brain vasculature segmentation at the capillary resolution. Bioinformatics. 2023;39(4):btad145. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 46. Tang Y, Yang F, Yuan S, Zhan C. A multi-stage framework with context information fusion structure for skin lesion segmentation, 2019 IEEE 16th International Symposium on Biomedical Imaging (ISBI 2019), Venice, Italy; 2019:1407-1410. doi: 10.1109/ISBI.2019.8759535
- 47. Agarwal A, Mishra A, Basavarajaiah M, Sharma P, Tanwar S.. Dilated volumetric network: an enhanced fully convolutional network for volumetric prostate segmentation from magnetic resonance imaging. Pattern Recognit Image Anal. 2021;31(2):228-239. [Google Scholar]
- 48. Bouget D, Pedersen A, Hosainey SAM, Solheim O, Reinertsen I.. Meningioma segmentation in T1-weighted MRI leveraging global context and attention mechanisms. Front Radiol. 2021;1:711514. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 49. Zhou T, Tan T, Pan X, Tang H, Li J.. Fully automatic deep learning trained on limited data for carotid artery segmentation from large image volumes. Quant Imaging Med Surg. 2021;11(1):67-83. [DOI] [PMC free article] [PubMed] [Google Scholar]





