Abstract
The computed tomography angiography (CTA) postprocessing manually recognized by technologists is extremely labor intensive and error prone. We propose an artificial intelligence reconstruction system supported by an optimized physiological anatomical-based 3D convolutional neural network that can automatically achieve CTA reconstruction in healthcare services. This system is trained and tested with 18,766 head and neck CTA scans from 5 tertiary hospitals in China collected between June 2017 and November 2018. The overall reconstruction accuracy of the independent testing dataset is 0.931. It is clinically applicable due to its consistency with manually processed images, which achieves a qualification rate of 92.1%. This system reduces the time consumed from 14.22 ± 3.64 min to 4.94 ± 0.36 min, the number of clicks from 115.87 ± 25.9 to 4 and the labor force from 3 to 1 technologist after five months application. Thus, the system facilitates clinical workflows and provides an opportunity for clinical technologists to improve humanistic patient care.
Subject terms: Image processing, Machine learning, Medical research, Neurology, Mathematics and computing
Manual postprocessing of computed tomography angiography (CTA) images is extremely labor intensive and error prone. Here, the authors propose an artificial intelligence reconstruction system that can automatically achieve CTA reconstruction in healthcare services.
Introduction
Computed tomography angiography (CTA) is a widespread, minimally invasive, and cost-efficient imaging modality that is employed in routine clinical diagnoses of head and neck vessels1, especially in cerebrovascular disease, which represents one of the leading causes of severe disability and mortality worldwide2. To visually analyze the vasculature of the head and neck more efficiently, image reconstruction is usually performed by experienced computed tomography technologists. However, as the number of requests for CTA examinations have increased, the postprocessing staff cohort has become overwhelmed due to the time-consuming manual process3. Considering that vessel imaging reconstruction is required in clinical settings, an automatic reconstruction system can be easily integrated into the clinical workflow if processed segmentations are available.
Deep learning-based segmentation approaches have drawn increasing interest due to their self-learning and generalization abilities from large data volumes4. A major technological challenge of this study is accurate vessel segmentation without interruption of the blood vessels which is difficult because the vessels are tortuous and branched. Although most rule-based approaches such as centerline tracking, active contour models, or region growth exploit various vessel image characteristics to reconstruct vessels5,6, they are either hand-crafted or insufficiently validated7,8; thus, achieving the desired level of robustness on full brain vessel segmentation is difficult. In addition, segmentation is easily affected by other tissues; for example, the CT value of the blood vessel at the intracranial entrance is highly similar to the CT value of the cranium, which causes extravascular tissue adhesion and interferes with subsequent disease diagnosis. Therefore, accurate segmentation extraction of arterial vessel status is not available; instead, a system with adaptive properties that is capable of automatic segmentation with continuity is required.
Deep neural network architectures are the most effective way to overcome the above technological roadblocks9. In the past few years, various deep learning models have shown significant potential for image classification and segmentation tasks in various fields, particularly in neuroimaging10. Specifically, U-net, which is a specialized convolutional neural network (CNN), has demonstrated its promise in medical image analysis11. Moreover, a 3D-CNN model that conforms to the physiological, anatomical, and morphological features of the objective images is specifically designed for segmentation tasks and has shown high segmentation performance for biomedical images12.
In this present study, we sought to develop an automatic imaging reconstruction system (CerebralDoc) based on an optimized anatomy prior-knowledge based 3D-CNN to reconstruct original head and neck CTA images, assist technologists in their daily work and establish a time-saving work process. A thorough quantitative assessment is performed from three perspectives to evaluate the strength of CerebralDoc, including the algorithmic performance of the model, whether the image reconstruction quality satisfies diagnostic needs, and the efficiency of its clinical application with model augmentation. Collectively, we argue that artificial intelligence (AI) technology can be integrated into the radiology workflow to improve workflow efficiency and reduce medical costs.
Results
Patients and image characteristics
A total of 18,766 patients consecutively subjected to head and neck CTA scans at five tertiary hospitals were included in this study. Due to poor image quality, 507 scans were excluded via manual inspection. The mean age of the participants was 63 ± 12 years: 9370 patients (51.3%) were male, and 8889 patients (48.7%) were female. The patient cohort characteristics and those of the head and neck CTA scans that were employed for training, validation, and independent testing data sets are summarized in Table 1. The overall experimental design of the workflow diagram is shown in Fig. 1.
Table 1.
Parameters | Patient (images) metric |
---|---|
Patients characteristics | |
Number of patients (training/independent testing/clinical evaluation and application data sets) | 16,433/1826/152/3430 |
Male to female (training/independent testing/clinical evaluation and application data sets) | 1.05:1 (6539:6242)/1.07:1 (2831:2647)/1.17:1 (82:70)/1.03:1(1738:1692) |
Age (y) (training/independent testing/clinical evaluation and application data sets) | 64 ± 11/62 ± 14/66 ± 13/64 ± 13 |
Different manufactures | Number of patients in training/testing data sets |
GE | 7071/1011 |
Siemens | 5271/523 |
Philip | 2556/197 |
Toshiba | 1535/95 |
Different diseases | Number of patients in training/testing/clinical evaluation/application data sets |
Atherosclerosis | 11,503/1223/128/2982 |
Cerebrovascular disease | 5752/548/52/1744 |
Arterial aneurysm | 679/70/7/176 |
Moyamoya disease | 159/17/3/45 |
Interventional therapy | 448/46/2/138 |
Vascular variation | 247/26/2/59 |
Model performance
A total of 91,295 augmented images were employed for training. The 3D-CNN model was trained using a modified U-net with the addition of bottleneck-ResNet (BR) which automatically achieved optimized model parameter selection. The core design of the CerebralDoc system was the deep learning model, which was divided into two parts: ResU-Net and connected growth prediction model (CGPM). ResU-Net was primarily responsible for bone segmentation and vessel extraction, while the proposed CGPM was utilized to ensure vessel integrity. In part one, for feature map extraction, we applied the distribution of BRs from the ResU-Net base, which exhibited the optimal segmentation performance of the model. The model training process included both bone segmentation and vessel segmentation, and all the curves reached convergence by the 20th epoch. The vessel segmentation attained dice similarity coefficient (DSC) of 0.975 and 0.944, weighted vessel score (V-scores) of −0.975 and −0.929, and recall of 0.961 and 0.933 for the training and validation sets, respectively. The bone segmentation obtained DSC of 0.979 and 0.960, recall of 0.984 and 0.976 for the training and validation sets, respectively. Figure 2 shows a detailed image demonstrating the evolution of the DSC, loss function value, and the segmentation recall over the training and validation sets. In the testing process, the model was improved with a DSC of 0.951, V-scores of −0.936, and recall of 0.952, compared with the original model (DSC of 0.925, V-scores of −0.916 and recall of 0.933). The overall accuracy of segmentation is 93.1%. 32 (1.75%) failed bone subtraction and 84 (5.15%) failed vessel segmentation.
Clinical evaluation of CerebralDoc
We assessed the model in terms of its ability to automatically achieve head and neck CTA reconstruction, evaluated whether the reconstructed image quality satisfies the radiologist’s diagnostic needs, and compared the performance to human output. The concordance between the two raters on the independent testing data set of CerebralDoc and human output was satisfactory (Fleiss’ κ = 0.91 and 0.88). Among the 152 scans, the overall qualification rates of CerebralDoc and human output were 92.1% (140/152) and 93.4% (142/152), respectively, according to whether the reconstructed images were sufficient for rendering clinical diagnosis possible. However, the excellent image quality score of 5 for CerebralDoc was higher than that for human output [70.4% (107/152) vs 52.6% (80/152), χ2 = 11.199, P = 0.01]. In the evaluation of the qualification rate for different diseases, including atherosclerosis, cerebrovascular disease, arterial aneurysm, and vascular variation, there was no statistically significant difference between CerebralDoc and human output (Figs. 3a and 4).
In a qualitative evaluation of volume rendering (VR), maximum intensity projection (MIP), curve planar reconstruction (CPR), and curved multiple planar reformation (MPR), there were statistical differences with respect to VR (P < 0.001) and MIP (P < 0.001) in favor of CerebralDoc (Supplementary Table 1). However, there were no significant differences in CPR and curved-MPR between the two outputs. The improved visualization of VR was the farther cerebral vascular branches and clearer vessel image, especially in the evaluation of compensatory circulation of the ischemic region (Fig. 3b (a, b)) and cavernous sinus segments [Fig. 3b (c, d)]. In the qualitative evaluation of MIP generated by manual processing, 73 (48.0%) were rated with a score of 3, 44 (28.9%) were rated as having moderate bone residue but do not affect vessel observation with a score of 2 and 35 (23.1%) were rated as having severe bone residue and do affect vessel observation [Fig. 3b (e–g)]. In CerebralDoc, all the MIP images were rated with a score of 3 due to automatic bone segmentation in the algorithm. Moreover, MIP generated by AI directly could observe the wall calcification (Fig. 3b (h)).
For the disqualified cases, 10 (6.6%) cases were rated as having weak or poor image quality and 2 cases failed to reconstruct in CerebralDoc. Among them, partial vessel interruption occurred in 6 cases (3.9%), veins were misidentified as arteries in 2 cases (1.3%) and curved-MPR was lacking in 2 cases (1.3%). A possible reason for the 2 failed cases was vascular abnormal routine and weak CT signals. Among the 10 failed cases of human output, the severe bone residual was detected in 6 cases, vein artifacts were observed in 3 cases and multiple vessel interruption occurred in 1 case (Fig. 5).
Clinical application efficiency of CerebralDoc
CerebralDoc was applied to all the patients who underwent head and neck CTA examinations, including patients with atherosclerotic ulceration, aneurysm, moyamoya disease, and cervical stent implantations (Fig. 6). From July to November 2019, a total of 3430 patients underwent head and neck CTA scans at Xuanwu Hospital in the clinical setting; of these, 3122 (91.0%) were successfully reconstructed by the software as evaluated by the experienced technologists. Finally, 2649 reconstructed images were applied in the clinical context (Fig. 7a). Moreover, as the doctors’ trust improved, the probability of the reconstructed images being pushed to the clinic gradually increased from 243 to 753 (Fig. 7b).
Comparing the performance of CerebralDoc with the manual process
The average postprocessing time of two technologists was 13.48 ± 3.67 and 14.95 ± 3.65 min. Compared with CerebralDoc, the average time consumed was reduced from 14.22 ± 3.64 to 4.94 ± 0.36 min (P < 0.001, Fig. 8a). We also found that the consumed postprocessing time and click numbers were different in patient and normal person by manually recognized (P < 0.001), while there was no significant difference in AI (Fig. 8b, d). Moreover, the number of clicks was reduced from 115.87 ± 25.9 to 4 (P < 0.001), and the labor saved reduced the necessary labor force from 3 technologists to 1 (Fig. 8c). In addition, the number of CTA examinations doubled during the five months during which CerebralDoc was applied, and the success rate of the software gradually increased.
Discussion
In this study, we innovatively developed and validated a deep learning algorithm through physiological anatomical-based 3D-CNN for CTA head and neck image postprocessing. The model was developed with an optimized network by the distribution of BRs, and we proposed the CGPM to revise vascular segmentation errors and avoid partially missing vasculature. Moreover, we applied CerebralDoc to clinical scenarios for five months in Xuanwu Hospital to use an innovative and optimized deep learning algorithm that uses artificial intelligence to reconstruct CTA images used directly for head and neck vascular disease diagnoses. Our results indicated that a 3D-CNN deep learning algorithm can be trained to complete bone and vessel segmentation automatically with high sensitivity and specificity in a wide variety of enhanced CTA scans, including the aorta, carotid artery, and intracranial artery. Strikingly, CerebralDoc’s reconstructed image quality performance was evaluated by two radiologists with significant clinical experience; the performance showed a qualification rate of 92.1%. Therefore, this creative pipeline for the reconstruction of head and neck CTA offers numerous advantages, including consistency of interpretation, labor and time savings, reduced radiation doses, and high sensitivity and specificity.
Prior studies that have used deep learning for vessel segmentation tasks have mostly focused on algorithm optimization and model organization, which encouraged our development of a purposed automated clinical vessel segmentation tool (Supplementary Table 2). For example, a Y-Net model was developed to achieve 3D intracranial artery segmentation from 49 cases of magnetic resonance angiography (MRA) data and has achieved a precision of 0.819 and a DSC of 0.828 for the testing set13. The performance of this model was shown to perform better than the three traditional segmentation methods in both binary classification and visual evaluation. In addition, Michelle et al. have proposed HalfU-net to perform the images reconstruction by 2D patches, which yielded high performance for time-of-flight (TOF)-MRA vessel segmentation in patients with cerebrovascular disease: a dice value of 0.88, a 95 Hausdorff distance (HD) of 47 voxels and an absolute volume difference (AVD) of 0.448 voxels9. Furthermore, MS-Net, which was proposed to improve segmentation accuracy and precision, and significantly reduces the supervision cost, was proposed for full-resolution segmentation14. These previous works demonstrated satisfactory performance in specific vessel areas segmentation, such as aortic15,16, carotid17,18, or intracranial vessels19, which particularly inspired our study. However, complete head and neck CT scans that include three different vessels with different size magnitudes (aorta, carotid arteries, and intracranial blood vessels) make it difficult for models to capture cross-size-grade vessel characteristics. In addition, most previous studies were derived from relatively small data sets that were limited and lacked validation from real clinical data. Therefore, a comprehensive pipeline with a guaranteed maximum patch-size (256 × 256 × 256 voxels) for automatic head and neck CTA postprocessing by optimized physiological anatomical-based 3D-CNN is of great significance in clinical practice.
The application of a modified U-net framework was suggested by Ronneberger et al. in 201511, and previous studies have used this strategy on large images cropped into smaller pieces, causing the loss of the structural characteristics of the entire vessel20–22. Considering the importance of vessel integrity, our strategy instead inputs the original 3D slices into the 3D CNN with cropping and modifies the U-net architecture with the distribution of BR to automatically find the optimized model parameters. This approach achieves significant parameter reduction and results in high quantitative performance for head and neck vessel segmentation. Additionally, connecting the three different size levels of vessels is still a challenge. Some studies used level sets for multi-size vessel segmentation tasks, but these methods were not strictly automated and required user intervention during the process23,24. Based on the physiological anatomical information, we divided the entire volume into three regions (aorta, carotid, and intracranial regions) for the network to achieve better feature extraction of vessels of different sizes. Moreover, we proposed the CGPM by learning from the input data of original CTA images without labeling, the current segmentation result generated by ResNet1, ResNet2, and ResNet3, and the labeled CTA images to revise vascular segmentation errors and effectively avoid partial missing vasculature.
Rarely have studies investigated vessel segmentation from the head and neck CTA scans due to the wide range from the aorta to skull and the tortuous and branched arterial brain vessel. Meijs et al.25 proposed a robust model for CTA cerebral vascular segmentation in suspected stroke patients, but this method used hand-crafted features, which are time-consuming and difficult to obtain. Michelle et al.9 utilized the U-Net deep learning framework to perform MRA vessel segmentation on 66 patients with cerebrovascular disease; this model served as the starting point for developing a model applicable to clinical settings. Our study is a multicenter, large sample study to complete the automatic framework of bone segmentation and vessel segmentation from the aortic arch to the intracranial artery in accordance with the clinical postprocessing process. We translated this process into clinically applicable software named CerebralDoc, which is available for research and clinical practice.
To provide a referenceable benchmark, we evaluated the performance of CerebralDoc on one prospective database. As expected, its overall qualification rate and performance in different diseases were not significantly different from those in manual processing. In the visualization of pathology, VR and MIP significantly improved CerebralDoc in the presentation of cerebral vascular branches, vessel cleanness, and bone removal compared to manual reconstruction. A previous study revealed the advantage of deep learning in providing a more reliable recognition performance by exploring the 3D structure26. It also worked in our study by improving the inspection of vessel cleanness; the 3D-CNN could learn the vessel feature instead of the threshold value to accurately improve and capture the vessel signals27. VR also improved the resolution of the distal and small cerebral vessels in CerebralDoc, which benefit from the ability of deep learning to learn the edge characters and morphological features of vessels28. Moreover, accurate bone removal is an immense challenge in head and neck CTA reconstruction because patient motion between the plain scan and enhanced scan is inevitable, which severely affects the subtraction result. Thus, the MIP sequence would contain severe bone residuals, which affect the observation of vessels in manual reconstruction. However, this pipeline can serve as an automatic bone removal tool by the proposed ResU-Net, which applied the enhanced images only instead of the enhanced and plain image subtraction and reduced the radiation dose. This combination of less bone residual and lower radiation dose could be of an important significance for clinical use. Besides, MIP contained the wall calcification in CerebralDoc, which is beneficial for radiologists to evaluate directly and intuitively. The vessel interruption caused by severe stenosis in CerebralDoc was mostly solved at the end of November by collecting these cases and re-training the model. Additionally, the reason for vein misidentification was the delayed contrast phase which was inevitable due to the algorithm limitation.
In this study, we developed a computational tool that can automatically segment and extract entire vessels from head and neck scans. This tool can potentially assist technologists in clinical practice by quickly pinpointing vessels in axial CT images automatically. It is time-consuming and experience-dependent for technologists to locate very small vascular regions; thus, our approach greatly reduces the time that technologists must spend on each image. Second, this tool can help technologists remove artifacts caused by veins or metals in most cases and readily provide complete reconstructed images. Above all, this tool can be used to access head and neck postprocessing CTA images, including CPR, VR, MIP, and curved-MPR, which are useful for assisting radiologists and clinicians in the diagnosis of vascular disease after the quality of the images has been evaluated.
Although the results of this study are promising, it does have some limitations. In this pilot study, images from patients with severely anomalous origins (abnormal vessel routine except for left vertebral arising directly from the aorta and severely stenotic origin) of head and neck arteries were not included in the training and validation sets. Therefore, further clinical collection and testing is needed to assess the clinical accuracy for various forms of skull and neck vessels. Due to the relatively low incidence of such deformities, this problem does not affect our overall conclusion. This study was performed for over 2 years, and CerebralDoc functioned properly for five months. However, further evaluation of the reconstruction tool is needed to assess long-term accuracy and stability needs. In addition, this deep learning algorithm is dependent on big data; therefore, it is not appropriate for small sample sizes. Noisy images also need to be tested in future work to more fully assess the system’s robustness. In addition, CerebralDoc currently works only on head and neck CTA images.
In conclusion, CerebralDoc is a practical system to provide head and neck CTA reconstruction. It offers a time-saving and subjectivity-independent method compared to currently available techniques to optimize the reconstruction of CTA images, saving costs, and increasing efficiency. Due to the automatic AI-based standardized U-net and the generation of visualized post-processed images, CerebralDoc has the potential to become a standard component of the daily workflow, leading to the establishment of a process for radiological operations that can support increased communication with patients allow more humanistic care to be provided in the future.
Methods
Data preparation
The study was approved by the institutional review board of the University Medical Center. Institutional Review Board (IRB)/Ethics Committee approvals and informed consents were obtained in Capital Medical University Xuanwu Hospital. The data set used in this study consisted of 18,766 raw contrast-enhanced head and neck CTA images collected retrospectively and randomly from five tertiary hospitals in China, including 8593 scans from Xuanwu Hospital of Capital Medical University, 4407 from Beijing Friendship Hospital affiliated to Capital Medical University, 2584 from Shanghai Sixth People’s Hospital, 2096 from Hebei General Hospital and 1086 from Shandong Provincial Hospital. All the head and neck CTA scans were performed as part of the patients’ routine clinical care. Each scan included the aorta, carotid artery, and intracranial artery with ~561–967 slices in the Z axis. We excluded 507 (2.7%) CTA scans based on poor imaging quality or severe artifacts via manual inspection. All the patient CTA images were in DICOM format, and the scans were acquired from GE, Siemens, Philip, and Toshiba equipment using 64-, 192-, 256- or 320-detector row CT scanners (GE Revolution, GE Healthcare, Boston, USA; Somatom Force, Siemens Medical Solutions, Forchheim, Germany; Somatom Definition Flash, Siemens Medical Solutions, Forchheim, Germany; Philips Brilliance ICT 256-slice, Philips Medical Systems, Best, the Netherlands; Toshiba Aquilion one, Toshiba Medical Systems, Tokyo, Japan). Data were acquired by using 64, 192, 256, or 320 × 0.5, 0.625, or 0.7 mm collimations and a rotation time between 0.5 and 0.7. The primary thickness ranged from 0.4 to 0.6 mm; and the contrast agent was iohexol (Guerbet, France) at a concentration of 300 mgI/ml and an injection rate of 5 ml/s.
In this study, we initially obtained 18,259 patients with 14,461,128 head and neck CTA scanning images from January 2018 to February 2019 acquired mainly from four different CT manufacturers. During the training process, scans of 16,433 patients were divided into two models. Model 1 consisted of scans from 6387 patients for bone segmentation and model 2 consisted of scans from 10046 patients for vessel segmentation. During the testing process, scans from another 1826 patients were assigned to an independent testing cohort and used for performance assessment.
For the clinical evaluation set, 152 additional head and neck CTA scanning images were prospectively selected from Xuanwu Hospital between May 2019 and June 2019 to evaluate the clinical performance of the model. These images were also reconstructed by the technologists as a comparable output. Two experienced radiologists were blinded to the imaging sources and assessed whether the imaging quality fulfilled the diagnostic requirements based on a 5-point scale randomly.
Finally, a clinical application set was collected mainly to verify the performance of the proposed model in real clinical application scenarios from July to November 2019, this set, included the number of postprocessing operations, the time saved, and the clinical productivity.
Radiologist annotations
We adopted a hierarchical labeling method for all the scans in the database to reduce human error. The radiologists involved with the project completed a stakeholder evaluation form. First, all 18,259 samples were obtained from the segmentation results generated by the proposed model and prelabeled using ITKSNAP software with masks marking the areas of the bone, aorta, carotid artery, and intracranial artery. They were multiclass classification with four classes. We annotated the arteries including the aorta, subclavian artery, common carotid artery, vertebral artery, internal carotid artery, basilar artery, posterior cerebral artery, middle cerebral artery, and anterior cerebral artery. Next, ten technologists with 2 or more years of postprocessing experience independently corrected the prelabeled images to obtain the preliminary labeling results. Then, two radiologists with 5 or more years of experience double-checked the accuracy of the preliminary labeling results and relabeled cases as needed to avoid mislabeling. When disagreements occurred between the two radiologists, an arbitration expert (a more senior radiologist with more than 10 years experience) made the final decision. Finally, we adopted the labeled source images that passed a quality inspection as the final labeling results. The average time required for each case is ~22 min. Thus, each CTA scan was annotated with four masks to obtain the reliable ground truth.
Data preprocessing and augmentation
Although the samples were labeled in a strict manner, scattered noise caused by operator misclicking cannot be entirely avoided; however, the effect was too small to determine through human inspection. To solve this problem, we used a connected component detection algorithm to eliminate scattered noise and improve the robustness of the data. In addition, data augmentation is a strategy to increase data diversity. Augmentation helps in training models more effectively and improves robustness properties. In this study, we augmented the training data by horizontal flipping, rotation with a maximum angle of 25°, shifting by horizontally and vertically translating the images (a maximum of twenty pixels), and random occlusion by selecting a rectangle region in an image and erasing its pixels with a random value. For a data set of size N, we generate a data set of 5 N size. The rotation and shift were applied to improve the situation caused by the different body positions in the scanning process. The purpose of applying random occlusion was to avoid the artifacts caused by metal implants. Moreover, to improve model robustness, Gaussian noise was added to the training data.
Model development
This pipeline developed an automatic segmentation framework based on three cascaded ResU-Net models, which can remove the entire bone and fully extract the vessel. ResU-net is a modified U-net architecture with an added BR that can optimize the network and gain accuracy from a considerably increased depth29,30. ResU-Net2 was specially designed for bone edge optimization. Additionally, we proposed using CGPM to eliminate vascular segmentation errors and effectively avoid partial or missing vascular segments. The input data are integrated and continuous 3D spacing vascular images. The ResU-Net1 and ResU-Net2 models are responsible for the segmentation of the whole bone using a semantic segmentation under the same high resolution and for learning the physiological anatomical structural features that divide the vessel into the aorta, carotid artery, and intracranial artery based on the size of the artery. The ResU-Net3 model was adopted to achieve vascular segmentation according to the physiological anatomical structure. Finally, the proposed CGPM was applied to fix any ruptured vessels and ensure that they can feasibly be predicted and recognized at the morphological level (Fig. 9a, b). The architecture of CGPM is based on the 3D CNN and input data of the CGPM including the original CTA image without labeling, the current segmentation result generated by the combination of ResU-Net1, ResU-Net2, and ResU-Net3, and the labeled CTA images. The CGPM learned the location of the accurate partial missing and then accomplished the replenishment of the missed vessel. We also performed the ablation study to justify the contribution of each component in the pipeline (Fig. 9c).
Training details
We used BR as the principal architect in our network. Each layer in the BR is activated by the LeakyReLU function, which improves the flow of information and gradients throughout the network. BR can also identify the lowest loss value in the model due to its stack of three layers with 1 × 1 × 1, 3 × 3 × 3, and 1 × 1 × 1 convolutions. In total, eight BRs are integrated into the subnetwork, each following four convolutional downsampling units and four convolutional upsampling units, respectively. This approach combines multiple network depths within a single network and represents complex functions at different scales. The feature volumes from each BR are first convolved to a smaller number of channels and then upsampled to the input size.
The input of the 3D CNN consisted of original CT slices with cropping and resolution normalization. First, we apply the layer thickness and space to normalize the CT volume size and crop it into patches with a dimension of 256 × 256 × 256. These patches are the input data of the model, we combine the model output into the original CT volume. The 3D CNN was trained using the SGD optimizer with a momentum of 0.9, a peak learning rate of 0.1 for randomly initialized weights, a weight decay of 0.0001, and an initial learning rate of 0.01 that shrank by 0.99995 after each training step of 200,000 iterations. Each convolutional layer was activated by LeakyReLU (1) , and the sigmoid function (2) was used to predict the final segment probability results for each pixel. In this segmentation task, the loss function minimized 1 − D (ypred, ytarget), where D is the Dice coefficient between the predicted and target segmentations. Thus, the Dice coefficient loss function is defined as follows: ytarget and ypred are the ground truth and binary predictions from the neural networks, respectively:
1 |
After training the model with the training set, we used the validation set to validate the model and fine-tune the hyperparameters to optimize the model. All the experiments throughout this study were implemented based on Python and PyTorch libraries using a Linux server equipped with eight NVIDIA GeForce GTX 1080Ti GPUs with 11 GB of memory, and all the networks were trained from scratch using six of the GPUs.
For the failed cases in the validation process, vessel adhesion and vessel interruption were two main reasons for its occurrence. The vessel adhesion mostly occurred when the bone residual appeared around the vessel. Thus, the ResU-Net1 and ResU-Net2 were added to achieve bone segmentation. The vessel interruption often occurred where the blood flow was insufficient. This problem was solved by increasing the weight of the central region of the artery in the loss function, which enable the model to learn more characteristics about the central region. In addition, the CGPM, which targeted the occasional vessel interruption in the normal area, was proposed. When these failed cases occurred, more similar cases were added to train the model, which leads to better performance of the model.
Auto postprocessing image layout
The system achieves productivity and quality assurance characteristics by enabling traceability. Each image on the film can be traced and redirected to its original location in our image set. This system operates by separating the filming output process into two sub-tasks: verification and export. In the verification task, our program first processes skull subtraction and vessel extraction and generates four enlarged output images, including a VR, MIP, CPR, and MPR, indicating the position of the disease. Then, the four output images are placed in the first row of the film produced by the AI and can be verified and overwritten by technologists. After verification by the technologists, the reconstructed images can be exported to the PACS system.
Model effectiveness
DSC: The performance of the proposed model was evaluated using DSC, which represents the overlap ratio between the ground truth and segmentation results and is defined as follows:
2 |
where Y and , respectively, represent the ground truth and the binary predictions from the neural networks. Here, indicates the sum of each pixel value after calculating the dot product between the ground truth and the prediction. Two DSCs evaluated on both training and validation sets were used to avoid the question of excessively fits in the algorithm.
V-score: The V-score was innovatively proposed to reveal the continuity and integrity of vessel segmentation and is calculated according to the anatomy and morphological location of the vessel. A higher V-score value is correlated with a more critical vessel position; for example, the V-score of the proximal segments of a vessel is higher than that of its distal segments. The V-score represents the differences between the continuity of labels and predictions and is defined as follows:
3 |
Recall: Recall (also called sensitivity) is the proportion of positive cases that are correctly predicted as positive31. For a particular category, when a test image is an input into a trained 3D-CNN model, the model outputs the probability that the image belongs to this category. A hard binary classification can be made by thresholding this probability p ≥ t, where t is a threshold value. Recall is defined in its various common appellations by the following equation:
4 |
Clinical evaluation
The automatic head and neck CTA postprocessing system was designed to create a productive, quality-assured, and unattended image reconstruction process. The reconstruction images achieved by the technologists were regarded as comparable output and two experienced radiologists (10 years of experience) who were blinded to the processing conditions assessed whether the imaging quality fulfilled the clinical diagnostic requirements based on all processed image types (VR, MIP, CPR, and curved-MPR). All data sets were randomized and the qualification rates of image reconstruction were calculated separately. Subjective evaluations were scored as follows32: 5 = excellent image quality and contrast ratio, good vascular delineation, no artifacts, easy to diagnose; 4 = good image quality and contrast ratio, normal vascular delineation, with a few artifacts, but adequate for diagnosis; 3 = satisfactory image quality and contrast ratio, some artifacts, the artery was not clearly displayed but was sufficient for the diagnosis; 2 = weak image quality and contrast ratio, obvious artifacts, the artery was not clearly displayed and was not sufficient for diagnosis; 1 = poor image quality and contrast ratio, severe artifacts, difficult to distinguish the small artery and diagnosis was not possible. When the results differed between the 2 raters, a consensus was obtained.
We also compared the image quality, which was separately rated on VR, MIP, CPR, and curved-MPR based on the clinical request, which includes the vascular integrity, main cerebral vascular branches, vessel cleanness, wall calcification, degree of stenosis, and bone segmentation error. VR was graded mainly according to the vascular integrity, a branch of the main cerebral vascular and vessel cleanness as follows: 3 = good vascular delineation without interruption, excellent vascular side branch reconstruction, and clear vessel image; 2 = normal vascular delineations with partial interruption, good vascular side branch reconstruction with few residuals; 1 = whole vessel interruption, poor vascular side branch reconstruction and many residuals. MIP was graded based on the wall calcification and bone segmentation error as follows: 3 = no bone segmentation error and excellent vessel presentation; 2 = moderate bone residue but does not affect vessel observation; and 1 = severe bone residue and does affect vessel observation. CPR and curved-MPR were graded based on the vascular integrity and location and degree of stenosis, using the source images as “ground truth”: 3 = good vascular delineation without interruption, the same degree of stenosis as the source image without omission; 2 = normal vascular delineation with partial interruption, and biased degree of stenosis compared to the source image without omission; 1 = whole vessel interruption, and severe narrow omission compared to the source image.
The clinical application value of CerebralDoc was assessed between July and November 2019 at Xuanwu Hospital, from the following aspects: 1. the coverage rate of all patients who underwent the head and neck CTA examination; 2. the overall, monthly, and daily postprocessing numbers, and the numbers successfully pushed to the PACS; 3. the average time required for postprocessing and the number of clicks by CerebralDoc and the technologists (for 100 randomly selected patients); 4. the clinical productivity over the tested months; and 5. the error rates of the model (the reasons were explored).
Statistical analysis
All the analyses were conducted using SPSS software version 23.0 (IBM, Armonk, New York). Categorical variables were expressed as frequencies (percentages, %), and continuous variables were expressed as the mean ± SD or as the interquartile range according to the normality of the data. Student’s t-tests or Mann–Whitney U-tests were used to analyze the continuous data, and Fisher’s test was used for the categorical data as appropriate. The interobserver reliability was assessed using kappa analysis.
Reporting summary
Further information on research design is available in the Nature Research Reporting Summary linked to this article.
Supplementary information
Acknowledgements
This work has been supported by “Beijing Natural Science Foundation (Z190014)” and “Beijing Municipal Administration of Hospital’s Ascent Plan” (DFL20180802) and “Beijing Municipal science and technology projects” (Z201100005620009).
Author contributions
The studies were conceptualized, results analyzed, and manuscript drafted by F.F., J.W. Y.X. and F.L. wrote the code, trained the models, and wrote the first draft of the manuscript, with guidance from J.L., M.Z., and D.R. Also, Y.L. and F.Y. provided imaging quality evaluation and data interpretation. C.Z. provided supervision of statistical analysis and interpretation of data. Y.S. provided case selection and annotations. Z.Y., Y.L, Y.C., and X.W. provided raw training data and overall study supervision. All authors were involved in critical revisions of the manuscript, and have read and approved the final version.
Data availability
Currently, the source imaging data from five tertiary hospitals cannot be made publicly accessible due to privacy protection. The CerebralDoc is accessible from the webserver created for this study at http://test.platform.shukun.net/login. User name: test, password: 123456. Request for software access and the remaining clinical evaluation images generated by technologists can be submitted to F.F. The source data underlying Supplementary Table 2, Figs. 3a, 7, and 8 are provided as a Source Data file with this paper.
Competing interests
The authors declare no competing interests.
Footnotes
Peer review information Nature Communications thanks Robert DeLaPaz, Michelle Livne, Ziyue Xu and the other, anonymous, reviewer(s) for their contribution to the peer review of this work. Peer reviewer reports are available.
Publisher’s note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Supplementary information
Supplementary information is available for this paper at 10.1038/s41467-020-18606-2.
References
- 1.Saxena A, Eyk N, Lim ST. Imaging modalities to diagnose carotid artery stenosis: progress and prospect. Biomed. Eng. Online. 2019;1:66. doi: 10.1186/s12938-019-0685-7. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 2.Xu G, Ma M, Liu X, Hankey GJ. Is there a stroke belt in China and why? Stroke. 2013;44:1775–1783. doi: 10.1161/STROKEAHA.113.001238. [DOI] [PubMed] [Google Scholar]
- 3.Hilkewich MW. Written observations as a part of computed tomography angiography post processing by medical radiation technologists: a pilot project. J. Med Imaging Radiat. Sci. 2014;45:31–36. doi: 10.1016/j.jmir.2013.10.012. [DOI] [PubMed] [Google Scholar]
- 4.McBee MP, et al. Deep learning in radiology. Acad. Radio. 2018;25:1472–1480. doi: 10.1016/j.acra.2018.02.018. [DOI] [PubMed] [Google Scholar]
- 5.Tian Y, et al. A vessel active contour model for vascular segmentation. Biomed. Res. Int. 2014;2014:106490. doi: 10.1155/2014/106490. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 6.Lesage D, Angelini ED, Bloch I, Funka-Lea G. A review of 3D vessel lumen segmentation techniques: models, features and extraction schemes. Med. Image Anal. 2009;13:819–845. doi: 10.1016/j.media.2009.07.011. [DOI] [PubMed] [Google Scholar]
- 7.Zhao FJ, Chen YR, Hou YQ, He XW. Segmentation of blood vessels using rule-based and machine-learning-based methods: a review. Multimed. Syst. 2019;25:109–118. doi: 10.1007/s00530-017-0580-7. [DOI] [Google Scholar]
- 8.Lesage D, et al. A review of 3D vessel lumen segmentation techniques: models, features and extraction schemes. Med. Image Anal. 2009;13:819–845. doi: 10.1016/j.media.2009.07.011. [DOI] [PubMed] [Google Scholar]
- 9.Livne M, et al. A U-Net deep learning framework for high performance vessel segmentation in patients with cerebrovascular disease. Front. Neurosci. 2019;13:97. doi: 10.3389/fnins.2019.00097. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 10.Gao J, Jiang Q, Zhou B, Chen D. Convolutional neural networks for computer-aided detection or diagnosis in medical image analysis: an overview. Math. Biosci. Eng. 2019;16:6536–6561. doi: 10.3934/mbe.2019326. [DOI] [PubMed] [Google Scholar]
- 11.Ronneberger O, Fischer P, Brox T. U-Net: convolutional networks for biomedical image segmentation. MICCAI. 2015;9315:234–241. [Google Scholar]
- 12.Kamnitsas K, et al. Efficient multi-scale 3D CNN with fully connected CRF for accurate brain lesion segmentation. Med. Image Anal. 2017;36:61–78. doi: 10.1016/j.media.2016.10.004. [DOI] [PubMed] [Google Scholar]
- 13.Chen L. et al. Y-net: 3D intracranial artery segmentation using a convolutional autoencoder. In IEEE International Conference on Bioinformatic and Biomedicine10.1109/bibm.2017.8217741 (2017).
- 14.Shah MP, Merchant SN, Awate SP. MS-Net: mixed-supervision fully-convolutional networks for full-resolution segmentation. MICCAI. 2018;13:379–387. [Google Scholar]
- 15.Graffy PM, Liu J, O’Connor S, Summers RM, Pickhardt PJ. Automated segmentation and quantification of aortic calcification at abdominal CT: application of a deep learning-based algorithm to a longitudinal screening cohort. Abdom. Radio. 2019;44:2921–2928. doi: 10.1007/s00261-019-02014-2. [DOI] [PubMed] [Google Scholar]
- 16.Cao L, et al. Fully automatic segmentation of type B aortic dissection from CTA images enabled by deep learning. Eur. J. Radio. 2019;121:108713. doi: 10.1016/j.ejrad.2019.108713. [DOI] [PubMed] [Google Scholar]
- 17.Zhou Z, Shin J, Feng R, Hurst RT, Kendall CB, Liang J. Integrating active learning and transfer learning for carotid intima-media thickness video interpretation. J. Digit Imaging. 2019;32:290–299. doi: 10.1007/s10278-018-0143-2. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 18.Wu J, et al. Deep morphology aided diagnosis network for segmentation of carotid artery vessel wall and diagnosis of carotid atherosclerosis on black-blood vessel wall MRI. Med. Phys. 2019;46:5544–5561. doi: 10.1002/mp.13739. [DOI] [PubMed] [Google Scholar]
- 19.Akkus Z, Galimzianova A, Hoogi A, Rubin DL, Erickson BJ. Deep learning for brain MRI segmentation: state of the art and future directions. J. Digit. Imaging. 2017;30:449–459. doi: 10.1007/s10278-017-9983-4. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 20.Fabijańska A. Segmentation of corneal endothelium images using a U-Net-based convolutional neural network. Artif. Intell. Med. 2018;88:1–13. doi: 10.1016/j.artmed.2018.04.004. [DOI] [PubMed] [Google Scholar]
- 21.Huang Q, Sun J, Ding H, Wang X, Wang G. Robust liver vessel extraction using 3D U-Net with variant dice loss function. Comput Biol. Med. 2018;101:153–162. doi: 10.1016/j.compbiomed.2018.08.018. [DOI] [PubMed] [Google Scholar]
- 22.Norman B, Pedoia V, Majumdar S. Use of 2D U-net convolutional neural networks for automated cartilage and meniscus segmentation of knee MR imaging data to determine relaxometry and morphometry. Radiology. 2018;288:177–185. doi: 10.1148/radiol.2018172322. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 23.Chen Z, Qiu N, Song H, Xu L, Xiong Y. Optically guided level set for underwater object segmentation. Opt. Express. 2019;27:8819–8837. doi: 10.1364/OE.27.008819. [DOI] [PubMed] [Google Scholar]
- 24.Liu Y, et al. Distance regularized two level sets for segmentation of left and right ventricles from cine-MRI. Magn. Reson. Imaging. 2016;34:699–706. doi: 10.1016/j.mri.2015.12.027. [DOI] [PubMed] [Google Scholar]
- 25.Meijs M, et al. Robust segmentation of the full cerebral vasculature in 4D CT of suspected stroke patients. Sci. Rep. 2017;7:15622. doi: 10.1038/s41598-017-15617-w. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 26.Titinunt K, et al. VesselNet: A deep convolutional neural network with multi pathways for robust hepatic vessel segmentation. Comput. Med. Imaging Graph. 2019;7:74–83. doi: 10.1016/j.compmedimag.2019.05.002. [DOI] [PubMed] [Google Scholar]
- 27.Yan ZQ, Yang X, Cheng KT. A three-stage deep learning model for accurate retinal vessel segmentation. IEEE J. Biomed. Health Inf. 2019;23:1427–1436. doi: 10.1109/JBHI.2018.2872813. [DOI] [PubMed] [Google Scholar]
- 28.Melinscak M, Prentasic P, Loncaric S. Retinal vessel segmentation using deep neural networks. Proc. 10th Int. Conf. Computer Vis. Theory Appl. 2015;2:577–582. [Google Scholar]
- 29.Sze V, et al. Efficient processing of deep neural networks: a tutorial and survey. Proc. IEEE. 2017;105:2295–2329. doi: 10.1109/JPROC.2017.2761740. [DOI] [Google Scholar]
- 30.He K, et al. Deep residual learning for image recognition. IEEE Comput. Soc. 2016;1:770–778. [Google Scholar]
- 31.Saito T, Rehmsmeier M. Precrec: fast and accurate precision–recall and ROC curve calculations in R. Bioformatics. 2017;33:145–147. doi: 10.1093/bioinformatics/btw570. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 32.Boos J, Fang J, Heidinger BH, Raptopoulos V, Brook OR. Dual energy CT angiography: pros and cons of dual-energy metal artifact reduction algorithm in patients after endovascular aortic repair. Abdom. Radio. 2017;42:749–758. doi: 10.1007/s00261-016-0973-7. [DOI] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Supplementary Materials
Data Availability Statement
Currently, the source imaging data from five tertiary hospitals cannot be made publicly accessible due to privacy protection. The CerebralDoc is accessible from the webserver created for this study at http://test.platform.shukun.net/login. User name: test, password: 123456. Request for software access and the remaining clinical evaluation images generated by technologists can be submitted to F.F. The source data underlying Supplementary Table 2, Figs. 3a, 7, and 8 are provided as a Source Data file with this paper.