Abstract
Abdominal aortic aneurysms (AAA) may remain clinically silent until they enlarge and patients present with a potentially lethal rupture. This necessitates early detection and elective treatment. The goal of this study was to develop an easy-to-train algorithm which is capable of automated AAA screening in CT scans and can be applied to an intra-hospital environment. Three deep convolutional neural networks (ResNet, VGG-16 and AlexNet) were adapted for 3D classification and applied to a dataset consisting of 187 heterogenous CT scans. The 3D ResNet outperformed both other networks. Across the five folds of the first training dataset it achieved an accuracy of 0.856 and an area under the curve (AUC) of 0.926. Subsequently, the algorithms performance was verified on a second data set containing 106 scans, where it ran fully automated and resulted in an accuracy of 0.953 and an AUC of 0.971. A layer-wise relevance propagation (LRP) made the decision process interpretable and showed that the network correctly focused on the aortic lumen. In conclusion, the deep learning-based screening proved to be robust and showed high performance even on a heterogeneous multi-center data set. Integration into hospital workflow and its effect on aneurysm management would be an exciting topic of future research.
Keywords: deep learning, computed X ray tomography, abdominal aortic aneurysm, image classification, interpretable artificial intelligence
1. Introduction
Abdominal aortic aneurysm (AAA) is a potentially life-threatening condition [1,2,3]. Possible rupture is associated with high mortality exceeding 50% [4,5,6]. In clinical routine, a small AAA, as a low prevalence disease, may be identified as a co-finding on abdominal computed tomography (CT) images performed for various reasons. The focus on other clinical questions and the time-consuming nature of a detailed AAA analysis might lead to underreporting and delayed diagnosis [7]. Therefore, patients might become discharged without detection of an early AAA. This may potentially lead to a delay in treatment since surveillance programs have been shown to be of benefit [8]. Ultimately, spontaneous rupture can be prevented if larger AAAs are treated surgically or interventionally [9]. Deep learning enables fast and highly accurate analysis of image data, and therefore, seems highly suited to contribute to the management of AAA.
With deep learning being introduced for various use-cases in medicine, a new and promising era of technical support and guidance for physicians is emerging [10]. Research effort has recently been made towards utilizing deep learning for AAA detection, segmentation, and prognostic evaluation. Mohammadi et al. [11] and López-Linares et al. [12] both described a 2D convolutional neural network (CNN)-based cascading pipeline for automated detection and segmentation of AAA in abdominal CT scans. Habijan et al. [13] have applied a 3D U-Net with deep supervision to segment AAAs in CT. These advances are useful for intervention planning [2]. Algorithm-based AAA growth prediction from CTA scans [14,15,16] and its potential for detection of aortic dissection and monitoring of endovascular aneurysm repair (EVAR) therapy complication has been shown [17,18,19].
A robust and automated algorithm that can be included in routine clinical workflow remains a great challenge [20]. The lack of algorithm generalizability is a central obstacle, which usually is caused by development on highly preselected data sets containing scans from a limited number of scanners and mostly exclusively CTA contrast phases.
The aim of this study is to develop and validate an easily trainable and fully automated deep learning 3D AAA screening algorithm, which can run as a background process in the clinic workflow. The main requirements are robustness, reliability, and precision as well as automation, whereas the algorithm training and clinical application should be feasible with minimal effort.
2. Materials and Methods
2.1. Patients and Data Sets
Institutional Review Board (IRB) approval for this study was obtained from the ethics committee II of the Medical Faculty Mannheim, Heidelberg University (2016-863R-MA, 17 November 2016).
We acquired the training data set from our radiology information system (RIS) based on keyword queries (for details see Appendix A). The presence of an AAA was reported by two radiologists, of which one had at least five years of experience. The studies were reviewed by a radiology resident with 2.5 years of experience in the interpretation of vascular CT scans to ensure data quality and reproducibility. The abdominal aorta was considered aneurysmatic when the aorta exceeded a 50% focal increase of its diameter [21]. Special care was taken to ensure that the data set contains scans of AAA of various size and shape with multiple co-findings and various devices such as intra-arterial stents and metallic interferences being present to represent a realistic spectrum of clinical data.
Annotation of the training data set was performed in two ways. Firstly, each CT scan was assigned to one of two classes (0: no AAA, 1: AAA). Secondly, the axial position of the origin of the most cranial left renal artery for the aorta was noted as an anchor point in every scan, which was subsequently used to automatically extract a sub volume of standardized size during training. Four example cases are shown in Figure 1. Distribution of classes and further statistics of our data set are listed in Table 1.
Table 1.
Training Set |
Verification Set | |||||
---|---|---|---|---|---|---|
TCIA | IRCAD | CHAOS | BTCV | Internal | ||
Cases (total) | 187 | 23 | 9 | 8 | 10 | 56 |
Cases in arterial phase |
85 | 0 | 0 | 0 | 0 | 31 |
Cases in venous phase |
102 | 23 | 9 | 8 | 10 | 25 |
Cases with AAA | 100 | 0 | 1 | 0 | 1 | 56 |
Voxel resolution X/Y | 0.9 ± 0.1 mm | 0.8 ± 0.1 mm | 0.7 ± 0.1 mm | 0.7 ± 0.0 mm | 0.7 ± 0.1 mm | 0.8 ± 0.1 mm |
Voxel resolution Z | 1.5 ± 0.5 mm | 1.0 ± 0.0 mm | 1.6 ± 0.8 mm | 1 ± 0.4 mm | 3 ± 0.6 mm | 1.5 ± 0.6 mm |
Slices [min, max] | [101, 2687] | [187, 310] | [79, 260] | [95, 266] | [42, 148] | [101, 1323] |
Data set statistics—resolution is given as median ± standard deviation.
For validation of the algorithm, we collected an additional verification data set from four publicly available sources (TCIA [22], IRCAD [23], CHAOS [24], BTCV [25]) and from our RIS as for the first data set. We excluded volumes from the public data, which do not cover the full abdomen, resulting in a total number of 50 cases (TCIA: 23, IRCAD: 9, CHAOS: 8, BTCV: 10). These scans originate from a broad spectrum of hardware of different manufacturers (Siemens Healthineers, Erlangen, Germany; Philips, Amsterdam, The Netherlands; Canon Medical Systems, Otawara, Japan). A total of 56 new cases were extracted from our PACS. Each CT scan was again assigned to one of the two classes (0: no AAA, 1: AAA). No further annotation was performed.
2.2. Networks, Preprocessing and Training
We extended three established architectures from literature to be applied to 3D image classification: AlexNet [26], VGG-16 [27] and ResNet [28]. We altered the described architectures to economize the memory consumption. All networks use the ReLU activation function, except for the last layer, where softmax is used. The architectures are shown in Figure 2.
We use stratified 5-fold cross-validation on the initial training data set. The data are split into five disjoint test sets, each consisting of 36–39 cases. For each fold, one of these test sets is used and 6 cases are selected for validation from the non-test data. The remaining non-test cases are used for training. The 3D patches of 320 × 384 × 224 voxels are extracted from the CT volumes. This reduces memory requirements and removes the parts of the image depicting air and the patient table. For prediction, the patch is centered on the center of the anchor point slice. The densities of the CT images are windowed to range between [−200HU,400HU], thus covering an extended soft tissue range, and then mapped to the interval [−1,1]. All images are resampled to a spacing of 0.9 × 0.9 × 1.5 mm, which equals the median resolution of the data set.
During training, the anchor point slice is shifted by up to ten voxels along the cranio-caudal axis and the patch center is randomly positioned on the slice. Furthermore, data augmentation is also performed via rotation around the cranio-caudal axis with angle α ∈ [−12.6°,12.6°] to mimic possible patient positions and scaling by up to 10% to mimic possible patient sizes. Additionally, density jittering by up to ± 3 HU is applied. Details of the implementation and training of the networks are provided in Appendix B and Appendix C.
2.3. Layer-Wise Relevance Propagation
Layer-wise relevance propagation (LRP) allows the calculation of voxel-wise decomposition of the decision of a CNN. It can thus be used to provide interpretability for CNNs [29]. The relevance is propagated layer-wise from the network output back through the network, until the input layer is reached. The result equals a relevance map, which provides a relevance value on the output class for each single input value. In our maps, positive relevance values indicate a relevance for the AAA class, while negative values indicate a relevance against the AAA class. We normalize the maps by their sum to standardize them [30,31].
2.4. Fully Automatic Screening
To apply the screening algorithm fully automatically, an automatic abdominal image region extraction was developed. The underlying algorithm first analyzes the HU distribution along the z-axis in the HU value range of soft tissue to determine a lower and upper bound of the abdomen. The abdomen center is then determined based on the distribution of high HU values, corresponding to the area between the top of the hip bone and the lower ribs. A subvolume is then extracted based on the abdomen center and fed to the network for classification. Using this fully automatic screening, the 106 cases from the second data set were processed.
2.5. Evaluation
We employ five metrics to assess the quality of the predicted classification. We use a discrimination threshold (DT) of 0.5 for all four binary metrics. They are derived from four outcomes: true positive (TP) represents a sample correctly identified as AAA, true negative (TN) denotes a non-AAA sample correctly classified as such, false positive (FP) is an AAA sample falsely identified as a non-AAA case and false negative (FN) denotes a non-AAA sample being misclassified as AAA. From this, we calculate the accuracy (A), the precision (P), the true positive rate (TPR), the false positive rate (FPR) and the F1 score (F1) [32]. These measures compare the predictions of the networks to the labels. The receiver operating characteristic (ROC) curve plots the ratio between TP and FP decisions, when the discrimination threshold of a binary classifier is varied. It evaluates the performance based on the predicted probabilities of the networks and the labels. The area under the curve (AUC) provides a metric for the classifier performance.
To assess the correspondence of the relevance maps with the aorta we use a 5-point Likert scale, which is used to measure the agreement between the decision relevant region of the algorithm with the aortic location determined by a human radiologist [33]. Positive and negative relevance values were considered equally for the assessment. The influence of the aorta on the network decision is scored as: 1 (no relevance in the aorta), 2 (low relevance in the aorta), 3 (medium relevance in the aorta), 4 (high relevance in the aorta) and 5 (very high relevance in the aorta).
3. Results
To assess the performance of the CNN classification we performed four experiments. Firstly, we trained and tested three architectures. This was only performed for one-fold of the initial data set. Secondly, LRP is applied to the best network from Experiment 1 to validate that the network decision is based on the correct region of interest (ROI). Relevance maps were scored by a radiologist with 2.5 years of experience in vascular imaging. Thirdly, the training of the best network is repeated for the four remaining folds to verify repeatability. In the fourth step, fully automatic screening using the best network is tested on the second verification data set.
3.1. Results of the First Experiment (Network Comparison)
The results of the evaluation metrics for the first experiment are listed in Table 2. The ROC curves for the three networks are shown in Figure 3. The 3D ResNet achieved the best results according to all five metrics. Out of the 3D ResNet’s five false positive cases, two cases showed a pre-aneurysmatic enlargement.
Table 2.
Network | A | P | TPR | F 1 | AUC |
---|---|---|---|---|---|
3D AlexNet | 0.769 | 0.773 | 0.810 | 0.791 | 0.899 |
3D VGG | 0.769 | 0.800 | 0.762 | 0.780 | 0.860 |
3D ResNet | 0.872 | 0.808 | 1.000 | 0.894 | 0.931 |
Experiment 1: Quantitative comparison of the classification quality of three different networks for one-fold. The best result for each metric is marked bold. A = accuracy; P = precision; TPR = true positive rate; F1 = F1 score; AUC = area under the curve.
3.2. Results of the Second Experiment (LRP Maps)
In the second experiment, the rating of the relevance maps resulted in an average score of 4.56 for correctly classified cases. The exact distribution can be seen in Figure 4c. Relevance maps for a case with and a case without AAA are shown exemplarily in Figure 4a,b, respectively. For the AAA class, the highest relevance values were present on the inner lumen of the aneurysm. There was one case where no relevance was present in the aorta. This patient had an extracorporeal membrane oxygenation (ECMO) tube placed in the vena cava. The relevance analysis showed that the network focused on the inferior vena cava instead of the aorta.
3.3. Results of the Third Experiment (Repeated Training of the Best Network) and the Fourth Experiment (Verification on a Additional Dataset)
The results from the third experiment-the evaluation metrics for the five folds and across all data (aggregated from the five folds) are listed in Table 3. Its performance on the verification dataset (fourth experiment) is listed in Table 4. Visual inspection of the results of the automated region extraction confirmed that a suitable subvolume was extracted for all cases in the verification dataset. The extraction had an average computation time of 2 s.
Table 3.
Fold | A | P | TPR | F1 | AUC |
---|---|---|---|---|---|
1 | 0.872 | 0.808 | 1.000 | 0.894 | 0.931 |
2 | 0.821 | 0.850 | 0.810 | 0.829 | 0.929 |
3 | 0.838 | 0.850 | 0.850 | 0.850 | 0.919 |
4 | 0.944 | 0.947 | 0.947 | 0.947 | 0.961 |
5 | 0.806 | 0.773 | 0.895 | 0.829 | 0.885 |
All | 0.856 | 0.841 | 0.900 | 0.870 | 0.926 |
Experiment 3: quantitative comparison of the classification performance of 3D ResNet across 5 folds. Last line shows the results computed across all test folds. A = accuracy; P = precision; TPR = true positive rate; F1 = F1 score; AUC = area under the curve.
Table 4.
Fold | A | P | TPR | F1 | AUC |
---|---|---|---|---|---|
All | 0.953 | 0.949 | 0.966 | 0.957 | 0.971 |
Experiment 4: quantitative analysis of the classification performance of 3D ResNet on the verification data set.
4. Discussion
The findings of this study show that our 3D ResNet demonstrated a high performance and robustness for fully automated AAA detection in abdominal CT scans. Our proposed network achieves an AUC of 0.971 and an accuracy of 0.953 on the verification data set. Based on these results, the architecture seems to be suitable for clinical screening purposes.
Ultrasound-based AAA screening has been described to decrease AAA mortality and increase effectiveness of treatment and has partly been introduced into clinical practice [34,35]. Automated screening for AAA on abdominal CT scans that have been acquired for various other reasons would add another chance of early detection of AAA, possibly supporting radiologists and clinicians in reporting, monitoring and the treatment of AAA [7].
Time-consuming manual annotation remains a major bottleneck of the training and validation of newly developed algorithms on large clinical data sets [36,37]. In contrast, the screening algorithm presented here was designed to be trainable on minimally annotated data. Setting of the anchor point was automated for the fully automated screening on the verification data via a heuristic approach but could also be realized by anatomical landmark detection [38]. Such automatization provides the opportunity for future training and validation on much larger data sets. A critical factor for the successful implementation of artificial intelligence systems into hospital workflows besides factors such as integration into the local standard operating procedures and clinical systems is a high robustness and reliability to establish the necessary trust into AI systems [39]. The LRP analysis presented here is able to generate a graphical analysis of the decision relevant areas of the network in the images and therefore, could be used for validation. PACS export of these results might significantly increase its clinical acceptance. Detailed effects on patient treatment remain unanswered and represent an exciting field of further research.
Different groups have already described pipelines and tools for detection and classification of AAA in CT scans. A general shift of focus towards compatibility with routine clinical implementation can be observed, however, this process remains challenging [11,20,40,41,42]. Different levels of performance have been reported for AAA related image processing tasks. Mohammadi et al. reported a high accuracy of successfully classifying 2D patches to show the aorta (0.986) [11]. For the prediction of an aortic dissection and aortic rupture using a 2D CNN a TPR of 0.900 and 0.889 as well as an AUC of 0.979 and 0.990 were reported by Harris et al., respectively [17]. Our 3D ResNet achieves a similar TPR and a comparable AUC value to these works. In general, the task of classifying an entire 3D volume is more complex than classifying selected 2D patches and we targeted a different problem. To the best of our knowledge, and based on our extensive literature research, the application of 3D CNNs to CT AAA screening has not been presented previously. Solutions which apply 3D CNNs for whole CT binary classification of other pathologies in CT data report accuracies of 0.918 (lung cancer [43]) and 0.93 (COVID-19 [44]).
It is important to note that the results achieved with the algorithm proposed in this study were derived from training on a heterogeneous data set of 187 images, which includes different contrast phases, scanners, and artifacts. Mohammadi et al. trained and tested their algorithm on a data set of solely ten patients with two patients having an “obvious” AAA [11]. Whereas in one of our previous studies [45], metallic interferences have been shown to negatively impact the CNN performance, in this study the accuracy remained high even with presence of these artifacts, most likely due to the inclusion of these in the training data set and improvements in data augmentation during training. Moreover, our results show—though only using a small dataset—that a robust algorithm could be trained that achieved an even higher accuracy on a differently composed unseen second dataset. Additional validation on a mixed dataset of external and internal abdominal CT scans on the one hand introduces the risk of verification on a dataset which does not entirely reflect clinical reality. The external data contained fewer metal artifacts and the AAA cases were often more simplistic cases. On the other hand, inclusion of these additional cases also increases CT data variability originating from international cohorts and different manufacturer hardware. It contributes to the understanding of the potential algorithm performance under different clinical conditions and the generalizability of the proposed method. The algorithm described here proved to be robust enough to achieve reliable results on the verification data set consisting of heterogeneous data from multiple sources.
The 3D ResNet, however, produced some FP results on the training data set. In two cases, early enlargements which were not aneurysmatic yet were detected. In another case, for one patient with an ECMO tube placed in the vena cava, the LRP revealed that the network focused on the inferior vena cava instead of the aorta. The classification result for this case was, however, correct. We assume that the ECMO tube was interpreted as an aortic stent by the network and therefore, caused the mislocated area of interest. On the verification data set, two FN cases were overserved. These cases showed very atypical shapes of the aneurysm where the aneurysmatic part was entirely thrombotic and the free-flown part had the shape of a regular aortic vessel. Further training on more data would probably solve this drawback. The aneurysm size and shape definition utilized for annotation seem critical for algorithm sensitivity, especially in small aneurysm.
4.1. Limitations
A limitation of our approach on the one hand is caused by the design of the algorithm as a binary classification task instead of object detection or segmentation. Screening could potentially also be solved with an object detection network, this would, however, require far more extensive annotations and is not guaranteed to provide a better performance. Our solution does not allow for automated segmentation and volumetry (which has already been achieved by other authors). While the very heterogenous datasets promote generalizability, the high differences in the slice thickness could also introduce a bias. However, in our approach, we interpolated the slice thickness as well as the in-plane resolution to the median resolution of the training data overcoming such problems. Future directions lie in the practical use of the technology developed within this study to notify clinicians in real-time. In addition, in cases of relevant AAA, a second pipeline could be triggered which includes fully automated AAA segmentation and analysis, similar to the one introduced by Lareyre et al. [41]. On the other hand, our study used a comparatively small patient cohort of 187 CT scans as a training data set, potentially limiting the dataset variability and thus, generalizability. However, the training was performed on a heterogeneous dataset derived from a large University Medical Center with a high variety of different findings. In addition, data augmentation methods were employed to simulate variations.
In summary, the training and additional validation on unfiltered clinical data sets mark a step forward towards clinical implementation. Further testing on a much larger international cohort seems paramount for integration into the workflow of different medical sites that might differ in various factors. Training of the developed methodology on a larger cohort would be possible with reasonable extra effort and therefore, should be performed to ensure a broad generalizability.
From a medical and clinical perspective, it will be exciting to investigate if earlier detection of small AAA contributes to a more effective patient treatment and ultimately improvement of prognosis. A possible prospective observer study monitoring the effects of the technology on patient care would be a logical and necessary next step. Besides the algorithm performance, the way of its clinical introduction represents a highly important factor influencing the clinical success and as a major remaining question should be defined more clearly in upcoming scientific works.
4.2. Conclusions
This study demonstrated the feasibility of CNN-based fully automated detection of AAA in an unselected clinical data set of CT images. The generation of relevance maps contributes to the explainability of the decision process. Integration of this deep learning screening for AAA into the routine workflow might lead to improved patient monitoring, earlier diagnosis, and improved patient treatment with possible reduction of rupture risk.
Acknowledgments
We gratefully acknowledge the support of NVIDIA Corporation with the donation of the NVIDIA Quadro P5000 used for this research.
Abbreviations and Acronyms
AAA | Abdominal aortic aneurysm |
AUC | Area under the curve |
CNN | Convolutional neural network |
CT | Computed tomography |
CTA | Computed tomography angiography |
DT | Discrimination threshold |
ECMO | Extracorporeal membrane oxygenation |
EVAR | Endovascular aneurysm repair |
LRP | Layer-wise relevance propagation |
PACS | Picture archiving and communication system |
ROC | Receiver operating characteristic |
ROI | Region of interest |
RIS | Radiology information system |
Appendix A. Data Aggregation and Dataset
For the data aggregation of the training dataset, keywords used to find suitable cases in the PACS were (translated): ”aortic aneurysm”, ”aneurysm”, ”abdominal aneurysm”, ”AAA”, ”no aneurysm”, ”no AAA” within the radiology report. This search strategy yielded a substantial number of reports confirming the presence of AAA and other reports denying presence of AAA. The scans for algorithm training were extracted from our picture archiving and communication system (PACS). The scans were derived from a total of seven different CT scanners (SOMATOM Force, SOMATOM Definition Flash, SOMATOM Definition AS 128, SOMATOM Sensation 64, SOMATOM Emotion 16, SOMATOM Emotion 16, Biograph mCT, all Siemens Healthineers, Erlangen, Germany) which were located at two different medical sites (being part of our department). Only contrast enhanced CT scans were included. Some scans were acquired 70 s post-contrast media application (venous), some in CTA contrast phase (arterial) and some in a mixed phase of 40–50 s post-application, which were classified as venous studies within this analysis.
Appendix B. Networks
Background information about the networks utilized: AlexNet was proposed by Krizhevsky et al. for image classification [26]. Its first five layers are convolutional layers with three max-pooling layers located in between. The final three layers are fully connected layers. In this work, we modified the network to make it suitable for 3D classification by reducing the number of filters and employing only two fully connected layers.
The VGG-16 architecture was introduced by Simonyan et al. [27]. In contrast to the AlexNet it uses multiple 3 × 3 convolutions replacing large kernel-sized filters. Our 3D implementation has a lower number of filters and layers. Additionally, the first convolutional layer uses a stride of two in every spatial direction.
He et al. developed a CNN with residual connections called ResNet [28]. The residual connections are formed by adding feature maps from the beginning of a block of convolutions to the final feature maps. Other than these connections, the ResNet also uses an average pooling instead of flattening the feature values before the first fully connected layer. The ResNet only uses max-pooling for downsizing in the first layer, subsequently downsizing is performed via strided convolutions. It also employs a batch-normalization [46]. Similarly, to the other networks, we adapted the number of channels and layers. We also increased the stride in the first layer to two.
The training was performed using the Adam optimizer with a learning rate of 10−3. We used the binary cross entropy loss and apply a L2 regularization on the network weights. The regularization term is weighted by 10−5. We applied an early stopping to the training process: after the initial 20 epochs, we tested for the convergence criterion at the end of each epoch. The criterion is defined as the change of validation accuracy over the last twenty epochs falling below 10−4. For the application we then selected the weights with the highest accuracy on the validation data from the last 20 epochs. The convergence of the networks is shown in Figure A1. The 3D ResNet converged after the highest number of epochs. The 3D AlexNet showed a low generalization capability during the training. The training time for each network was 34 to 57 h. The average prediction time was 3 s.
Appendix C. Implementation
The described networks were implemented using TensorFlow 1.12 [47] and Python 3.6. For the evaluation and the image processing we used SimpleITK 1.2.4, which provides a simplified interface to the Insight Toolkit (ITK) [48]. LRP was realized using the iNNvestigate Neural Networks! Toolbox 1.0.9 [49]. Training and testing were performed on a Windows Server 2016 with an Intel Core i7-7700K CPU, 64GB of RAM and a NVIDIA Quadro P5000 graphics card with 16 GB VRAM.
Author Contributions
Conceptualization, A.-K.G., S.J.D., F.G.Z. and J.S.R.; formal analysis, T.R. and D.F.B.; funding acquisition, S.J.D., S.O.S. and L.R.S.; investigation, T.R. and D.F.B.; methodology, A.-K.G. and J.S.R.; project administration, J.S.R.; resources, S.J.D., L.R.S. and F.G.Z.; software, A.-K.G.; supervision, S.J.D., S.O.S., M.K. and L.R.S.; validation, A.-K.G., M.F.F., S.J.D., S.O.S., M.K. and F.G.Z.; visualization, A.-K.G.; writing—original draft, A.-K.G. and J.S.R.; writing—review and editing, A.-K.G., C.T., T.R., D.F.B., M.F.F., S.J.D., S.O.S., M.K., L.R.S., F.G.Z. and J.S.R. All authors have read and agreed to the published version of the manuscript.
Funding
This research project is part of the Research Campus M2OLIE and funded by the German Federal Ministry of Education and Research (BMBF) within the Framework Forschungscampus: Public–Private Partnership for Innovations” under the funding Code 13GW0388A.
Institutional Review Board Statement
The study was conducted according to the guidelines of the Declaration of Helsinki and approved by the Institutional Review Board “Ethics Committee II” of the Medical Faculty Mannheim, Heidelberg University (2016-863R-MA, 17 November 2016).
Informed Consent Statement
Patient consent was waived due to the retrospective nature of data aggregation with no connection to clinical patientcare.
Data Availability Statement
The sources of the publicly available datasets used for composition of the verification dataset are listed in the references [22,23,24,25]. The data retrieved from our own clinical systems used for training and partly verification presented in this study is not publicly available due to patient privacy concerns, publication would not be covered by IRB statement.
Conflicts of Interest
The authors declare no conflict of interest.
Footnotes
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.
References
- 1.Nordon M., Hinchliffe R.J., Loftus I.M., Thompson M.M. Pathophysiology and epidemiology of abdominal aortic aneurysms. Nat. Rev. Cardiol. 2011;8:92–102. doi: 10.1038/nrcardio.2010.180. [DOI] [PubMed] [Google Scholar]
- 2.Chaikof E.L., Dalman R.L., Eskandari M.K., Jackson B.M., Lee W.A., Mansour M.A., Mastracci T.M., Mell M., Murad M.H., Nguyen L.L., et al. The Society for Vascular Surgery practice guidelines on the care of patients with an abdominal aortic aneurysm. J. Vasc. Surg. 2018;67:2–77.e2. doi: 10.1016/j.jvs.2017.10.044. [DOI] [PubMed] [Google Scholar]
- 3.Turton E., Scott D., Delbridge M., Snowden S., Kester R. Ruptured Abdominal Aortic Aneurysm: A Novel Method of Outcome Prediction Using Neural Network Technology. Eur. J. Vasc. Endovasc. Surg. 2000;19:184–189. doi: 10.1053/ejvs.1999.0974. [DOI] [PubMed] [Google Scholar]
- 4.Schmitz-Rixen T., Keese M., Hakimi M., Peters A., Böckler D., Nelson K., Grundmann R.T. Ruptured abdominal aortic aneurysm—epidemiology, predisposing factors, and biology. Langenbeck’s Arch. Surg. 2016;401:275–288. doi: 10.1007/s00423-016-1401-8. [DOI] [PubMed] [Google Scholar]
- 5.Peters A.S., Hakimi M., Erhart P., Keese M., Schmitz-Rixen T., Wortmann M., Bischoff M.S., Böckler D. Current treatment strategies for ruptured abdominal aortic aneurysm. Langenbeck’s Arch. Surg. 2016;401:289–298. doi: 10.1007/s00423-016-1405-4. [DOI] [PubMed] [Google Scholar]
- 6.Lindholt J.S., Søgaard R., Laustsen J. Prognosis of ruptured abdominal aortic aneurysms in Denmark from 1994–2008. Clin. Epidemiol. 2012;4:111–113. doi: 10.2147/CLEP.S31098. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 7.Claridge R., Arnold S., Morrison N., van Rij A.M. Measuring abdominal aortic diameters in routine abdominal computed tomography scans and implications for abdominal aortic aneurysm screening. J. Vasc. Surg. 2017;65:1637–1642. doi: 10.1016/j.jvs.2016.11.044. [DOI] [PubMed] [Google Scholar]
- 8.Oliver-Williams C., Sweeting M., Jacomelli J., Summers L., Stevenson A., Lees T., Earnshaw J.J. Safety of Men With Small and Medium Abdominal Aortic Aneurysms Under Surveillance in the NAAASP. Circulation. 2019;139:1371–1380. doi: 10.1161/CIRCULATIONAHA.118.036966. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 9.Salata K., Hussain M.A., de Mestral C., Greco E., Aljabri B.A., Mamdani M., Forbes T.L., Bhatt D.L., Verma S., Al-Omran M. Comparison of outcomes in elective endovascular aortic repair vs open surgical repair of abdominal aortic aneurysms. JAMA Netw. Open. 2019;7:2. doi: 10.1001/jamanetworkopen.2019.6578. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 10.Lundervold S., Lundervold A. An overview of deep learning in medical imaging focusing on mri. Z. Med. Phys. 2019;29:102–127. doi: 10.1016/j.zemedi.2018.11.002. [DOI] [PubMed] [Google Scholar]
- 11.Mohammadi S., Mohammadi M., Dehlaghi V., Ahmadi A. Automatic segmentation, detection, and diagnosis of abdominal aortic aneurysm (aaa) using convolutional neural networks and hough circles algorithm. Cardiovasc. Eng. Technol. 2019;10:490–499. doi: 10.1007/s13239-019-00421-6. [DOI] [PubMed] [Google Scholar]
- 12.López-Linares K., Aranjuelo N., Kabongo L., Maclair G., Lete N., Ceresa M., García-Familiar A., Macía I., González Ballester M.A. Fully automatic detection and segmentation of abdominal aortic thrombus in post-operative cta images using deep convolutional neural networks. Med. Image Anal. 2018;46:202–214. doi: 10.1016/j.media.2018.03.010. [DOI] [PubMed] [Google Scholar]
- 13.Habijan M., Galić I., Leventić H., Romić K., Babin D. Abdominal aortic aneurysm segmentation from ct images using modified 3d u-net with deep supervision; Proceedings of the IEEE 2020 International Symposium ELMAR; Zadar, Croatia. 14–15 September 2020; pp. 123–128. [Google Scholar]
- 14.Zhang L., Jiang Z., Choi J., Lim C.Y., Maiti T., Baek S. Patient-Specific Prediction of Abdominal Aortic Aneurysm Expansion Using Bayesian Calibration. IEEE J. Biomed. Heal. Inform. 2019;23:2537–2550. doi: 10.1109/JBHI.2019.2896034. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15.Do H.N., Ijaz A., Gharahi H., Zambrano B., Choi J., Lee W., Baek S., Choi J. Prediction of Abdominal Aortic Aneurysm Growth Using Dynamical Gaussian Process Implicit Surface. IEEE Trans. Biomed. Eng. 2018;66:609–622. doi: 10.1109/TBME.2018.2852306. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 16.García G., Maiora J., Tapia A., De Blas M. Evaluation of Texture for Classification of Abdominal Aortic Aneurysm after Endovascular Repair. J. Digit. Imaging. 2012;25:369–376. doi: 10.1007/s10278-011-9417-7. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 17.Harris R.J., Kim S., Lohr J., Towey S., Velichkovich Z., Kabachenko T., Driscoll I., Baker B. Classification of aortic dissection and rupture on post-contrast ct images using a convolutional neural network. J. Digit. Imaging. 2019;32:939–946. doi: 10.1007/s10278-019-00281-5. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 18.Cao L., Shi R., Ge Y., Xing L., Zuo P., Jia Y., Liu J., He Y., Wang X., Luan S., et al. Fully automatic segmentation of type B aortic dissection from CTA images enabled by deep learning. Eur. J. Radiol. 2019;121:108713. doi: 10.1016/j.ejrad.2019.108713. [DOI] [PubMed] [Google Scholar]
- 19.Hahn S., Perry M., Wshah S., Morris C.S., Bertges D.J. SS26. Machine Deep Learning Accurately Detects Endoleak After Endovascular Abdominal Aortic Aneurysm Repair. J. Vasc. Surg. 2019;69:e202–e203. doi: 10.1016/j.jvs.2019.04.291. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 20.Hong H.A., Sheikh U.U. Automatic detection, segmentation and classification of abdominal aortic aneurysm using deep learning; Proceedings of the 2016 IEEE 12th International Colloquium on Signal Processing & Its Applications (CSPA); Melaka, Malaysia. 4–6 March 2016; pp. 242–246. [Google Scholar]
- 21.Maiora J., Graña M. A Hybrid Segmentation of Abdominal CT Images. In: Corchado E., Snášel V., Abraham A., Woźniak M., Graña M., Cho S.B., editors. Hybrid Artificial Intelligent Systems. Vol. 7209. Springer; Berlin, Heidelberg, Germany: 2012. [DOI] [Google Scholar]
- 22.Roth H., Farag A., Turkbey E.B., Lu L., Liu J., Summers R.M. The Cancer Imaging Archive. The University of Arkansas for Medical Sciences; Little Rock, AR, USA: 2016. [(accessed on 1 November 2021)]. Data From Pancreas-CT. Available online: https://wiki.cancerimagingarchive.net/display/Public/Pancreas-CT#225140400c26eab54502412cbbd0e1c0fddd917b. [DOI] [Google Scholar]
- 23.Soler L., Hostettler A., Agnus V., Charnoz A., Fasquel J.B., Moreau J., Osswald A.B., Bouhadjar M., Marescaux J. 3D Image Reconstruction for Comparison of Algorithm Database: A Patient Specific Anatomical and Medical Image Database. Les Hôpitaux Universitaires; Strasbourg, France: 2010. [(accessed on 21 October 2021)]. Available online: https://www.ircad.fr/fr/recherche/3d-ircadb-01-fr/ [Google Scholar]
- 24.Kavur E., Selver M.A., Dicle O., Barış M., Gezer N.S. CHAOS-Combined (CT-MR) Healthy Abdominal Organ Segmentation Challenge Data (Version v1.03) [Data Set] Zenodo, European Organization for Nuclear Research and OpenAIRE; Genève, Switzerland: 2019. [(accessed on 21 October 2021)]. Available online: [DOI] [Google Scholar]
- 25.Landman A., Xu Z., Igelsias J.E., Styner M., Langerak T.R., Klein A. MICCAI Multi-Atlas Labeling beyond the Cranial Vault-workshop and Challenge. Synapse, Sage Bionetworks; Seattle, WA, USA: 2015. [(accessed on 21 October 2021)]. Available online: [DOI] [Google Scholar]
- 26.Krizhevsky A., Sutskever I., Hinton G.E. Imagenet Classification with Deep Convolutional Neural Networks. In: Pereira F., Burges C.J.C., Bottou L., Weinberger K.Q., editors. Proceedings of the NIPS 2012; Lake Tahoe, NV, USA. 3–6 December 2012; San Diego, CA, USA: Curran Associates Inc.; 2012. pp. 1097–1105. [Google Scholar]
- 27.Simonyan K., Zisserman A. Very deep convolutional networks for large-scale image recognition. In: Bengio Y., LeCun Y., editors. Proceedings of the 3rd International Conference on Learning Representations, ICLR 2015; San Diego, CA, USA. 7–9 May 2015; La Jolla, CA, USA: International Conference on Representation Learning; 2015. [(accessed on 21 October 2021)]. Available online: https://dblp.org/db/conf/iclr/iclr2015.html. [Google Scholar]
- 28.He K., Zhang X., Ren S., Sun J. Deep residual learning for image recognition; Proceedings of the CVPR 2016; Los Alamitos, CA, USA. June 2016; Washington, DC, USA: IEEE Computer Society; 2016. pp. 770–778. [Google Scholar]
- 29.Bach S., Binder A., Montavon G., Klauschen F., Müller K.-R., Samek W. On pixel-wise explanations for non-linear classifier decisions by layer-wise relevance propagation. PLoS ONE. 2015;10:e0130140. doi: 10.1371/journal.pone.0130140. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 30.Chlebus G., Abolmaali N., Schenk A., Meine H. Relevance analysis of mri sequences for automatic liver tumor segmentation; Proceedings of the MIDL 2019; London, UK. 8–10 July 2019. [Google Scholar]
- 31.Schnurr A.-K., Schöben M., Hermann I., Schmidt R., Chlebus G., Schad L.R., Zöllner F.G. Relevance analysis of mri sequences for ms lesion detection; Proceedings of the 2020 ESMRMB Congress; Online. 2020. pp. 77–78. [Google Scholar]
- 32.Hossin M., Sulaiman M. A review on evaluation metrics for data classification evaluations. Int. J. Data Min. Knowl. Manag. Process. 2015;5:1. [Google Scholar]
- 33.Likert R. A technique for the measurement of attitudes. Arch. Psychol. 1932;22:55. [Google Scholar]
- 34.Bains P., Oliffe J.L., Mackay M.H., Kelly M.T. Screening older adult men for abdominal aortic aneurysm: A. scoping review. Am. J. Mens Health. 2021;15:15579883211001204. doi: 10.1177/15579883211001204. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 35.Sweeting M., Marshall J., Glover M., Nasim A., Bown M.J. Evaluating the cost-effectiveness of changes to the surveillance intervals in the uk abdominal aortic aneurysm screening programme. Value Health. 2021;24:369–376. doi: 10.1016/j.jval.2020.10.015. [DOI] [PubMed] [Google Scholar]
- 36.Yasaka K., Abe O. Deep learning and artificial intelligence in radiology: Current applications and future directions. PLoS Med. 2018;15:e1002707. doi: 10.1371/journal.pmed.1002707. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 37.Saba L., Biswas M., Kuppili V., Cuadrado Godia E., Suri H.S., Edla D.R., Omerzu T., Laird J.R., Khanna N.N., Mavrogeni S., et al. The present and future of deep learning in radiology. Eur. J. Radiol. 2019;114:14–24. doi: 10.1016/j.ejrad.2019.02.038. [DOI] [PubMed] [Google Scholar]
- 38.Ghesu F., Georgescu B., Zheng Y., Grbic S., Maier A., Hornegger J., Comaniciu D. Multi-scale deep reinforcement learning for real-time 3dlandmark detection in ct scans. IEEE Trans. Pattern Anal. Mach. Intell. 2019;41:176–189. doi: 10.1109/TPAMI.2017.2782687. [DOI] [PubMed] [Google Scholar]
- 39.Schmuelling L., Franzeck F.C., Nickel C.H., Mansella G., Bingisser R., Schmidt N., Stieltjes B., Bremerich J., Sauter A.W., Weikert T., et al. Deep Learning-based Automated Detection of Pulmonary Embolism on CT Pulmonary Angiograms: No Significant Effects on Report Communication Times and Patient Turnaround in the Emergency Department Nine Months after Technical Implementation. Eur. J. Radiol. 2021;141:109816. doi: 10.1016/j.ejrad.2021.109816. [DOI] [PubMed] [Google Scholar]
- 40.Raffort J., Adam C., Carrier M., Ballaith A., Coscas R., Jean-Baptiste E., Hassen-Khodja R., Chakf´e N., Lareyre F. Artificial intelligence in abdominal aortic aneurysm. J. Vasc. Surg. 2020;72:321–333. doi: 10.1016/j.jvs.2019.12.026. [DOI] [PubMed] [Google Scholar]
- 41.Lareyre F., Adam C., Carrier M., Dommerc C., Mialhe C., Raffort J. A fully automated pipeline for mining abdominal aortic aneurysm using image segmentation. Sci. Rep. 2019;9:13750. doi: 10.1038/s41598-019-50251-8. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 42.Hirata K., Nakaura T., Nakagawa M., Kidoh M., Oda S., Utsunomiya D., Yamashita Y. Machine learning to predict the rapid growth of small abdominal aortic aneurysm. J. Comput. Assist. Tomogr. 2020;44:37–42. doi: 10.1097/RCT.0000000000000958. [DOI] [PubMed] [Google Scholar]
- 43.Polat H., Danaei Mehr H. Classification of pulmonary ct images by using hybrid 3d-deep convolutional neural network architecture. Appl. Sci. 2019;9:940. doi: 10.3390/app9050940. [DOI] [Google Scholar]
- 44.Singh D., Kumar V., Kaur M. Classification of covid-19 patients from chest ct images using multi-objective differential evolution–based convolutional neural networks. Eur. J. Clin. Microbiol. Infect. 2020;39:1–11. doi: 10.1007/s10096-020-03901-z. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 45.Golla K., Bauer D.F., Schmidt R., Russ T., Nörenberg D., Chung K., Tönnes C., Schad L.R., Zöllner F.G. Convolutional neural network ensemble segmentation with ratio-based sampling for the arteries and veins in abdominal ct scans. IEEE Trans. Biomed. Eng. 2020;68:1518–1526. doi: 10.1109/TBME.2020.3042640. [DOI] [PubMed] [Google Scholar]
- 46.Ioffe S., Szegedy C. Batch normalization: Accelerating deep network training by reducing internal covariate shift; Proceedings of the ICML 2015; Lille, France. 6–11 July 2015; pp. 448–456. [Google Scholar]
- 47.Abadi M., Agarwal A., Barham P., Brevdo E., Chen Z., Citro C., Corrado G.S., Davis A., Dean J., Devin M., et al. TensorFlow: Large-Scale Machine Learning on Heterogeneous Systems. 2015. [(accessed on 21 October 2021)]. Available online: Tensorflow.Org.
- 48.Lowekamp B.C., Chen D.T., Ibáñez L., Blezek D. The design of simpleitk. Front. Neuroinform. 2013;7:45. doi: 10.3389/fninf.2013.00045. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 49.Alber M., Lapuschkin S., Seegerer P., Hägele M., Schütt K.T., Montavon G., Samek W., Müller K.-R., Dähne S., Kindermans P.-J. Innvestigate neural networks! J. Mach. Learn. Res. 2019;20:1–8. [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Data Availability Statement
The sources of the publicly available datasets used for composition of the verification dataset are listed in the references [22,23,24,25]. The data retrieved from our own clinical systems used for training and partly verification presented in this study is not publicly available due to patient privacy concerns, publication would not be covered by IRB statement.