Abstract
In addition to lung cancer, other thoracic abnormalities, such as emphysema, can be visualized within low-dose CT scans that were initially obtained in cancer screening programs, and thus, opportunistic evaluation of these diseases may be highly valuable. However, manual assessment for each scan is tedious and often subjective, thus we have developed an automatic, rapid computer-aided diagnosis system for emphysema using attention-based multiple instance deep learning and 865 LDCTs. In the task of determining if a CT scan presented with emphysema or not, our novel Transfer AMIL approach yielded an area under the ROC curve of 0.94 ± 0.04, which was a statistically significant improvement compared to other methods evaluated in our study following the Delong Test with correction for multiple comparisons. Further, from our novel attention weight curves, we found that the upper lung demonstrated a stronger influence in all scan classes, indicating that the model prioritized upper lobe information. Overall, our novel Transfer AMIL method yielded high performance and provided interpretable information by identifying slices that were most influential to the classification decision, thus demonstrating strong potential for clinical implementation.
Subject terms: Computed tomography, Translational research
Introduction
In 2021, the US Preventative Services Task Force (USPSTF) expanded eligibility requirements for lung cancer screening due to programs such as the International-Early Lung Cancer Action Program (I-ELCAP) and the National Lung Screening Trial (NLST)1–3. These initiatives found improved patient outcomes through lung cancer screening via low-dose CT (LDCT) acquisition, which provides a clinically useful image at reduced radiation dose to the patient than a standard dose CT scan. In addition to lung cancer, other thoracic abnormalities can be visualized within the CT scan range and opportunistic evaluation of these diseases may be highly valuable4,5. Emphysema can be identified on CT scans and shares risk factors with lung cancer, and current literature suggests that presence of emphysema may increase risk of lung cancer development6–11. Thus, emphysema evaluation on LDCT is appropriate and desirable. However, a manual assessment for each scan is tedious and often subjective, thus an automatic, rapid computer-aided diagnosis system should be investigated12,13.
Multiple instance learning (MIL) is a deep learning scheme commonly used in digital pathology that utilizes weak annotations to train models by evaluation of instances (e.g., CT slices) to form a collective classification decision of a bag (e.g., CT scan)14. Wang discussed key MIL schemes, mi-Net and MI-Net, which classify scans based on individual instance classifications and pooled instance representations, respectively15. Ilse improved MIL schemes through attention-based multiple instance learning, which utilizes attention mechanisms to identify and more heavily weight key instances of whole slide images for cancer detection16.
Deep learning, including MIL schemes, have been utilized to automate emphysema evaluation in standard diagnostic and lung screening CT scans. Humphries utilized a convolutional neural network and long short-term memory architecture to classify visual emphysema pattern on CT and Oh used the same model to compare visual emphysema progression with functional impairment and mortality17,18. Negahdar automatically segmented lung volumes on chest CT and classified patches of lung tissue based on visual emphysema pattern to quantify severity19. Chepylgina and Orting utilized human-engineered features based on histogram features acquired from filtered lung ROIs in a multiple instance learning scheme to characterize COPD and emphysema, respectively, in low-dose CT scans20,21. Tennakoon expanded their work to incorporate deep MIL on 3D LDCT patches to classify emphysema presence22.
In our work, we utilize deep MIL with transfer learning and attention-based pooling (Transfer AMIL) to evaluate emphysema in LDCT scans and compare performances in classification of disease.
Methods
We utilized MIL to characterize emphysema at LDCT presentation through the use of convolutional neural networks (CNN) with transfer learning. In this study, the standard MIL terms “instance” and “bag” are used synonymously with CT slice and scan, respectively.
LDCT imaging data
The data in this retrospective, HIPAA-compliant study consisted of 865 LDCT scans obtained as part of the International-Early Lung Cancer Action Program. IRB approval and informed consent was waived by the University of Chicago Biological Sciences Division/University of Chicago Medical Center IRB due to deidentification of images prior to obtention and confirm that all experiments were performed in accordance with relevant guidelines and regulations. The image selection criteria were as follows: using the database of all participants enrolled in the Early Lung Cancer Action Program at Weill Cornell Medical College and the Icahn School of Medicine at Mount Sinai between 1992 and July 2017, the most recent LDCT scan of 865 participants with LDCT acquisition were selected that included all participants with moderate emphysema and severe emphysema, and randomly selected the remaining participants among 3696 available participants with no emphysema. Emphysema was evaluated by visual scoring according to previously published criteria which, in brief, categorizes CT scans into mild/moderate categories if less than one-half of the lung volume presents with regions of decreased attenuation and severe if more than one-half has decreased attenuation23. The scans were non-gated, non-contrast enhanced, and acquired following the parameters and patient information given in Table 1. A radiologist further characterized the emphysema by dominant phenotype (centrilobular, panlobular, paraseptal) and other pulmonary CT findings non-specific to emphysema utilizing the Fleischner criteria24. Imaging data is currently not available publicly.
Table 1.
Number of cases | 865 (5 excluded from 865) |
Dates of acquisition | 1997–1999 (8) |
2000–2004 (161) | |
2005–2009 (192) | |
2010–2014 (128) | |
2015–2017 (326) | |
NA (50) | |
Sex at birth | Male (384) female (431) NA (50) |
Smoking status | Current (257) former (469) never (89) NA (50) |
Age | Mean (66.8) SD (11.4) range (33–93) |
Pack-years of smoking | Mean (36.2) SD (30.8) range (0–199) |
Scanner manufacturer | GE medical systems siemens |
Exposure time | Range (250–2100) |
kVp | (100, 120, 140) |
Slice thickness | Range (0.5–10 mm) |
Emphysema severity | None (508) mild/moderate (240) severe (117) |
Dominant emphysema phenotype | Centrilobular (284) paraseptal (33) panlobular (40) |
Multiple instance learning (MIL)
Typically, MIL is posed as a binary classification problem in which the data are composed into bags each of which is composed of N instances 15,16. The corresponding instance truths are unknown, but the bag truth is determined from the instance truths by the binary decision rule
MIL can be broken down into three key steps as: (1) Extraction of instance representations, (2) transformation from instance representations to bag representation through MIL pooling, and (3) classification of bag representation for clinically relevant decision25. In all, the process is described by
where is the predicted bag label, is the set of input CT slices (images) that are transformed to instance representations via , pooled via matrix , and transformed to a bag prediction via 25.
Transfer learned instance feature extraction (transfer MIL)
In our study, instant representations of CT slices are acquired through transfer learning from a pre-trained VGG19 architecture26. Transfer learning utilizes large models with deep, hierarchical features after pre-training for a similar task, in this case image classification but on the ImageNet database set of natural objects27–29. In situations where little training data are available, transfer learning allows for the extraction of more complex, rich data representations than can be achieved by training a model from scratch. In this study, we utilized a VGG-19 architecture pre-trained for natural image classification on ImageNet to extract quantitative features similar to the scheme proposed by Antropova30.
The instance representations were then input to two fully connected layers with ReLU activation with a dropout rate of 0.5.
Attention-based MIL pooling (AMIL)
Attention mechanisms have been widely utilized in deep learning to both improve performance and provide interpretability of model predictions31. In our study, the pooling matrix P was constructed through the MIL attention mechanism in which a bag representation was acquired through a weighted average of instance representations:
for learned parameters and with N input instances with dimension 512 and hidden dimension 128. The attention weights also provided interpretable output inherent to the decision task in the form of influential instances (i.e., slices), which were evaluated separately for model validation and interpretability.
The attention weights for different scan classes (dominant emphysema phenotypes of centrilobular, panlobular, and paraseptal) were evaluated by scaling attention weights for a given scan to the range [0, 1] and plotting as a function of the axial depth to determine regions of high and low influence. Influence was quantified by three metrics: (1) depth maximum attention of fit curve, (2) weighted average of slice depths weighted by attention, and (3) range of fit curve attention values. The full workflow of Transfer AMIL is provided in Fig. 1.
Implementation details
All models were trained in Keras (2.2.4) with Tensorflow backend (2.2.0) in Python (3.7) and optimized by binary cross entropy loss calculated for bag predictions. Adam optimization was utilized with parameters β1 = 0.9 and β2 = 0.99 and initial learning rate of 0.0001. Early stopping was initiated if the validation loss did not improve after 7 epochs. All learned parameters were initialized by sampling a normal distribution.
Training, testing, and statistical evaluation
Models were trained with 70%, 10%, and 20% of the available cases serving for training, validation, and testing, respectively, repeated 5 times with different randomly generated splits. The mean and variance of the area under the ROC curve (AUC) were obtained across the five models. AUCs were compared through the Delong test on each of the five training passes with the median p-value serving as the metric for significance32.
Compared methods
We compared Transfer AMIL to other approaches which required only scan annotations. A 3D CNN classifier was trained by interpolating to a fixed input size of 128 slices and scan presence of emphysema serving as binary class. Additionally, a standard 2D classifier was trained by assigning the scan ground truth class to all slices within the scan regardless of emphysema presence within that slice; this caused noisy labels during training, particularly with many false positive slices for severe emphysema cases.
Results
Binary classification performance
In the task of determining if a CT scan presented with emphysema or not, the Transfer AMIL approach yielded an area under the ROC curve of 0.94 ± 0.04, which was a statistically significant improvement compared to other methods evaluated in our study following the Delong Test with correction for multiple comparisons (Table 2). Transfer AMIL performed better than or similar to other published work, including shallow, human-engineered MIL methods, as well as other deep MIL approaches, although it is important to note that others’ evaluations were on different datasets.
Table 2.
Algorithm | AUC from ROC analysis | Human-engineered features | Deep CNN features | Interpretable | Transfer learning |
---|---|---|---|---|---|
Transfer AMIL | 0.94 ± 0.04 | X | X | X | |
Noisy 2D Classifier | 0.85 ± 0.06 | X | |||
Fully 3D Classifier | 0.58 ± 0.16 | X | |||
AMIL | 0.69 ± 0.05 | X | X | ||
Mean pooling | 0.90 ± 0.02 | X | X | ||
Max pooling | 0.88 ± 0.02 | X | X | ||
Cheplygina20 | 0.78 ± 0.04 | X | X | ||
Orting21 | 0.88 ± – | X | |||
Tennakoon22 | 0.95 ± – | X |
Attention weights across emphysema classes
Attention weight curves were calculated to demonstrate the influence of disease type localized throughout the lung. The attention weights demonstrated a stronger influence for slices in the upper lung in all scan classes, indicating that the model prioritized upper lobe information (Table 3, Fig. 2). This agrees with published literature trends that note an upper lobe predominance for emphysema, particularly centrilobular, the most common phenotype in this dataset33–36. Recall, influence is quantified by three metrics: (1) depth maximum attention of fit curve, (2) weighted average of slice depths weighted by attention, and (3) range of fit curve attention values.
Table 3.
Scans evaluated | Maximum attention lung depth (%) | Weighted average of lung depth (%) | Range of attention values (%) |
---|---|---|---|
Positive scans | 15.6 | 38.5 | 33.6 |
Negative scans | 32.9 | 38.7 | 47.2 |
Positive: centrilobular | 19.3 | 39.1 | 34.8 |
Positive: panlobular | 17.2 | 46.2 | 20.0 |
Positive: paraseptal | 12.8 | 37.6 | 34.1 |
Recall, influence is quantified by three metrics: (1) depth maximum attention of fit curve, (2) weighted average of slice depths weighted by attention, and (3) range of fit curve attention values.
By phenotype, the centrilobular and paraseptal attention average depths (39.1%, 37.6%) aligned with expected upper lobe predominance compared to panlobular (46.2%). Further, the panlobular scans tended to more heavily influence slices throughout the lung range as demonstrated by the reduced range of attention values (20.0%) compared to the other phenotypes (34.8%, 34.1%). Note that any given scan did not necessarily present with only one phenotype; for example, the scans labeled panlobular-dominant may also present with other phenotypes. This and the model’s learned predisposition to more highly weight the upper lobe slices (as conveyed by the quantification of negative scan attention) may account for the relative importance of the upper lobes even in the panlobular-dominant scans.
CT image features emphasized by attention weights
The top-k influential slices according to attention weights were evaluated to determine which CT imaging features drew the most attention and to identify potential sources of misclassification. The prevalence of image features that were present in the top-k selected attention weighted slices are shown in Fig. 3. Different features were likely to have different prevalence within each scan (e.g., nodules were local abnormalities while architectural distortions were generally more widespread structural changes), thus the prevalence of each imaging feature as identified by a radiologist is presented for comparison. Bronchial disease and architectural distortions demonstrated the largest change in importance for the top-k attended slices compared to the human reader with changes of 23.7% ± 0.03% and 22.7% ± 0.01%, respectively. The prevalence of each feature did not significantly change when including more slices in the attention analysis; however, bronchial disease and architectural distortion features were attended to much more frequently than their frequency in entire scans while the opposite occurred for ground glass opacities. This may suggest that the model was balanced between identifying features indicative of emphysema presence, such as regions of hypoattenuation and structural changes, while maintaining a general representation of the entire CT scan.
Discussion
In this study, we present a novel CT slice-based Transfer AMIL approach for evaluating emphysema on LDCT scans acquired for lung cancer screening. The model provides strong classification performance compared to models with similar label constraints, including models evaluated for this study and those published in the literature. The attention module also provides interpretable information for verifying model performance by identifying slices that were most influential to the classification decision. Indeed, the attention weight trends for different subsets of the LDCT scans agreed with expectations in terms of the most likely regions to find emphysema, including when different classes of emphysema were dominant. A further investigation into the attention weights also revealed which CT image features were most useful for the model prediction and may provide insight into what potential cases will be problematic for automatic evaluation, particularly considering the lung cancer screening population.
Importantly, the developed model requires a relatively small amount of computing power compared to other modern deep learning computer vision tasks. The AUC performance achieved by the Transfer AMIL was either comparable or outperformed other models in this study, including standard 2D classification models with noisy labels and 3D image classifiers. Note that this performance may only hold true for the data available for this study; a large amount of data would likely improve the non-transfer learning models more than transferred models because during training the number of training images approaches the number of trainable parameters.
The pre-trained VGG19 feature extraction model parameters (20 M) were fixed from the pre-training task with no additional training; additional training would further underdetermine the model considering the limited dataset of 860 utilized scans (of 865 total with 5 excluded). With the feature extractor fixed, the additional fully connected layers and attention module require only 1.15 M trainable parameters; still an underdetermined system, but at a greatly reduced risk of overfitting. While larger standard architectures such as ResNet50 and DenseNet121 could be utilized for feature extraction, this study demonstrates that even the use of smaller, less complex models can achieve competitive performance. Note, these architectures were evaluated but no performance gain was observed thus the least computationally expensive model was utilized. Further, the reduced model capacity and use of transfer learning with a common architecture encourage wider implementation of this technique because the compute power needed to run the model is generally attainable by today’s standards and feature extractor does not require local training.
While the attention module interpretable output validates model performance compared to literature, it also encourages clinical implementation as the attention weights can be added as an optional part of the lung screening workflow for a radiologist to further investigate the classification decision, specifically by review of the slices influential to the classification decision. This review process can lead to radiologist trust and understanding of clinical implementation of the algorithm and has the potential to improve clinical workflow in terms of both reading time and performance, although this would require a prospective reader study to confirm.
The theme of improved performance also aligns with the attention module’s ability to identify which cases may be problematic for classification. For example, the model tended to more heavily weight slices with bronchial disease and architectural distortions, which are nonspecific to emphysema patients, and which often appear similar to typical presentation of emphysema (e.g., regions of hypoattenuation and structural changes). This also suggests that patients with these presentations caused by non-emphysematous conditions may be difficult for the model to classify.
Future work should prospectively utilize this model in a reader study to evaluate its impact on radiologist performance and radiological workflow as well as include images acquired from multiple institutions to assess model generalizability. This is especially important as the data in this study were limited (single institution, limited N). Further, this study only evaluated binary classification decisions and does not consider relationships between slices when calculating attention weights; multi-class variants of MIL as well as more complex attention-based pooling functions. Despite these limitations, the Transfer AMIL method achieved strong performance as determined from ROC analysis and the attention weight investigations performed in this study demonstrated strong potential for clinical implementation.
Acknowledgements
Partially funding for this work was provided by the NIH S10-OD025081, and by the previous Awards S10-RR021039 and P30-CA14599 awards. The contents of this paper are solely the responsibility of the authors and do not necessarily represent the official views of any of the supporting organizations. Additional funding through NIH Contracts 75N92020C00008 and 75N92020C00021. Also supported by UL1TR000430.
Author contributions
Project was initially conceived by C.I.H., D.F.Y., and M.L.G. R.Y., A.C.J., C.I.H., and D.F.Y. collected imaging and clinical data. F.L. read images for ground truth slice information. J.D.F. designed the method, conducted all analysis and interpretation, and served as primary writer. J.D.F., R.Y., Y.Z., A.C.J., C.I.H., D.F.Y., and M.L.G. were involved in ongoing discussions of project progress, writing, and revision.
Data availability
The data generated and analyzed during this study may be made available by contacting the corresponding author with reasonable request.
Competing interests
Dr. David Yankelevitz is a named inventor on a number of patents and patent applications relating to the evaluation of diseases of the chest including measurement of nodules. Some of these, which are owned by Cornell Research Foundation (CRF) are non-exclusively licensed to General Electric. As an inventor of these patents, Dr. Yankelevitz is entitled to a share of any compensation which CRF may receive from its commercialization of these patents. Consultant and shareholder for Accumetra LLC. Advisory Board and shareholder HeartLung Technologies. Advisory Board Median Technologies. Dr. Henschke is the President and serves on the board of the Early Diagnosis and Treatment Research Foundation. She receives no compensation from the Foundation. The Foundation is established to provide grants for projects, conferences, and public databases for research on early diagnosis and treatment of diseases. Dr. Claudia Henschke is also a named inventor on a number of patents and patent applications relating to the evaluation of pulmonary nodules on CT scans of the chest which are owned by Cornell Research Foundation (CRF). Since 2009, Dr. Henschke does not accept any financial benefit from these patents including royalties and any other proceeds related to the patents or patent applications owned by CRF. Maryellen L. Giger: Stockholder and receives royalties, Hologic, Inc.; Equity holder and co-founder, Quantitative Insights, Inc (now Qlarity Imaging).; Shareholder, QView Medical, Inc.; Royalties, General Electric Company, MEDIAN Technologies, Riverain Technologies, LLC, Mitsubishi, and Toshiba. Other authors declare no competing interests.
Footnotes
Publisher's note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
References
- 1.US Preventive Services Task Force Screening for lung cancer: US preventive services task force recommendation statement. JAMA. 2021;325:962–970. doi: 10.1001/jama.2021.1117. [DOI] [PubMed] [Google Scholar]
- 2.Henschke CI, et al. Early lung cancer action project: Overall design and findings from baseline screening. Lancet. 1999;354:99–105. doi: 10.1016/S0140-6736(99)06093-6. [DOI] [PubMed] [Google Scholar]
- 3.The National Lung Screening Trial Research Team Reduced lung-cancer mortality with low-dose computed tomographic screening. N. Engl. J. Med. 2011;365:395–409. doi: 10.1056/NEJMoa1102873. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 4.Pickhardt PJ. Value-added opportunistic CT screening: State of the art. Radiology. 2022;303:241–254. doi: 10.1148/radiol.211561. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 5.Yip R, et al. Added benefits of early detection of other diseases on low-dose CT screening. Transl. Lung Cancer Res. 2021;10:1141–1153. doi: 10.21037/tlcr-20-746. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 6.Steiger D, et al. The importance of low-dose CT screening to identify emphysema in asymptomatic participants with and without a prior diagnosis of COPD. Clin. Imaging. 2021;78:136–141. doi: 10.1016/j.clinimag.2021.03.012. [DOI] [PubMed] [Google Scholar]
- 7.González J, et al. Emphysema phenotypes and lung cancer risk. PLoS ONE. 2019;14:e0219187. doi: 10.1371/journal.pone.0219187. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8.Yong PC, et al. The effect of radiographic emphysema in assessing lung cancer risk. Thorax. 2019;74:858–864. doi: 10.1136/thoraxjnl-2018-212457. [DOI] [PubMed] [Google Scholar]
- 9.Labaki WW, et al. Quantitative emphysema on low-dose CT imaging of the chest and risk of lung cancer and airflow obstruction: An analysis of the national lung screening Trial. Chest. 2021;159:1812–1820. doi: 10.1016/j.chest.2020.12.004. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 10.Mouronte-Roibás C, et al. COPD, emphysema and the onset of lung cancer: A systematic review. Cancer Lett. 2016;382:240–244. doi: 10.1016/j.canlet.2016.09.002. [DOI] [PubMed] [Google Scholar]
- 11.de Torres JP, et al. Assessing the relationship between lung cancer risk and emphysema detected on low-dose CT of the chest. Chest. 2007;132:1932–1938. doi: 10.1378/chest.07-1490. [DOI] [PubMed] [Google Scholar]
- 12.Hersh CP, et al. Interobserver variability in the determination of upper lobe-predominant emphysema. Chest. 2007;131:424–431. doi: 10.1378/chest.06-1040. [DOI] [PubMed] [Google Scholar]
- 13.Hartman JE, et al. HRCT characteristics of severe emphysema patients: Interobserver variability among expert readers and comparison with quantitative software. Eur. J. Radiol. 2021;136:109561. doi: 10.1016/j.ejrad.2021.109561. [DOI] [PubMed] [Google Scholar]
- 14.Carbonneau M-A, Cheplygina V, Granger E, Gagnon G. Multiple instance learning: A survey of problem characteristics and applications. Pattern Recogn. 2018;77:329–353. doi: 10.1016/j.patcog.2017.10.009. [DOI] [Google Scholar]
- 15.Wang X, Yan Y, Tang P, Bai X, Liu W. Revisiting multiple instance neural networks. Pattern Recogn. 2018;74:15–24. doi: 10.1016/j.patcog.2017.08.026. [DOI] [Google Scholar]
- 16.Ilse, M., Tomczak, J. & Welling, M. Attention-based deep multiple instance learning. in International Conference on Machine Learning 2127–2136 (PMLR, 2018).
- 17.Humphries SM, et al. Deep learning enables automatic classification of emphysema pattern at CT. Radiology. 2020;294:434–444. doi: 10.1148/radiol.2019191022. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 18.Oh AS, et al. Emphysema progression at CT by deep learning predicts functional impairment and mortality: Results from the COPD gene study. Radiology. 2022;304:213054. doi: 10.1148/radiol.213054. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 19.Negahdar, M., Coy, A. & Beymer, D. An end-to-end deep learning pipeline for emphysema quantification using multi-label learning. in 2019 41st Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC) 929–932 (2019). 10.1109/EMBC.2019.8857392. [DOI] [PubMed]
- 20.Cheplygina, V. et al. Classification of COPD with multiple instance learning. in 2014 22nd International Conference on Pattern Recognition 1508–1513 (2014). 10.1109/ICPR.2014.268.
- 21.Ørting, S. N., Petersen, J., Thomsen, L. H., Wille, M. M. W. & Bruijne, M. de. Detecting emphysema with multiple instance learning. in 2018 IEEE 15th International Symposium on Biomedical Imaging (ISBI 2018) 510–513 (2018). 10.1109/ISBI.2018.8363627.
- 22.Tennakoon R, et al. Classification of volumetric images using multi-instance learning and extreme value theorem. IEEE Trans. Med. Imaging. 2020;39:854–865. doi: 10.1109/TMI.2019.2936244. [DOI] [PubMed] [Google Scholar]
- 23.Zulueta JJ, et al. Emphysema scores predict death from COPD and lung cancer. Chest. 2012;141:1216–1223. doi: 10.1378/chest.11-0101. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 24.Lynch DA, et al. CT-definable subtypes of chronic obstructive pulmonary disease: A statement of the Fleischner society. Radiology. 2015;277:192–205. doi: 10.1148/radiol.2015141579. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 25.Shao, Z. et al.TransMIL: Transformer based Correlated Multiple Instance Learning for Whole Slide Image Classification. http://arxiv.org/abs/2106.00908 (2021).
- 26.Simonyan, K., Vedaldi, A. & Zisserman, A. Deep Inside Convolutional Networks: Visualising Image Classification Models and Saliency Maps. http://arxiv.org/abs/1312.6034 (2013).
- 27.Giger ML. Machine learning in medical imaging. J. Am. Coll. Radiol. 2018;15:512–520. doi: 10.1016/j.jacr.2017.12.028. [DOI] [PubMed] [Google Scholar]
- 28.Sahiner B, et al. Deep learning in medical imaging and radiation therapy. Med. Phys. 2019;46:e1–e36. doi: 10.1002/mp.13264. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 29.Litjens G, et al. A survey on deep learning in medical image analysis. Med. Image Anal. 2017;42:60–88. doi: 10.1016/j.media.2017.07.005. [DOI] [PubMed] [Google Scholar]
- 30.Antropova N, Huynh BQ, Giger ML. A deep feature fusion methodology for breast cancer diagnosis demonstrated on three imaging modality datasets. Med. Phys. 2017;44:5162–5171. doi: 10.1002/mp.12453. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 31.Niu Z, Zhong G, Yu H. A review on the attention mechanism of deep learning. Neurocomputing. 2021;452:48–62. doi: 10.1016/j.neucom.2021.03.091. [DOI] [Google Scholar]
- 32.DeLong ER, DeLong DM, Clarke-Pearson DL. Comparing the areas under two or more correlated receiver operating characteristic curves: A nonparametric approach. Biometrics. 1988;44:837–845. doi: 10.2307/2531595. [DOI] [PubMed] [Google Scholar]
- 33.Araki T, et al. Paraseptal emphysema: Prevalence and distribution on CT and association with interstitial lung abnormalities. Eur. J. Radiol. 2015;84:1413–1418. doi: 10.1016/j.ejrad.2015.03.010. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 34.Wyatt JP, Fischer VW, Sweet HC. Panlobular emphysema: Anatomy and pathodynamics. Dis. Chest. 1962;41:239–259. doi: 10.1378/chest.41.3.239. [DOI] [PubMed] [Google Scholar]
- 35.Anderson AE, Foraker AG. Centrilobular emphysema and panlobular emphysema: Two different diseases. Thorax. 1973;28:547–550. doi: 10.1136/thx.28.5.547. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 36.Foster WL, et al. Centrilobular emphysema: CT-pathologic correlation. Radiology. 1986;159:27–32. doi: 10.1148/radiology.159.1.3952318. [DOI] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Data Availability Statement
The data generated and analyzed during this study may be made available by contacting the corresponding author with reasonable request.