Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2019 Dec 1.
Published in final edited form as: Clin Cancer Res. 2018 Oct 11;24(23):5902–5909. doi: 10.1158/1078-0432.CCR-18-1115

Deep Learning to Distinguish Recalled but Benign Mammography Images in Breast Cancer Screening

Sarah S Aboutalib 1, Aly A Mohamed 2, Wendie A Berg 2,3, Margarita L Zuley 2,3, Jules H Sumkin 2,3, Shandong Wu 4,*
PMCID: PMC6297117  NIHMSID: NIHMS1503183  PMID: 30309858

Abstract

Purpose:

False positives in digital mammography screening lead to high recall rates, resulting in unnecessary medical procedures to patients and health care costs. This study aimed to investigate the revolutionary deep learning methods to distinguish recalled but benign mammography images from negative exams and those with malignancy.

Experimental Design:

Deep learning convolutional neural network (CNN) models were constructed to classify mammography images into malignant (breast cancer), negative (breast cancer free), and recalled-benign categories. A total of 14860 images of 3715 patients from two independent mammography datasets: Full-Field Digital Mammography Dataset (FFDM) and a digitized film dataset, Digital Dataset of Screening Mammography (DDSM), were used in various settings for training and testing the CNN models. The receiver operating characteristic curve was generated and the area under the curve (AUC) was calculated as a metric of the classification accuracy.

Results:

Training and testing using only the FFDM dataset resulted in AUC ranging from 0.70 to 0.81. When the DDSM dataset was used, AUC ranged from 0.77 to 0.96. When datasets were combined for training and testing, AUC ranged from 0.76 to 0.91. When pre-trained on a large non-medical dataset and DDSM, the models showed consistent improvements in AUC ranging from 0.02 to 0.05 (all p > 0.05), compared to pre-training only on the non-medical dataset.

Conclusions:

This study demonstrates automatic deep learning CNN methods can identify nuanced mammographic imaging features to distinguish recalled-benign images from malignant and negative cases, which may lead to a computerized clinical toolkit to help reduce false recalls.

Keywords: breast cancer, digital mammography, screening, recall, deep learning

Introduction

Mammography is clinically used as the standard breast cancer screening exam for the general population and has been shown effective in early detection of breast cancer and in reduction of mortality (13). High recall (asking a woman back for additional workup after a screening mammogram) rates are, however, a concern in breast cancer screening. On average, approximately 11.6% of women in the U.S. (over 5 million women annually) who are screened using digital mammography are recalled and over 70% of more than 1 million breast biopsies performed annually are benign (4,5), resulting in unnecessary psychological stress, medical costs, and clinical workload. Thus, reducing false recalls from screening mammography is of great clinical significance.

Observer performance in breast cancer detection in screening mammography varies widely and is influenced by experience, volume, and subspecialization among other factors (6). One approach to improving mammographic interpretation is to provide the radiologist with powerful computerized tools to aid in image interpretation and decision-making. In order to help radiologists make more accurate decisions on whether to recall a patient, we focused on building computer-aided models/classifiers that can distinguish subtle imaging characteristics of mammograms from patients who were recalled but biopsied with benign results from mammograms read as negative from the onset, and mammograms from recalled cases with biopsy-proven breast cancer. Computerized classifiers can assist radiologists in predicting which patients/images may be recalled but most likely benign.

Deep learning (7) coupled with a big dataset has shown promising performance in many artificial intelligence applications and is entering the field of biomedical imaging (810). The main architecture of deep learning for image data is the convolutional neural network (CNN) (11). The most distinguishing strength of the CNN is that it can automatically learn and hierarchically organize features from a large dataset without manual feature engineering (7,12,13) and outperformed, in many scenarios, traditional manual-engineered imaging features/descriptors. While deep learning CNN can be used as an offline feature-extractor where features are coupled with traditional classifiers (14), it is more common to build end-to-end deep learning classification models (15,16). Studies applying deep learning to digital mammography images have focused on mass and microcalcification detection and classification, distinction between tumor and benign/normal tissue (1721), breast tissue segmentation (22), and classification of breast anatomy (9).

Recently, a number of groups competed under the Digital Mammography DREAM challenge (23) to distinguish malignant from negative mammogram images. While important, distinguishing potentially recalled but biopsy benign images from both malignant and negative images represents a critical need and a practical approach to aid radiologists. The purpose of this study was to investigate end-to-end deep learning CNN models for automatic identification of nuanced imaging features to distinguish mammogram images belonging to negative, recalled-benign, and malignant cases aimed to improve clinical mammographic image interpretation and reduce unnecessary recalls.

Materials and Methods

We performed a retrospective study that received Institutional Review Board (IRB) approval at our institution. This study was compliant to the Health Insurance Portability and Accountability Act (HIPAA) and the U.S. Common Rule. Informed consent from patients was waived due to the retrospective nature.

Study cohort and datasets

We used two independent mammography datasets to develop and evaluate deep learning classifiers. These two independent mammogram datasets include a total of 3715 patients and 14860 images. A patient case typically contains a single patient exam with the standard four screening mammography views including left and right breast with craniocaudal (CC) and mediolateral oblique (MLO) views. Malignant images were taken from patients determined to have breast cancer based on pathology results (only images of the cancer-affected breast were used). Negative images were from those who maintained a breast cancer-free status after at least a one-year follow-up. Recalled-benign images were taken from patients who were recalled based on the screening mammography exam but later determined as benign (biopsy-proven or diagnostically confirmed).

Full-field Digital Mammography (FFDM) Dataset:

This is a retrospective cohort of 1303 patients (5212 mammogram images) who underwent standard digital mammography screening (2007–2014) at our institution: 552 patients were evaluated as negative in the initial screen; 376 patients were recalled-benign; 375 patients were evaluated as positive for breast cancer (101 or 26.9% ductal carcinoma in situ and 274 or 73.1% invasive) based on pathology results.

Digital Database of Screening Mammography (DDSM) Dataset:

The DDSM dataset (2426) is a large collection of digitized film mammography images. 9648 images consisting of 2412 patient cases were used from this dataset. 695 cases were negative, 867 malignant, and 850 recalled-benign.

Deep learning approach for building classifiers

We built end-to-end two-class and three-class CNN models to investigate six classification scenarios: Five binary classifications – Malignant vs Recalled-Benign+Negative (i.e. Recalled-benign and Negative are merged into one class), Malignant vs Negative, Malignant vs Recalled-Benign, Negative vs Recalled-Benign, and Recalled-Benign vs Malignant+Negative (Negative and Malignant are merged into one class), as well as one triple classification – Malignant vs Negative vs Recalled-Benign.

The CNN used a modified version of the AlexNet (11,27) model. The CNN structure consists of five convolutional layers (includes max-pooling in the 1st, 2nd, and 5th convolutional layers) followed by two fully connected layers and a fully connected output layer with a final softmax function. CNN model parameters were fixed in all experiments: batch size of 50 for stochastic gradient descent, a weight decay of 0.001, and a momentum of 0.9. For the learning rate, we started with 0.001 and dropped the rate by a factor of ten every 2500 iterations. To speed up training, rectified linear units were used as the activation function in place of traditional tangent sigmoid functions (7). To maximize performance and increase computational efficiency of the AlexNet network, images were pre-processed using standard techniques including histogram equalization, mean subtracting, and down-sampling using standard bicubic interpolation to 227×227 pixels from original resolution.

Transfer learning was used to enhance model training. We pre-trained all the CNN models with a large existing image dataset (ImageNet (11), 1.3 million non-medical images) and then fine-tuned the models with our own mammography dataset. In a novel approach, we also tested incremental transfer learning strategies: first pre-training the network using ImageNet and then continuing with the DDSM dataset, and finally fine-tuning and testing on the FFDM dataset.

The CNN was trained with the goal of increasing variation of the data and avoiding overfitting. Internal validation is based on 6-fold cross validation of training data in the CNN model training phase, reducing overfitting and calibrating accuracy of the training process. External evaluation of the CNN models was performed using unseen testing data.

The deep learning network was implemented using the Caffe platform running on a desktop computer system with the following specifications: Intel® Core™ i7–2670QM CPU@2.20GHZ with 8 GB RAM and a Titan X Pascal Graphics Processing Unit (GPU).

Evaluation and statistical analysis

An independent test cohort consisting of approximately 5% of our datasets was used for testing (95% data for training). We used this setting in order to maximize the amount of data for CNN model training. Table 1 summarizes the number of images used in our independent training and testing datasets in each experiment. As an additional robustness analysis, we also repeated the experiments by using 10% and 15% of the data as independent sets for testing. In all settings, testing data did not include any images used in training.

Table 1.

Number of training and testing images from the full-field digital mammography (FFDM) and digital database of screening mammography (DDSM) datasets used for each experiment. The Negative vs Recalled-Benign scenarios (Negative vs Recall-Benign) have more data and thus are listed separately. For simplicity, all other scenarios (Malignant vs Negative+Recalled-Benign, Malignant vs Negative, Malignant vs Recalled-Benign, Malignant vs Negative vs Recalled-Benign, Recalled-Benign vs Malignant+Negative) are listed under “Others”. Both the total number of images in the scenario as well as the images per category are displayed.

Number of Training and Testing Images Used
Dataset / Experiment Scenario Training Testing
Total Per Category Total Per Category
FFDM Only (Train & Test) Negative vs Recalled-Benign 3040 1520 160 80
Others 1734 867 100 50
DDSM Only (Train & Test) Negative vs Recalled-Benign 5282 2641 278 139
Others 3294 1647 172 86
FFDM+DDSM (Train & Test) Negative vs Recalled-Benign 8322 4161 438 219
Others 5028 2514 272 136
Pre-train on ImageNet and DDSM, test on FFDM Negative vs Recalled-Benign 3040 1520 160 80
Others 1734 867 100 50

We performed 4 different experiments/strategies utilizing the two datasets in testing the CNN model’s performance in six classification scenarios. These 4 experiments included: (I) Using the FFDM dataset only for training and testing; (II) Using the DDSM dataset only for training and testing, in order to determine how our CNN model performs on an independent dataset; (III) Using the FFDM and DDSM datasets together by combining the datasets (mixing up) for both training and testing and lastly (IV) using incremental transfer learning by pre-training the CNN model on the DDSM dataset after pre-training on the ImageNet dataset. Note that in all four experiments the base CNN models are pre-trained on the ImageNet and fine-tuned with the FFDM dataset.

To explore clinical caution for reducing the risk of false negatives in recalling a woman, we performed additional classification experiments using CNN models to distinguish false negatives from recalled-benign cases and also from negative cases. Here we used interval cancers as the false negative cases in the experiments.

The receiver operating characteristic (ROC) curve (28) was generated and area under the curve (AUC) was calculated as a metric of classification accuracy. For the triple classification, since ROC is a binary-class evaluation method, we followed the common practice in literature of generating an ROC curve for each binary-class combination and then reporting the average of the AUCs. 95% Confidence intervals (CI) were calculated for AUC values using bootstrapping methods (29). DeLong test (30) was used to evaluate the statistical significance when comparing differences of AUCs. All statistical tests were two-sided.

Results

Results on FFDM dataset

As shown in Figure 1, we found that all categories can be well distinguished (AUC ranging from 0.66 to 0.81) in the FFDM dataset. The identified imaging features between recalled-benign and negative were most distinguishing (AUC=0.81; 95% CI 0.75–0.88). Recalled-Benign vs Malignant+Negative (AUC=0.76; 95% CI 0.67–0.86) had the second best performance, followed by Malignant vs Negative (AUC=0.75; 95% CI 0.65–0.84). Malignant vs Recalled-Benign had performance result of AUC=0.74 (95% CI 0.65–0.84). Malignant vs Negative+Recalled-Benign had AUC=0.70 (95% CI 0.60–0.80). Triple classification (Malignant vs Negative vs Recalled-Benign) had an average AUC of 0.66.

Figure 1.

Figure 1.

Performance results for deep learning convolutional neural network (CNN) models for classification on the full-field digital mammography (FFDM) dataset. (Left) Receiver operating characteristic (ROC) curves for the binary classification scenarios and corresponding area under the curves (AUC). (Right) ROC curves for the triple-class classification scenario and averaged AUC.

Results on DDSM dataset

Figure 2 shows the results on the DDSM dataset. We found AUC ranged from 0.77 to 0.90. Negative vs Recalled-Benign and Malignant vs Negative showed the best performances with AUC of 0.96 (95% CI 0.94–0.98) and 0.93 (95% CI 0.89 – 0.97) respectively. Triple classification had an average AUC of 0.85. Recalled-Benign vs Malignant+Negative showed the next best performance at 0.83 (95% CI 0.77 – 0.89). Malignant vs Recalled-Benign and Malignant vs Negative+Recalled-Benign showed similar performance with AUCs of 0.78 (95% CI 0.71 – 0.85) and 0.77 (95% CI 0.70 – 0.84) respectively.

Figure 2.

Figure 2.

Performance results for deep learning convolutional neural network (CNN) models for classification on the digital database of screening mammography (DDSM) dataset. (Left) Receiver operating characteristic (ROC) curves for the binary classification scenarios and corresponding area under the curves (AUC). (Right) ROC curves for the triple-class classification scenario and averaged AUC.

Results on combined FFDM and DDSM datasets

Figure 3 shows performance results when the two datasets were combined (mixed together) for both training and testing. AUC ranged from 0.76 to 0.91. Best performance was observed for Negative vs Recalled-Benign (AUC 0.91; 95% CI 0.89 – 0.94). With an AUC of 0.84 (95% CI 0.79 – 0.88), Malignant vs Negative had second best performance, followed by Malignant vs Negative vs Recalled-Benign with an average AUC of 0.79, then Recalled-Benign vs Malignant+Negative (AUC 0.78; 95% CI 0.73 – 0.83). Both Malignant vs Recalled-Benign and Malignant vs Negative+Recalled-Benign had an AUC of 0.76 (95% CI 0.70 – 0.82). It is observed that when FFDM and DDSM were combined for training and testing, overall classification showed AUC values between FFDM-only and DDSM-only results.

Figure 3.

Figure 3.

Performance results for deep learning convolutional neural network (CNN) models for classification using combined Full-field Digital Mammography (FFDM) and digital database of screening mammography (DDSM) datasets for training and testing. (Left) Receiver operating characteristic (ROC) curves for the binary classification scenarios and corresponding area under the curves (AUC). (Right) ROC curves for the triple-class classification scenario and averaged AUC.

Incremental transfer learning using the DDSM dataset

DDSM dataset was used to further pre-train the base CNN network (after originally pre-trained on the ImageNet dataset) which was then fine-tuned with the FFDM dataset. Testing data was also from the FFDM dataset. As can be seen in Figure 4, when compared with results using ImageNet pre-trained model, all scenarios showed an increase in performance using this incremental transfer learning strategy, although not statistically significant (p > 0.05). Malignant vs Recalled-Benign showed the greatest increase in performance increasing from 0.75 to 0.80 (95% CI 0.71 – 0.88, p = 0.10) or 5%, followed by Malignant vs Negative+Recalled-Benign (AUC 0.70 to 0.74; 95% CI 0.65 – 0.84, p = 0.15). Malignant vs Negative and Recalled-Benign vs Malignant+Negative both had their AUC increase by 0.03, 0.74 to 0.77 (95% CI 0.68 – 0.87, p = 0.30) and 0.76 to 0.79 (95% CI 0.70 – 0.87, p = 0.31) respectively. Negative vs Recalled-Benign and the triple classification (Malignant vs Negative vs Recalled-Benign) had the least increase (2% or 0.02 in AUC, 0.66 to 0.68 for triple classification), although Negative vs Recalled-Benign had the best performance of all the scenarios (AUC 0.83 [95% CI 0.76 – 0.87, p = 0.34] compared with 0.81 for ImageNet pre-trained model).

Figure 4.

Figure 4.

Comparison of performance results of deep learning convolutional neural network (CNN) models on different pre-training strategies: using original ImageNet pre-trained model vs. using model pre-trained on ImageNet and digital database of screening mammography (DDSM) dataset. All the area under the curves (AUCs) were results based on training (fine-tuning) and testing on the full-field digital mammography (FFDM) dataset.

Robustness analysis

In the robustness analysis using 10% and 15% data for testing (Figure 5), as expected, we see a slight decrease in the AUC performance in almost all scenarios compared to those using 5% for testing. This is partly due to the reduced amount of data for training. We also found that the performances overall remained relatively stable, demonstrating the robustness of our CNN models on a varied amount of data for training and testing.

Figure 5.

Figure 5.

Comparison of performance results using varying amounts (5%, 10%, and 15%) of testing data across all models: (Top-Left) FFDM trained model; (Top-Right) DDSM trained model; (Bottom-Left) FFDM + DDSM trained model; (Bottom-Right) Incrementally pre-trained CNN models in all scenarios.

False negative analysis

The CNN models were further tested on correctly identifying false negative cases/images. In our FFDM dataset, there are 34 interval cancer patients examined here as false negative cases. We re-trained the CNN models (both the Malignant vs Negative scenario and the Malignant vs Recalled-Benign scenario) on the FFDM dataset only and excluding those interval cancer patients, and then tested these models by inputting the false negative images to the models. Our results showed that 71.3% (by the Malignant vs Negative model) and 63.6% (by the Malignant vs Recalled-Benign model) of the entire false negative images can be correctly identified by our models. Furthermore, when we re-trained the CNN model by including 50% of the false negative cases and used the rest of the unseen 50% for testing, we see improved results, as expected, i.e., 72.8% (by the Malignant vs Negative model) and 68.4% (by the Malignant vs Recalled-Benign model) of the entire false negative images can be correctly identified. The CNN models trained by other strategies (i.e., experiments III and IV in Section “Evaluation and statistical analysis”) also achieved similar performance (results not shown).

Discussion

In this study, we present a novel investigation showing that automatic deep learning CNN methods can identify nuanced mammographic imaging features to distinguish negative, recalled-benign, and malignant images. We demonstrated that it is feasible to discover subtle imaging distinction for six classification scenarios on two different imaging datasets.

Among the six scenarios, Negative vs Recalled-Benign showed the best performance. This scenario also had the greatest amount of data, which may have contributed to this improved performance. The distinction in this scenario implies that certain imaging features may result in recalled-benign images to be recalled rather than being determined negative in the first place. In general, the relatively higher AUCs in Negative vs Recalled-Benign and also Malignant vs Recalled-Benign indicate that there are imaging features unique to recalled-benign images that the CNN-based deep learning can identify and potentially use to help radiologists in making better decisions on whether a patient should be recalled or is more likely a false recall. The six classification scenarios were designed to reveal different aspects of performance of the CNN models. It however remains to be determined which scenario, especially in terms of binary-classification or triple-classification, would be the most useful and meaningful choice in implementing a real-world CNN model to help radiologists make recall decisions in a clinical setting. There are noticeable variations in the classification performance between binary and triple classification. This is something to be further investigated in conjunction perhaps with reader studies to evaluate their clinical effects in depth. Of note, with the ability of the CNN model to distinguish between negative and malignant images, the results indicate that deep learning can also perform well in computer-aided diagnosis of breast cancer.

In terms of diagnosis (i.e., Malignant vs Negative), literature has reported a great reader variability of radiologists in sensitivity and specificity in screening mammography (4). Using previously-reported radiologists’ overall sensitivity 86.9% as a reference threshold (4), our best deep learning model for classifying Malignant vs Negative yields a specificity of 87%, comparable to radiologists’ overall specificity of 88.9%. This is encouraging for the current scale of our datasets. With enlarged datasets we would expect further improvement in our model’s performance. In terms of our major motivation of this work for reducing recalls, the lack of exact radiologists’ performance data in the literature prevents us from making a reasonable comparison to our CNN models. We aim to further investigate this with a reader study in future work.

Our CNN models demonstrated encouraging results in the false negative analysis. Prevention of false negatives may be an important reason for the high recall rate in current clinical practice. By examining classifications of the interval cancer cases, it indicates a potential of the CNN models to help correctly identify as malignant a majority of false negative cases from recalled-benign cases or from negative cases. This also implies that some cases may be recalled due to certain imaging indications besides the intention of preventing false negatives. Further study into this issue on a larger cohort of false negative cases is of great clinical importance.

In terms of dataset, we started with a digital mammography dataset, FFDM, as it is the current standard screening mammography examination. To demonstrate that our CNN models can be used on a dataset from an independent institution, we tested our models on the DDSM dataset. The DDSM alone showed the best performance overall. This may be due to the larger dataset size or something intrinsic to the characteristics of the DDSM dataset. When the FFDM and DDSM datasets were combined, overall improved performance was observed compared with using FFDM alone. This indicates that our CNN model is robust to an external dataset and its performance can be further improved by including additional data.

Since the DDSM dataset is based on digitized film mammography images and current clinical practice has moved to digital mammography, we wanted to determine the best use of the DDSM to improve results on the FFDM dataset. The most straightforward way was simply to combine it with the FFDM datasets, which improved results when testing on the combined datasets. More importantly, we demonstrated a novel approach of incremental transfer learning using the DDSM dataset, which enhanced the performance consistently on the FFDM dataset in all six scenarios. The incremental transfer learning we used was a two-phase transfer learning approach using two different datasets consecutively (i.e., ImageNet then DDSM). The improvement may be due to using DDSM to fine-tune the model’s weights rather than directly in training, allowing the noise from being a different type of mammography dataset to be dampened. It could also be related to the strengths of DDSM as a medical imaging modality that is close to the target modality (i.e., digital mammography) in fine-tuning the weights learned from pre-training on the non-medical imaging dataset of ImageNet. Many studies have shown the benefit of transfer learning in medical domains with limited data (31). Our results in this study provide deeper insights in developing more optimized transfer learning strategies. However, the incremental transfer learning and the observations made here needs to be evaluated by further analyses and comparative studies in future work.

In comparing the various deep/transfer learning strategies, we show that adding/mixing a larger independent dataset, even if it is not exactly the same imaging modality, helped improve the CNN models in our classification tasks. Furthermore, using this kind of additional dataset in an incremental transfer learning approach has shown a trend of boosted model performance. Although the AUC increases are not statistically significant, additional data will help further evaluate this finding. As to mixing two datasets or using transfer learning, it depends on the specific scenarios and their actual classification performance. While we have shown encouraging results in this study, we believe this is still an open question meriting further investigation.

In traditional computer-aided-detection or -diagnosis, the models are usually based on predefined features, which require preemptive determination of which features will work best for the task at hand. In contrast, with the deep learning method we utilized, predefinition of the imaging features is not necessary and are learned automatically from labeled data. Deep learning allows nuanced features to be determined by the learning algorithm for the targeted task, where intrinsic features that may not be identifiable by human visual assessment can be automatically identified and used for imaging interpretation. This study illustrated the encouraging effects of such automatic feature identification by deep learning.

Our study has some limitations. While we have two independent datasets, additional datasets, especially another digital mammography dataset, can further bolster the evaluation of the deep learning models. We are currently trying to obtain external datasets for such experiments. Also, it is still not clear why the DDSM performed substantially better than the FFDM dataset; this requires further exploration. Of note, digital breast tomosynthesis is increasingly being used in clinical practice and it has been shown to reduce recall rates (32); thus incorporation of tomosynthesis data in our study will be important future work. While we utilized AlextNet CNN structures in this study, comparison with other network structures such as Residual Network (33), VGG (34), or GoogLeNet (35) will be useful to gain further insights on the potential of deep learning.

Finally, it would be useful to present to radiologists the subtle imaging features found by our CNN models in distinguishing the different groups. However, we are not yet able to clearly visualize and clinically interpret what the identified nuanced imaging features are for recalled but benign images or for the other classification categories. At this phase, deep learning is often referred to as a “black-box” due to the lack of interpretability of the identified features. A current area of investigation we are exploring is to visualize the CNN-identified features to be more intuitively perceived by radiologists. The complexity of deep learning network structures, parameters, and data evolving process across different network layers, however, make feature visualization very complicated, requiring in-depth research. Further technical advancement in this active research area is expected to contribute to addressing this important issue.

In summary, we showed that the three different imaging reading categories (malignant, negative, and recalled-benign) could be distinguished using our deep learning-based CNN models. We believe our study holds great potential to incorporate deep learning-based artificial intelligence into clinical workflow of breast cancer screening to improve radiologist interpretation of mammograms, ultimately contributing to reducing false recalls.

Translational relevance.

Breast cancer screening mammography is currently affected by high false recall rates resulting in unnecessary stress to patients, increased medical costs and increased clinical workload. Deep learning convolutional neural networks (CNNs) can be used to recognize the nuanced imaging features that distinguish recalled but benign mammography images, which may not be identifiable by human visual assessment. These imaging features can then be used for imaging interpretation and lead to computational tools that aid radiologists in distinguishing these images and thus help in the reduction of the false recall rate. In addition, with the ability of the CNN model to distinguish between negative and malignant images, deep learning can also perform well in computer-aided diagnosis of breast cancer.

Acknowledgments

This work was supported by National Institutes of Health (NIH)/National Cancer Institute (NCI) grants (#1R01CA193603, #3R01CA193603-03S1, and #1R01CA218405) (S. Wu, A. Mohamed, W. Berg, M. Zuley), a National Library of Medicine (T15LM007059) grant (S. Aboutalib), a Radiological Society of North America (RSNA) Research Scholar Grant (#RSCH1530) (S. Wu), a University of Pittsburgh Cancer Institute Precision Medicine Pilot Award from The Pittsburgh Foundation (#MR2014-77613) (S. Wu, A. Mohamed), and a University of Pittsburgh Physicians (UPP) Academic Foundation Award (S. Wu). We gratefully acknowledge the support of NVIDIA Corporation with the donation of the Titan X Pascal GPU used for this research.

Footnotes

We have no conflicts of interests to disclose.

References

  • 1.Tabar L, Fagerberg G, Chen HH, et al. Efficacy of breast cancer screening by age: New results from the Swedish Two-County Trial. Cancer. 1995;75(10):2507–17. [DOI] [PubMed] [Google Scholar]
  • 2.Preventive Services Task Force US. Screening for breast cancer: U.S. Preventive Services Task Force recommendation statement. Ann Intern Med. 2016;164(4):279–96. [DOI] [PubMed] [Google Scholar]
  • 3.Coldman A, Phillips N, Wilson C, et al. Pan-Canadian study of mammography screening and mortality from breast cancer. 2014;106(11). [DOI] [PubMed] [Google Scholar]
  • 4.Lehman D, Arao RF, Sprague BL, et al. National Performance Benchmarks for Modern Screening Digital Mammography: Update from the Breast Cancer Surveillance Consortium Constance. Radiology. 2017;283(1):49–58. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5.Silverstein MJ, Lagios MD, Recht A, et al. Image-detected breast cancer: state of the art diagnosis and treatment. J Am Coll Surg. 2005;201(4):586–97. [DOI] [PubMed] [Google Scholar]
  • 6.Elmore JG, Wells CK, Howard DH. Does diagnostic accuracy in mammography depend on radiologists’ experience? J Womens Health. 1998;7:443–449. [DOI] [PubMed] [Google Scholar]
  • 7.LeCun Y, Bengio Y, Hinton G. Deep learning. Nature. 2015;521:436–444. [DOI] [PubMed] [Google Scholar]
  • 8.Wang D, Khosla A, Gargeya R, Irshad H, Beck A. Deep Learning for Identifying Metastic Breast Cancer. ARXIV. 2016. [Google Scholar]
  • 9.Dubrovina A, Kisilev P, Ginsburg B, Hashoul S, Kimmel R. Computational mammography using deep neural networks. Comp Methods in Biomechanics and Biomed Engineering: Imaging & Visualization. 2016:1–5. [Google Scholar]
  • 10.Huval B, Coates A, Ng A. Deep learning for class-generic object detection. AIRXIV. 2013. [Google Scholar]
  • 11.Krizhevsky A, Sutskever I, Hinton GE. Imagenet classification with deep convolutional neural networks. Advanced in Neural Information Processing Systems (NIPS). 2012;25:1106–1114. [Google Scholar]
  • 12.Litjens G, Kooi T, Bejnordi BE. A survey on deep learning in medical imag analysis. CoAA. 2017. [DOI] [PubMed] [Google Scholar]
  • 13.Deng L, Yu D. Deep Learning: methods and applications. Foundations and Trends in Signal Processing. 2014;7(3–4):197–387. [Google Scholar]
  • 14.Li H, Giger M, Antropova N. Deep Learning in breast cancer risk assessment: evaluation of convolutional neural networks on a clinical dataset of full-field digital mammograms. Journal Med Imagning. 2017;4(4). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15.Carneiro G, Nascimento J, Bradely AP. Unregistered multiview mammogram analysis with pre-trained deep learning models. International Conference on Medical Image Computing and Computer-Assisted Intervention 2015:652–660. [Google Scholar]
  • 16.Mohamed A, Berg W, Peng H, Luo Y, Jankowitz R, Wu S. A deep learning method for classifying mammographic breast density categories. Medical Phyiscs. 2017. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17.Dhungel N, Carneiro G, Bradley AP. Automated mass detection in mammograms using cascaded deep learning and random forests. IEEE International Conference on Digital Image Computing: Techniques and Applications 2015:1–8. [Google Scholar]
  • 18.Ertosun MG, Rubin DL. Probabilistic visual search for masses within mammography images using deep learning. IEEE International Conference on Bioinformatics and Biomedicine 2015:1310–1315. [Google Scholar]
  • 19.Levy D, Jain A. Breast mass classification from mammograms using deep convolutional neural networks. arXiv:1612.00542 2016. [Google Scholar]
  • 20.Arevalo J, Gonzalez FA, Ramos-Poll R, Oliveira JL, Lopez MAG. Representation learning for mammography mass lesion calssification with convolutional neural networks. Computer methods and programs in biomedicine. 2016;127:248–257. [DOI] [PubMed] [Google Scholar]
  • 21.Mordang JJ, Janssen T, Bria A, Kooi T, Gubern-Merida A, Karssemeijer N. Automatic microcalcification detection in multivendor mammography using convolutional neural networks. International Workshop on Digital Mammograph 2016:35–42. [Google Scholar]
  • 22.Petersen K, Nielsen M, Diao P, Karssemeijer N, Lillholm M. Breast Tissue Segmentation and Mammographic Risk Scoring Using Deep Learning. International Workshop on Digital Mammography 2014;8539:88–94. [Google Scholar]
  • 23.Digital Mammography DREAM Challenge. 2016-2017. Available at: https://www.synapse.org/Digital_Mammography_DREAM_Challenge.
  • 24.Heath M, Bowyer K, Kopans D, et al. Current status of the Digital Database for Screening Mammography. Proceedings of the Fourth International Workshop on Digital Mammography 1998:457–460. [Google Scholar]
  • 25.Heath M, Bowyer K, Kopans D, Moore R, Kegelmeyer PW. The Digital Database for Screening Mammography. Proceedings of the 5th International Workshop on Digital Mammography 2001:212–218. [Google Scholar]
  • 26.Sharma A DDSM Utility. GitHub; 2015. Available at: https://github.com/trane293/DDSMUtility. [Google Scholar]
  • 27.Available at: https://github.com/BVLC/caffe/tree/master/models
  • 28.Metz CE. ROC methodology in radiologic imaging. Investigative radiology. 1986;21(9):720–733. [DOI] [PubMed] [Google Scholar]
  • 29.Robin X, Turck N, Hainard A. pROC: an open-source package for R and S+ to analyze and compare ROC curves. BMC Bioinformatics. 2011;7(77). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 30.DeLong E, DM D, Clarke-Pearson D. Comparing areas under two or more correlated reciever operating characteristics curves: a nonparamentric approach. Biometrics. 1988;44(3):837–845. [PubMed] [Google Scholar]
  • 31.Zhou Z, Shin J, Zhang L. Fine-tuning Convolutional Neural Network for Biomedical Image Analysis: Actively and Incrementally. In Proc. IEEE CVPR 2017. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 32.McDonald E, Oustimov A, Weinstein S, Synnestvedt M, Schnall M, Conant E. Effectiveness of Digital Breast Tomosynthesis Compared With Digital Mammography: Outcomes Analysis From 3 Years of Breast Cancer Screening. JAMA Oncology. 2016;2(6):737–743. [DOI] [PubMed] [Google Scholar]
  • 33.He K, Zhang X, Shaoqing R, and Sun J. Deep Residual Learning for Image Recognition. CoRR. 2015;abs/1512.03385. [Google Scholar]
  • 34.Simonyan K, Zisserman A. Very Deep Convolutional Networks for Large-Scale Image Recognition. CoRR. 2014;abs/1409.1556. [Google Scholar]
  • 35.Szegedy C, Liu W, Jia Y, et al. Going Deeper with Convolutions. CoRR. 2014;abs/1409.4842. [Google Scholar]
  • 36.Weaver DL, Rosenberg RD, Barlow WE, et al. Pathologic findings from the Breast Cancer Surveillance Consortium: population-based outcomes in women undergoing biopsy after screening mammography. Cancer. 2006;106(4):732–42. [DOI] [PubMed] [Google Scholar]
  • 37.Deng J, Dong W, Socher R, Li LJ, Li K, Fei-Fei L. ImageNet: a large-scale hierarchical image database. Paper presented at: CVPR, 2009. [Google Scholar]
  • 38.Berg WA, D’Orsi CJ, Jackson VP, et al. Does training in the Breast Imaging Reporting and Data System (BI-RADS) improve biopsy recommendations or feature analysis agreement with experienced breast imagers at mammography? Radiology. 2002;224(3):871–80. [DOI] [PubMed] [Google Scholar]
  • 39.Roque AC, Andre TC. Mammography and computerized decision systems: a review. Acad Sci. 2002;980:83–94. [DOI] [PubMed] [Google Scholar]
  • 40.Huynh BQ, Li H, Giger L. Digital mammographic tumor classification using transfer learning from deep convolutional neural networks. Journal of Medical Imaging. 2016;3(3). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 41.Durand M, Haas B, Yao X, et al. Early Clinical Experience with Digital Breast Tomosynthesis for Screening Mammography. Radiology. 2015;274(1):85–92. [DOI] [PubMed] [Google Scholar]

RESOURCES