Abstract
Background
The limitations of traditional computer-aided detection (CAD) systems for mammography, the extreme importance of early detection of breast cancer and the high impact of the false diagnosis of patients drive researchers to investigate deep learning (DL) methods for mammograms (MGs). Recent breakthroughs in DL, in particular, convolutional neural networks (CNNs) have achieved remarkable advances in the medical fields. Specifically, CNNs are used in mammography for lesion localization and detection, risk assessment, image retrieval, and classification tasks. CNNs also help radiologists providing more accurate diagnosis by delivering precise quantitative analysis of suspicious lesions.
Results
In this survey, we conducted a detailed review of the strengths, limitations, and performance of the most recent CNNs applications in analyzing MG images. It summarizes 83 research studies for applying CNNs on various tasks in mammography. It focuses on finding the best practices used in these research studies to improve the diagnosis accuracy. This survey also provides a deep insight into the architecture of CNNs used for various tasks. Furthermore, it describes the most common publicly available MG repositories and highlights their main features and strengths.
Conclusions
The mammography research community can utilize this survey as a basis for their current and future studies. The given comparison among common publicly available MG repositories guides the community to select the most appropriate database for their application(s). Moreover, this survey lists the best practices that improve the performance of CNNs including the pre-processing of images and the use of multi-view images. In addition, other listed techniques like transfer learning (TL), data augmentation, batch normalization, and dropout are appealing solutions to reduce overfitting and increase the generalization of the CNN models. Finally, this survey identifies the research challenges and directions that require further investigations by the community.
Electronic supplementary material
The online version of this article (10.1186/s12859-019-2823-4) contains supplementary material, which is available to authorized users.
Keywords: Mammograms (MGs), Breast cancer, Deep learning (DL), Convolutional neural networks (CNNs), Machine learning (ML), Transfer learning (TL), Computer-aided detection (CAD), Classification, Feature detection
Background
Breast cancer is the second most common cause of cancer death in women. According to the American cancer society’s latest statistics, it is estimated that 40,610 women in the USA are expected to die in 2017 from breast cancer. As of March 2017, there are more than 3.1 million women with a history of breast cancer in the USA [1]. Mammography is one of the most widely used methods for breast cancer screening and has contributed significantly to the reduction of the mortality rate through early detection of cancer [2]. However, the complexity of mammograms (MGs) and the high volume of exams per radiologist can result in false diagnosis [3, 4].
Computer-aided detection (CAD), which employs image processing techniques and pattern recognition theory, has been introduced to provide an objective view to radiologists [2]. Studies have shown the effectiveness of CAD models; however, accurate detection of breast cancer has remained challenging [2]. Recent studies show that CAD models cannot improve significantly the diagnostic accuracy of mammography [5]. The biggest challenge in using CAD for abnormality detection in MGs is the high false positive rates (FPR). False positives result in patient anxiety, additional radiation exposure, unnecessary biopsies, high callback rates, increased health care costs, and additional assessment [4]. In the USA, millions of women undergo screening mammography each year, as a result, even a small reduction in the FPR result in a widespread benefit [1, 6]. The limitations of current CAD indicate the need for new, more precise detection methods.
Recent advances in computational technologies, significant progress in machine learning and image processing techniques, and prevalence of digital MG images have opened up an opportunity to address the challenging issue of early detection of breast cancer using deep learning (DL) methods [7–10]. Recently, DL methods, specially convolutional neural networks (CNNs, also known as ConvNets) have gained lots of attentions to CAD for MGs as they help overcome CAD systems’ limitations [2, 8, 9, 11]. CNNs achieve higher detection accuracy than CAD models, and help radiologists provide more accurate diagnosis by delivering quantitative analysis of suspicious lesions [10, 12–14]. A recent research study shows that using DL methods drop human error rate for breast cancer diagnoses by 85% [15]. Current CNN models are designed to improve radiologists’ ability to find even the smallest breast cancers at their earliest stages alerting the radiologist to the need for further analysis [12, 15].
Recent studies used CNNs to generate a standard description of lesions, which can help radiologist in making a more accurate decision [12, 14]. Moreover, advances in CNNs can not only aid radiologists, but also eventually make diagnosis systems to read MGs independently in the near future [12]. In the last few years, CNNs have led to breakthroughs in a variety of pattern recognition and classification problems for natural images due to the availability of big data repositories, fast graphical processing units, and the power of parallel and distributed computing [7, 10, 16, 17].
Training a deep CNN model with a limited number of medical data is very challenging, which has been addressed by using transfer learning (TL) and augmentation techniques [7, 16, 18]. Studies show that CNN methods that compare images from left and right breasts [19] and also the craniocaudal (CC) and mediolateral-oblique (MLO) view of each breast can improve the accuracy of detection and reduce the false positives [15, 20–25]. CNNs have also been used in the risk assessment applications to increase the accuracy of early detection breast cancer by radiologist [26–35]. In this work, we summarize almost all contributions, as of November 2017, to the field of DL in MGs, in particular using CNNs.
Methods
Criteria for inclusion/exclusion of studies in the survey
We carried out a comprehensive literature research, using the defined keywords given in Table 1, on journals and proceedings of scientific conferences including, but are not limited to the following scientific databases: Scopus, ACM Digital Library, Science Direct, IEEE Explore Digital Library, PubMed, Web of Science. In total, we considered 83 studies from the time period of 1995 to Nov 2017. These studies focus on implementing CNNs for lesion localization and detection, risk assessment, image retrieval, high resolution image reconstruction and classification tasks in MG images. The inclusion/exclusion criteria we used for this review are presented in Table 1. Figure 1, shows a breakdown of the studies included in this survey in the year of publication grouped by their neural network task.
Table 1.
Category | Criteria |
---|---|
Time period | Published from 1995 to the present (Nov 2017). |
Databases | Private and public databases. |
Publication | English articles in print. |
Excluded articles accepted for publication before appearance in journals or conferences as of Aug 2017. | |
Research focus | All Implementation of CNNs for breast cancer in Mammography. |
Keywords | Deep learning, convolutional neural networks, breast cancer, mammography and transfer learning. |
Abnormalities | Mass, calcification, architectural distortion and asymmetries. |
In this study, we addressed the following research questions:
Does this study focus on using a CNN for detecting abnormalities in MGs?
What is the task of the implemented CNN?
What are the databases, database size, image resolution, image type, abnormalities involved in the development of the CNN?
What are the methodologies used for the setup and pre-processing of the data-set?
Can deep networks perform well on medical images specifically MGs?
What are the learning methods used for training the CNNs?
What are the best practices that were applied to increase the accuracy of detection of abnormalities?
What are the advantages and limitations presented by the methodologies employed in CNNs?
Is it an end-to-end (E2E) training method?
Is transfer learning from natural imagery to the medical domain relevant?
Is combining learned features with hand-crafted features will enhance the accuracy of certain mammographic task?
What are the common toolkits used in mammography?
What are the challenges to train deep neural network for mammography data-set?
How imbalanced data-sets impact the performance of CNNs?
What is the common cross-validation method used with MGs?
Which activation functions are commonly used for training MGs?
Breast cancer digital repositories
Mammographic databases play an important role in training, testing, and evaluation of DL methods. The amount of data needed to train a DL network is massive compared to the data needed to train traditional neural networks. The availability of comprehensive annotated databases is critical for advancing DL development in medical imaging. The most common findings seen on mammography are abnormal areas of mass, calcifications (MCs), architectural distortion (AD), and asymmetries. There are common publicly available databases for MGs: the Mammographic Image Analysis Society (MIAS) database [36], Digital Database for Screening Mammography (DDSM) [19], INbreast database [37], Breast Cancer Digital Repository (BCDR) [38], Image Retrieval in Medical Applications (IRMA) [39].
Table 2 compares the publically available MG databases according to the origin, the number of images, size of images, views (CC, MLO), digital or film database, the format of images, resolution of images, and the distribution of normal, benign and malignant images. Other databases used in literature are private and restricted to individual organizations [21, 26, 27, 31, 34, 40–46]. The public databases present a wide variability of patients’ cases and a mixture of normal, benign, and malignant cases. Annotations include the location and boundaries of the lesions performed by imaging specialists. The public repositories have collected film screen MGs (FSMs) [36, 38, 39], and/or digital mammography (FFDM) [37–39, 47] with different resolutions. Digital MG images are usually saved in the DICOM format that gathers not only the image but also some related meta-data as in [37, 38, 47]; however, some databases use different formats [36, 38, 39, 48].
Table 2.
Database | Image-size | Views | Type | Format | bpp | #Normal | #Benign | #Malignant |
---|---|---|---|---|---|---|---|---|
DDSM | 3118 ×5001 | Both | FSM | LJPEG | 12 | 914 | 870 | 695 |
IRMA | Several | Both | Both | PNG | 12 | 1108 | 1284 | 1284 |
INbreast | Several | Both | FFDM | DICOM | 16 | 67 | 220 | 49 |
MIAS | 1024 ×1024 | MLO | FSM | PGM | 8 | 207 | 69 | 56 |
BCDR-F01 | 720 ×1168 | Both | FSM | TIF | 8 | 0 | 187 | 175 |
BCDR-F02 | 720 ×1168 | Both | FSM | TIF | 8 | 0 | 426 | 90 |
BCDR-F03 | 720 ×1168 | Both | FSM | TIF | 8 | 0 | 426 | 310 |
BCDR-D01 | Several | Both | FFDM | DICOM | 14 | 0 | 85 | 58 |
BCDR-D02 | Several | Both | FFDM | DICOM | 14 | 0 | 405 | 51 |
BCDR-DN01 | Several | Both | FFDM | DICOM | 14 | 200 | 0 | 0 |
The images of the MIAS database are of low resolution and have strong noise. The MIAS database is an old database that contains a limited number of images. Despite all these drawbacks, it has been widely used in literature until now [49–51]. DDSM is a huge repository used in many studies [23, 24, 32, 49, 52–65]. DDSM images are saved in non-standard compression files that require use of decompression codes. Moreover, the Region of Interest (ROI) annotations for the abnormalities in the DDSM images indicate general position of lesions, without precise segmentation of them. The IRMA project is a combination of a number of databases of different resolution and sizes. The ROI annotations for these databases are more precise making them more accurate for supervised DL methods. The INbreast database is gaining more attention nowadays and used in [25, 32, 57, 66–70]. Its advantages are high resolution and accurate segmentation of lesions; however, its small size and the limited shape variations of the mass are its drawbacks. BCDR is a promising database but still is in its development phase. BCDR has been used in few studies [71–74]. The strengths and limitations of these databases are summarized in Table 3.
Table 3.
Database | Strength | Limitation |
---|---|---|
DDSM | Big widely used database. | Non-standard format. |
Shape variations of different lesions. | Not precise position of lesions. | |
IRMA | Accurate position of lesions. | Non-standard format. |
High resolution. | ||
INbreast | Accurate position of lesions. | Limited size. |
Limited mass shape variations. | ||
Standard file format. | ||
Old database. | ||
No more supported. | ||
MIAS | Still widely used. | Limited size. |
Images are of low resolution. | ||
Has MLO view only. | ||
Different resolutions. | ||
BCDR | Accurate position of lesions. | Limited size. |
Standard file format. | ||
Still in their development phase. |
Convolutional neural networks
In fact, DL is not a new idea, which even dates back to 1940s [7, 75] for medical images. Shallow layer CNNs were used to investigate breast cancer in 1995 [40, 76]. Famous CNNs such as Alex-Net [16], ZF-Net [77], GoogLeNet [78], VGG-Net [79] and ResNet [80] have brought about breakthroughs in processing images. Alex-Net architecture is extensively used in medical imaging for breast cancer detection. DL is a subset of machine learning that requires a huge number of labeled data to train the models. The term “deep” usually indicates the number of hidden layers in neural networks, e.g. ResNet has a depth of 152 layer which is 8 × deeper than VGG-Net. Since 2012, CNNs have become more popular and have attracted more attention because of the increasing computing power, availability of lower cost hardware, open source algorithms, and the rise of big data [16].
The structure of CNNs is very similar to that of ordinary neural networks. The basic CNN architecture is a stack of convolutional layer (Conv), nonlinear layer (e.g. ReLU), pooling layer (e.g. Max-pooling), and a loss function (e.g. SVM/Softmax) on the last fully connected (FC) (Fig. 2). The output can be a single class (e.g. normal, benign, malignant) or a probability of classes that best describes the image. The input to a convolutional layer is a W1 ×H1 ×D1 image where W1 is the width and H1 is the height of the image and D1 is the number of channels, e.g. an RGB image has D1=3. The convolutional layer will have F filters (e.g. 12 filters) of size N ×N×D1 where N is smaller than the dimension of the image and D1 is the same as the number of channels (e.g. 5 ×5×3 (i.e. 5 pixels width and height, and 3 because images have depth 3, the color channels).
During the convolution operation, each of the F filters convolves with the image to produce K feature maps of volume size W2 ×H2 ×D2 where: W2=H2=(W1-F+2P) /S+1, S is the number of strides, D2=F, and P is the amount of zero padding. For each feature map, a non-linear activation function is applied (e.g. ReLU). A non-linear activation function leaves the size of the volume unchanged (W2 ×H2 ×D2). After applying ReLU, a down-sampling operation called Pool is applied along the spatial dimensions (width, height) of the result feature map. After pooling, there may be any number of fully connected layers that compute the class scores (Fig. 2). More details about the architecture of CNNs can be found in [16, 81].
Popular CNNs
Alex-Net [16], ZF-Net [77], GoogLeNet [78], VGG-Net [79] and ResNet [80] have been extensively used as a pre-trained networks to classify images for medical domains instead of training a network from scratch. Table 4 shows the configurations of the most popular CNNs. Generally, training a deep CNN requires extensive computational and memory resources. Training these networks from scratch typically takes days or weeks on modern GPUs (Table 4). All these networks were trained on the 1000 object category classification on the ImageNet data-set [82]. The ImageNet data-set consists of a 1.2M image training set, a 50K image validation set, and a 100K image test set. Two error rates are reported for these networks: top-1 and top-5, where the top-5 error rate is the fraction of test images for which the correct label is not among the five labels considered most probable by the model. All these network architectures use the data augmentation technique to prevent overfitting with dropout initially set to 0.5.
Table 4.
AlexNet [16] | ZF-Net [77] | GoogLeNet [78] | VGG-Net [79] | ResNet [80] | |
---|---|---|---|---|---|
Year | 2012 | 2013 | 2014 | 2014 | 2015 |
Image Resolution | 227 ×227 | 227 ×227 | 224 ×224 | 224 ×224 | 2244 ×224 |
Number of layers | 8 | 8 | 22 | 19 | 152 |
Number of Conv-Pool layers | 5 | 5 | 21 | 16 | 151 |
Number of FC layers | 3 | 3 | 1 | 3 | 1 |
Full connected layer size | 4096,4096,1000 | 4096,4096,1000 | 1000 | 4096,4096,1000 | 1000 |
Filter Sizes | 3, 5, 11 | 3, 5, 11 | 1,3,5,7 | 3 | 1,3,7 |
Number of Filters | 96 - 384 | 96 - 384 | 64 - 384 | 64 - 512 | 64 - 2048 |
Strides | 1, 4 | 1, 4 | 1, 2 | 1 | 1, 2 |
Data Augmentation | + | + | + | + | + |
Dropout | + | + | + | + | + |
Batch Normalization | - | - | - | - | + |
Number of GPU | 2 GTX | 1 GTX | A few high-end | 4 Nvidia | |
580 GPUs | 580 GPUs | GPUs | Titan Black GPUs | Titan Black GPUs | 8 GPUs |
Training Time | 5:6 days | 12 days | 1 week | 2:3 weeks | 2:3 weeks |
Top-5 error | 16.40% | 11.2% | 6.70% | 7.30% | 3.57% |
Alex-Net [16] was the first CNN to win the ImageNet Challenge in 2012. AlexNet’s CNN consists of five Conv layers and three fully connected (FC) layers. Within each Conv layer, there are 96 to 384 filters and the filter size 3 ×3, 5 ×5, 11 ×11, with 3 to 256 channels each. A ReLU non-linearity is used in each layer. Max-pooling of 3 ×3 is applied to the outputs of layers 1, 2 and 5. Alex-Net used a stride of 4 at the first layer of the network. AlexNet’s model requires 61M weights to process one 227 ×227 input image (top-5 error of 16.40%). ZF-Net [77] is a slightly modified version of Alex-Net model and uses an interesting way of visualizing their feature maps. In ZF-Net, the used visualization technique give insight into the function of intermediate feature layers and the operation of the used classifier. The VGG-Net [79] model reinforces that the CNNs have to have a deep network of layers. GoogLeNet [78] has 22 layers. It introduced an inception module to the CNN model. It has pieces of the network that are working in parallel in contrast to previous CNN models, which have only a single serial connection. ResNet [80], also known as Residual Net, uses residual connections to go even deeper. ResNet determines an object’s exact location, which is a huge jump in CNNs. ResNet is 8 × deeper than VGG-Net with lower complexity. The ResNet with 152 layers was the winner of the ImageNet challenge 2015 [82] (top-5 error of 3.57%). it has 60M weights. YOLO is another famous CNN named that is recently used for object classification and localization while processing the image only once, as is implied by it’s name, You Only Look Once [83, 84]. Table 4 shows that the number of layers are going deeper and deeper within the newer implementations as in ResNet.
Results
CNNs best practices
In this section, we explain the practices that contribute to improve the performance of CNNs for MGs. It goes beyond the scope of this paper to discuss all the best practices done in CNNs in general, but we are going to highlight and focus on some of them that show significant changes in the classification accuracy when applied to MG images. Recent survey papers [7, 8, 85] discuss more trends for natural images.
Data preparation
Pre-processing of MG images
Pre-processing of MG images is an essential task before training CNNs [63, 66, 71, 72, 86]. The pre-processing consists of contrast enhancement, noise removal, and breast segmentation. Breast segmentation includes the remove of the background area, labels, artifacts, and pectoral muscle which disturb the detection of Mass/MCs [45, 50]. It is important to have good separation between foreground and background pixels and do not remove the important information in images [59, 87, 88]. The commonly used filters for image enhancement and noise reduction are the adaptive mean filter, median filter, and contrast limited adaptive histogram equalization (CLAHE) [62, 89–92].
Image size, cropping, and down-sampling
Most studies have used segmented ROIs in order to reduce the computation of the CNNs and to avoid the issue of small training data. These ROIs can be obtained by a manual segmentation of the images using the available ground truth data, or an automatic detection system. The ROIs are cropped and re-scaled to r ×r pixels with the lesion centered within the image. However, using very small subsampled (e.g. 32 ×32) patches may not contain enough detail to improve the classification results as in [40, 41, 44, 63, 66, 67, 70, 74, 93].
Two strategies have been utilized to use full image size for training CNNs on MGs instead of ROIs. The first strategy, down-sample high resolution images to ≈250×250. However, the requirement to find small mass regions or MCs clusters in down-sampled high resolution images is unlikely to be successful for MGs [65]. The second strategy, train a patch-level CNN classifier, which is then used as feature extractor to an image-level model. In the image-level model, each image is partitioned into a set of patches with a minimal overlap such that each patch is contained entirely within the image. Final classification involves aggregation across patches and the CC & MLO views [65].
Mixing databases
In literature, researchers mix several databases to analysis their CNNs. The fusion from different image type (FSM and FFDM) assists CNNs in term of detection rate. Researchers in [32, 49, 51, 52, 57, 94, 95] compared both image quality and detection on FFDM and FSM databases. They have shown that a CNN using FFDM images gives better detection rate than using FSM images. Moreover, these studies show that DL training using the fusion of both FFDM and FSM lower the number of false detections [93, 94].
Learned and hand-crafted features
The hand-crafted features (i.e. Haar-like features, histogram of oriented gradients (HOG), and histogram of the gradient divergence (HGD)) are commonly used with traditional machine learning approaches for object recognition like support vector machines. CNNs are able to extract features from the input image data-sets. Thus, CNNs remove the necessity of the time-consuming hand-crafted features.
However, the authors in [21, 23, 54, 71, 87, 96–100] have demonstrated the importance of combining the extracted features using deep CNNs with hand-crafted features like texture, and shape. Interestingly, the combination of both representations (learned and hand-crafted features) resulted in a better descriptor for Mass/MCs lesion classification [71, 100]. The reason behind using hand-crafted features is that the learning process should be guided by a training data-set that has a wide variability of texture and shape features. For example, Dhungel et al. [97] proposed two-step training process involving pre-training based on a large set of hand-crafted features. The second stage fine-tunes the features learned in the first stage to become more specialized for the classification problem.
Using hand-crafted features depends on the size of the data-set. With small training data-set, generating hand-crafted images could result in a better model for Mass/MCs lesion classification. Also, employing some hand-crafted features that specifically target small and missed lesions is a more effective strategy than adding extra cases to the train a data-set. Thus, the performance of CNNs trained with small data-set can be improved by incorporating hand-specified features to deal with cases that cause false positives or false negatives [23].
Hyper-parameters
Hyper-parameters are variables which determine the network structure (e.g. number of hidden layers), and the variables which determine how the network is trained (e.g. learning rate). Hyper-parameters are manually chosen before training the CNNs.
Data augmentation
Data augmentation is an appealing solution to reduce overfitting and increase the generalization of the model and boost the performance. Overfitting happens in CNNs when the models learn too well the details from training data, but they do not generalize well from the training data, in order to make good predictions about the future unseen data. As a result, the performance of the trained model is poor for testing data. That usually happens when the size of training data-set is too small compared with the number of model parameters that need to be learned.
Data augmentation artificially creates new sample images by applying transformations like flipping, and rotation to the actual data. Common data augmentation techniques for mammography images are horizontally flipping, rotations (90, 180, and 270 degrees), jittering, and random scaling. Such data augmentation generates relevant training samples because tumors may present in various orientations and sizes. Thus, augmentation techniques do not change the underlying pathology of the masses. Data augmentation has been employed by many studies [12, 22, 23, 30, 33, 34, 40–45, 50, 53–55, 57, 58, 63, 66, 70–73, 96, 97, 101–110].
Going deeper
In CNN, the design of the network architecture completely depends on the model requirements and the size of the data-set. The CNNs in [53, 66, 96] have a fewer number of layers but show good accuracy. However, the work done in [63, 72, 73, 78] shows that we can get better performance in term of higher area under the ROC curve (AUC) as the architecture goes deeper and trained on more data. Deep architectures can lead to abstract representations because more abstract shapes can often be constructed in terms of less abstract ones captured in earlier layers. Adding more layers will help the model to extract more features. But adding more layers can be done to a certain extent and there is a limit. After that, instead of extracting features, it results in overfitting the network that can lead to false positives. Adding more hidden layers will promote the accuracy for large data-sets. Adding layers unnecessarily to a CNN will increase the number of parameters, and for a smaller data-set, it will reduce accuracy of the test data. Deep architectures are often challenging to train effectively, and this has been the subject of more recent research. Choosing a smaller network or a larger one cannot be estimated theoretically. A trade-off between accuracy and deep networks need to be done with trial and error method and some experience and practice on the basis of the data-set.
Learning rate
Learning rate (LR) is one of the most important hyper-parameters, which influences the CNNs’ performance. Deep learning models are typically trained by a stochastic gradient descent optimizer. There are many variations of stochastic gradient descent as Adam, RMS Prop, Adagrad, etc. All these optimizers let users set the learning rate. Learning rate controls how much the network parameters are adjusted in order to minimize the network’s loss function. If the LR is too small, the CNN will converge after many iterations to the best values. However, if LR is too high, it can cause undesirable divergent behavior in the loss function. Famous learning rate policies are step decay, quadratic decay, square root decay and linear decay [85]. A common practice when dealing with MG images, is to use a step decay rate where the LR is reduced by some percentage after a set number of training epochs. For example, Yi et al. [23] used a learning rate of 0.001 with decay rate of 0.99 per epoch, and a regularization coefficient of 10−5 for training their CNN. Another common practice is to use a small learning rate (e.g. 0.001) to train a pre-trained network, since we expect well-adjusted pre-trained weights compared to randomly initialized weights.
Activation functions
Recently, many variations of rectified linear unit (ReLU) function have been proposed for activation function such as leaky ReLU, parametric ReLU, and randomized ReLU [111]. There are other popular activation functions such as sigmoid, and tanh. The activation functions bring non-linearity into CNNs. Sigmoid presents a serious disadvantage called the vanishing gradient problem. In the vanishing gradient problem, the gradient of small input values to sigmoid functions tends to get smaller (close to zero) as gradients are computed backward through the hidden layers, resulting in slow learning in the earlier layers of the model. Slow learning is highly avoided in DL since it results in expensive and tedious computations [112].
ReLU became a popular choice in DL and even nowadays provides outstanding results as it solves the vanishing gradient problem [111]. ReLU has gradient one for positive inputs and zero for negative inputs. As long as values are above zero, the gradient of the activation function will be one, meaning that it can learn anyways. This solves the vanishing gradient problem present in the sigmoid activation function. On the downside, once the gradient is zero the corresponding nodes do not have any influence on the network anymore, which is known as “dying ReLU” problem. Leaky ReLU is one attempt to overcome the dying ReLU problem [113]. Instead of the output of ReLU being zero when input is less than zero, a leaky ReLU will provide a small negative slope (α of 0.01, or so). This small slop reduces the sparsity but, on the other hand, makes the gradient more robust for optimization, since in this case, the weight will be adjusted for those nodes that were not active with ReLU. When the slop is not constant (e.g. 0.01) then it is called randomized ReLU.
A detailed explanation of the advantages and disadvantages of different activation functions are discussed in [16, 111, 112]. Theoretically, leaky ReLU is in general better than ReLU. However, ReLU has been chosen as an activation function in most of the CNNs for MGs as it allows faster learning [58, 64, 65, 70, 114, 115].
Techniques for improving the CNNs performance
Dropout
Dropout is a regularization technique proposed in [116] that superior the other regularization methods (L1, L2, Max norm). Dropout prevents a CNN model from overfitting. This technique randomly selects neurons and ignore them during training. They are “dropped-out” randomly. This means that their contribution to the activation of downstream neurons is temporally removed on the forward pass and any weight updates are not applied to these neurons on the backward pass [16]. Smirnov [117] has shown a comparison of regularization methods with deep CNNs and showed that the dropout technique is in general better than other regularization techniques. The authors in [12, 22, 25, 44, 58, 70, 73, 94, 106, 106, 108] have used dropout in their work with MGs. The dropout of 0.5 is a common value for mammography images.
Batch normalization
In a CNN model, a batch normalization (BN) layer normalizes input variables across a mini-batch (a subset of the training data-set). First, the BN layer normalizes the activations of each channel by subtracting the mini-batch mean and dividing by the mini-batch standard deviation. Then, the BN layer shifts the input by a learnable offset β and scales it by a learnable scale factor γ, thus reduces the networks’ internal covariant shift. BN speeds up training of CNNs and reduce the sensitivity to network initialization. According to [118], BN allows the use of much higher learning rates and less care about initialization as it acts a regularize. BN results in faster convergence and as a consequence overall faster training for a CNN. Besides that, BN regulates the values going into each activation function. With BN, saturating nonlinear activation functions (e.g. sigmoid) that do not work well in deep networks tend to become viable again. Similar to dropout, BN adds some noise to each hidden layer’s activations. Therefore, using BN causes less dropout value. BN has been used in CNNs for MG images [65, 73, 101]. For mammography, it is recommended to not depend only on BN for regularization; and to use it together with dropout.
Transfer learning
Training a deep CNN requires large amounts of labeled training data [11]. Only few studies train an entire CNNs from scratch with random initialization; and the rest use TL approaches either fine-tune a pre-trained network [46, 52, 53, 58, 60, 63, 72, 73, 94, 110, 119, 120] or use a pre-trained network as feature extractor [15, 32, 46, 70]. Recent overviews of TL in deep network models are given in [37, 45, 46, 65]. The need for TL in medical domain occurs because data are scarce and expensive, they are not publicly available, and it is time-consuming to collect and label them by professional radiologists [17, 46, 55, 121–124]. Moreover, training a deep CNN requires extensive computational and memory resources [16, 17, 78].
References [60, 77, 125] show that the main power of a CNN lies in its deep architecture. Extracted features of earlier layers of a pre-trained CNN (i.e. on natural images) contain more generic features (e.g. edge detectors or blob detectors) that are useful for many tasks; but in later layers, generic features are combined and become more specific to the details of the classes contained in the training data-set. Thus, a deep CNN allows extracting a set of discriminating features at multiple levels of abstraction which can be transferable from one domain to another. However, the required level of fine-tuning differs from one application to another. Tajbakhsh et al. [125] show that neither shallow tuning nor deep tuning may be the optimal choice for a particular application. Moreover, layer-wise fine-tuning may offer a practical way to reach the best performance for a certain application and should be chosen experimentally. In addition, the work in [21, 106, 109] has achieved a good performance on a small data-set by pre-training the network on a large data-set of general medical images.
Most of the studies employed TL have used ImageNet’ data-set [82] for pre-training their network [46, 58, 60, 72, 94, 95, 110, 126–128]. The commonly used pre-trained CNNs architectures for mammography are Alex-Net [46, 50, 58, 60, 72, 94, 95, 110, 127, 128], VGG16 [50, 127, 129], ResNet50 [127, 129] and GoogLeNet [58, 72, 127]. All the deep CNN architectures that are pre-trained using ImageNet are designed for a 1000-class classification task. To adapt them to the task at hand, the last three layers are removed from each network and a three new layers (FC layer, soft-max layer, and classification layer) are appended to the remaining structure of each network.
Until large-scale medical image data-sets for mammography became available, the combination of TL and data augmentation is a very promising approach for training deep CNNs. By visualizing the features learned at different layers during the training process, a model can be monitored to closely observe and track its performance [23]. Learned features can indicate whether a model is successfully learning or not, allowing a user to stop the training process early [130].
Cross-validation
Cross-validation is a statistical technique to evaluate predictive models by partitioning the original samples into a training set to train the model, and a test set to evaluate it. There are three common types used in literature for validation, the hold-out splits [76, 131], three-way data splits [8, 22, 58, 65, 96], the K-fold cross-validation [20, 23, 25, 26, 49, 94, 110, 115, 132, 133]. In the hold-out data splits, data is split into training set and test set (e.g. 80%, and 20%, respectively). The training set is used to train the model and the test set is used to estimate the error rate of the trained model. In the three-way data splits, data are randomly split into training, validation and testing sets. The CNN model is trained on the training set and is evaluated on the validation set. Training and validation may be iterated a few times till the best model is found. The final model is assessed using the test set.
In the K-folds cross validation, data are split into k different subsets (or folds). The cross-validation process is repeated K times (the folds), with each of the K sets used exactly once as the test set. The K error estimates from the folds can then be averaged to produce a single estimation. Cross-validation avoids overfitting and gives a less biased estimate of the performance of the model [67, 134]. In practice, the choice of the number of folds depends on the size of the data-set. In literature, the common strategy is to use K-fold cross validation for mammography. For large data-sets, it is a common choice to use 3 to 5-fold cross-validation. For small mammography data-set, it is a common choice to use 10-fold cross-validation.
Context and patient information
Integrating some information such as patient age, breast density and other context like the view type (CC or MLO) into a CNN method can improve the detection rate of CNNs [96]. Multi-modal machine learning aims to build models that can process and relate information from multiple modalities (e.g. images and text) with a score level fusion at the final prediction results.
Multi-view and single-view images
It is a good practice to use both CC & MLO views to detect abnormalities. A true abnormality can usually be detected on two different views of a MG. Recent studies in [15, 20–25, 95, 107] lead to significant improvements of multi-view (MV) approaches compared to single-view (SV) ones, demonstrating that the high-level features of the individual CNN models provide a robust representation of the input images. Comparing two views can aid in the reduction of false positives and false negatives.
Balanced and imbalance distribution
A couple of publicly available databases (e.g. INbreast, DDSM) are constructed to include approximately the same proportions of normal and abnormal cases, which is a balanced distribution of classes. Other databases called imbalance distribution (natural distribution) databases, which include unequal proportions of normal and abnormal cases. Training CNN models directly on imbalanced data-sets may bias the prediction towards the more common classes like normal, resulting in false negatives. Whereas the minority ones are misclassified frequently [135]. The authors in [20, 21, 32, 45, 46, 53, 56, 58, 73, 74] have pointed out that the balance of the number of samples per class has a great impact on the performance of the system. However, the authors in [22, 44, 96] used a natural distribution databases. According to [136] choosing a wrong distribution or objective function while developing a classification model can introduce bias towards potentially uninteresting class (non-cancerous). For MG images, it is preferable to use a balanced data-set. Different approaches to handle imbalanced data-sets include random under-sampling and random over-sampling techniques [135]. Random under-sampling aims to balance class distribution by randomly eliminating majority class samples (normal cases). This is done until the majority and minority class instances are balanced out as done in [74]. In the other side, over-sampling increases the number of instances in the minority class (abnormal cases) by randomly replicating them in order to present a higher representation of the minority class. Unlike under-sampling, over-sampling leads to no information loss.
The appropriate approach (random under-sampling or random over-sampling) depends on the amount of available data-set and the specific problem at hand. Researchers empirically test each approach and select the one that gives them the best results. In the case of using an imbalance data-set, accuracy is not a right metric to evaluate the performance of the model. There are more appropriate scores when using imbalanced data-sets such as F1-score [136] that combines the trade-offs of precision and recall, and outputs a single number reflecting the goodness of a classifier.
Multi-stage and end-to-end (E2E) methods
A multi-stage pipeline used for detection and classification of a lesion consists of multiple stages such as pre-processing, image segmentation, feature detection, feature selection and classification stages [137, 138]. End-to-end (E2E) deep learning methods take all these multiple stages and replace it with just a single neural network. Researchers in [12, 15, 24, 40–42, 60, 67, 102, 105] have used one or more stages of this multi-stage pipeline in their CNN systems. In their multi-stage method, a CNN is trained to determine whether a small patch has Mass and/or MCs. Other researchers focused on training a deep CNN for classifying a small ROI or full image into benign or malignant, assuming an existing Mass/MCs detection system as in [23, 25, 43, 50, 55, 56, 62, 66, 71, 72, 87, 104]. In multi-stage methods for CNNs, several cascaded classifiers are trained independently, each classifier makes a prediction, and all predictions are combined into one using different strategies. Dhungel et al. has found that the multi-stage methods are effective in the reduction of false positive detection [97]. Moreover, researchers in [22, 25, 30, 45, 57, 96, 98, 107] used the E2E methods.
E2E methods for MGs are better than multi-stage method when training a CNN with a large data-set. But if the data-set is small in size, then the learning algorithm cannot capture much insight from data. Excluding potentially useful hand-crafted features that are very helpful if well designed is the downside of the E2E approaches. Therefore, the key parameter to choose using E2E deep learning approach is having sufficient data to learn the model.
Toolkits and libraries for deep learning
Implementing a DL network from scratch is an exhausting process and probably beyond the skills of most medical imaging researchers. It is much more efficient to utilize the publicity available resources. Some criteria should be considered while choosing a library and toolkit including its programming language for interface, the quality of documentation of the toolkit, the ease of programming, the runtime to do thousands of calculations per pixel, the training speed, GPU support for faster performance [17], and lastly its popularity among experts. Recent surveys done in [139, 140] discus the most famous and recent toolkits and libraries used generally for DL. The common toolkits used in training CNNs for mammography are Tensorflow [141], Keras, Caffe [142, 143], PyTorch [144] and MatConvNet [145]. Table 5 gives a comparison between these libraries and their ranking based on the forks received by the community on GitHub.
Table 5.
Interface | Languages | Open source | CUDA support | Pre-trained models | Forks (Github) | Contributions (Github) | |
---|---|---|---|---|---|---|---|
TensorFlow | Python | C++, Python | Yes | Yes | Yes | 63,603 | 1,481 |
Keras | Python, R | Python | Yes | Yes | Yes | 11,203 | 681 |
Cafee | Python, Matlab, C++ | C++, Python | Yes | Yes | Yes | 14,868 | 267 |
PyTorch | Python | C, Python, CUDA | Yes | Yes | Yes | 3,592 | 644 |
MatConvNet | Matlab | CUDA | Yes | Yes | Yes | 651 | 24 |
Tensorflow is one of the most popular DL libraries, it was developed by the Google Brain team and open-sourced in 2015 [141]. Tensorflow is a Python-based library capable of running on multiple CPUs and GPUs. It can be used directly to create deep learning models, or by using wrapper libraries (e.g. Keras) on top of it. Tensorflow does not contain many pre-trained models and there’s no support for external data-sets, like Caffe. The framework is written in C++ and Python and has large amount of available documentation. As of today it is the most commonly used deep learning framework.
Keras is a very lightweight open source library, easy to use, and pretty straightforward to learn. It was built as a simplified interface for building efficient deep neural networks in just a few lines of code and use Tensorflow as back-end.
Caffe is one of the first deep learning libraries developed mainly by Berkeley vision and learning center (BVLC) [142, 143]. It is a C++ library which also has a Python interface and finds its primary application in modeling CNNs. Caffe provides a number of pre-trained networks directly from the Caffe Model Zoo, available for immediate use.
PyTorch is a Python library enabling GPU accelerated tensor computation, similar to NumPy. A few advantages of using PyTorch are it’s multi-GPU support, dynamic computational graphs, custom data loaders, optimization of tasks, and memory managements. PyTorch provides a rich API for neural network applications [144]. PyTorch is used by many companies such as Twitter, Facebook and Nvidia to train DL models.
MATLAB has a neural network toolbox that provides algorithms to create, train, visualize deep neural networks. TL can be done with pre-trained deep CNNs models (including Inception-v3, ResNet-50, ResNet-101, GoogLeNet, Alex-Net, VGG-16, and VGG-19) and models imported from Keras or Caffe. MATLAB allows computations and data distribution a across multi-core processors and GPUs with the parallel computing toolbox. MatConvNet [145] is an open source implementation of CNNs with a deep integration in the MATLAB environment.
Applications of deep CNNs for mammography
After describing deep CNNs in the previous section, and different practices that are famous for mammography, we will now turn our focus to how these are used for recognition purposes for mammography. More specifically, we review recent deep CNNs’ applications in mammography such as classification, localization, image retrieval, high resolution image reconstruction and risk analysis. We summarized these recent works in Additional file 1: Table S1.
Lesion classifications and detection
The detection of lesions in mammography is a common task for CNNs. In contrast to lesion detection, classification of MGs into benign and malignant is a challenging task that many studies try to address it. The authors in [12, 15, 20, 21, 24, 40, 41, 44, 52, 60, 67, 70, 90, 93, 102, 105, 120] are interested in lesion classification into two classes. They developed a CNN to predict a probability of being normal (NL), contain mass and/or MCs. The studies in [23, 43, 46, 49, 50, 55, 56, 58, 59, 62, 63, 66, 69, 71–73, 94, 99, 100, 104, 132, 146–150] present deep CNN methods to classify the MG images into 2 classes (benign or malign), or three classes (benign, malign or without tumor). The authors in [32, 95] studied the development of malignancy of mass(es). The authors in [15, 40, 42, 44, 151] are interested in the classification and detection of MCs in mammography. Chan et al. [40] introduced one of the earliest application of CNNs to detect clustered MCs. The authors applied enhancement filters for noise reduction on fifty-two FSM images. They observed that the shape of MCs in the breast is randomly oriented, thus they introduced an augmentation technique. Sahiner et al. [41] demonstrated the great effect of mixing CNN representation features and textural features (AUC of 0.873). Lo et al. [102] introduced a multiple circular path CNN coupled with morphological features of ROIs (AUC of 0.89). Sharma et al. [59] extracted geometrical features from MG images and used it with the representation features of their CNN. Their work demonstrates that DL methods are superior to traditional classifiers. Domingues et al. [67] used a shallow CNN that did not outperform traditional CAD methods, as they used a very small data-set to train their network and the selected normal ROIs did not represent every possible aspect of healthy breast tissue. Antropova et al. [100] developed a system incorporating both deep CNN and conventional CAD methods that performed statistically better than either one separately.
Sert et al. [63] stated that human level recall performance in detecting breast cancer considering MCs from MGs has a recall value between 74.5% and 92.3%. In [63], the authors reached a recall value of 94.0% above human level performance. Wang et al. [15] showed that breast arterial calcifications (BACs), detected in MGs, can be useful for identifying risk markers for having cancer. The authors in [15] showed that their CNN method achieves a level of detection similar to the human experts. Kooi et al. [12] employed a deep CNN with a large augmented data-set. Similar to the work of [15], the network in [12] performs similar to experienced radiologists, achieving AUC of 0.87 while the mean AUC of the experienced radiologists is 0.84. In [96], Kooi et al. proposed to use a random forest classifier for mass detection followed by a deep CNN that classifies each detected mass. Their method relies on a manually extracted features and features extracted from CNN layers. In [96], Kooi et al. trained their model on a large data-set and integrated additional information such as lesion location and patient information. Kooi et al. [149] following their work in [12, 96] employed a conditional random field (CRF) that is trained on top of CNN to model contextual interactions such as the presence of other suspicious regions. In [21], Kooi et al. employed a deep MV CNN using a pre-trained network on medical domain. They combined the extracted features using the deep CNN with hand-crafted features.
The studies in [46, 50, 58, 64, 72, 94, 106, 109, 120] demonstrated the use of TL in their work. The authors in [46, 50, 72] showed that CNNs in addition to TL can superior current CAD methods for tumor detection and classification based on small data-sets. Samala et al. [106] demonstrated that MGs can be useful for pre-training a deep CNN for mass detection in digital breast tomosynthesis (DBT). The similarity between masses in mammography and DBT can be observed from the ability of the DCNN in recognizing masses in DBT. In [94], Samala et al. demonstrated that CNNs with TL achieve better generalization to unknown cases than networks without TL. Similar to [94, 106], Hadad et al. [109] described a TL approach for using a pre-trained deep CNN on MGs to improve the detection accuracy of fine-tuned CNN on breast MRI lesions. Suzuki et al. [120] developed a deep CNN pre-trained on natural images, then the authors modified the last fully connected layer and subsequently train the modified CNN using 1,656 ROIs. Similar to [120], Jiao et al. [55] achieved an accuracy of 96.7% by applying fine-tuning on a pre-trained CNN on natural images to extract features for the next procedures. Jiao et al. [64] following his work in [55] proposed metric learning layers to further improve performance of the deep structure and distinguish malignant instances from benign ones. Levy and Jain [58] demonstrated that a fine-tuned pre-trained network significantly outperforms shallow CNNs.
Abbas [49] used speed-up robust features and local binary pattern variance descriptors that are extracted from ROIs. After that, they constructed deep invariant features in supervised and unsupervised fashions through a multilayer CNN architecture. Valvano et al. [151] achieved accuracy of 83.7% for MCs detection using a deep CNN. Jamieson et al. [43] introduced a four-layer unsupervised adaptive deconvolution network to learn the image representation using 739 FFDM images. Sun et al. [105] developed a graph-based semi-supervised learning (SSL) method using a deep CNN, their method allows the users to include the unlabeled data into the DL training data-set. In contrast, Arevalo et al. [69] used supervised training in their method using ROIs annotated manually made by expert radiologists, achieving AUC of 0.86. Arevalo et al. [71] following their work in [69], used a hybrid supervised CNN classifier along with an extensive enhancement pre-processing process. Dubrovina et al. [104] presented a supervised CNN for region classification into semantically coherent tissues. The authors overcame the difficulty involved in a medium-size database by training the CNN in an overlapping patch-wise manner. Teare et al. [62] proposed dual supervised CNNs for classifying full MG images to normal, benign and malignant classes. In their work a random forest classifier was trained, taking the outputs of the two-deep CNNs.
The authors in [42, 44, 57, 66, 69, 70, 72, 73, 90, 146, 147] applied pre-processing, augmentation, normalization, regularization, mixing FSM and FFDM MG images, and other techniques to better implement their network. Ge et al. [42] compared the performance of CNNs on pairs of FFDM and SFM obtained from the same patients with a time span of less than 3 months. Their results show that the CNN with FFDM images (AUC of 0.96) detect more MCs than the CNN with FSM images (AUC of 0.91). Hepsaug [74] achieved an accuracy of 88% when training separate deep CNN on only mass ROIs and 0.84% on training deep CNN on only MCs ROIs in the BCDR database. On the other hand, the accuracy results show that classifying only mass or only MCs is more successful compared to classifying mass and MCs data. Zhu et al. [20] conducted mass detection for whole MG images. Their deep multi-instance network uses linear regression with weight sharing for the malignant probability of each position from the CNN’s feature maps. The authors in [50, 146] trained a multi-stage CNN network for the classifications of lesions in MGs. Bekker et al. [56] presented a deep MV CNN for the classification of clustered breast MCs to two classes. Their results show that classification based on MV MGs show promising results. Carneiro et al. [32] addressed the classification of mass(es) using a pre-trained MV CNN. Their model classifies a full MG by extracting features from each view of the breast (train a separate CNN for each view) and combining these features in a joint CNN model to output a prediction that estimates the patient’s risk of developing breast cancer. Carneiro et al. [95] following his work [32] build a fully automated pre-trained CNN for detecting masses and MCs in MV MG images. Geras et al. [22] developed a MV CNN that utilizes large high-resolution images without downscaling. They showed that the accuracy of detecting and classifying MGs clearly increases with the size of the training data-set and that the best performance can only be achieved using the images in the original resolution. Yi et al. [23] utilized a deep MV learning by averaging the probability scores of both views to make the final prediction. Lotter et al. [65] introduced a multi-scale deep CNN trained with a curriculum learning strategy. Lotter et al. first train CNN-based patch classifiers on ROIs, and then use the learned features to initialize a scanning-based model that renders a decision on the whole image, having final results by averaging final scores across MV of the breast. Dhungel et al. [97, 98] presented an cascade DL networks for detecting, segmenting and classifying breast masses from MGs with minimal user intervention. Dhungel et al. [25], following their work in [52, 97, 98], implemented a MV deep residual neural network for the fully automated classification of MGs as either malignant or normal/benign (AUC of 0.8).
Risk assessment
The studies in [26, 27, 33, 34, 107, 115, 133] have demonstrated that applying CNNs methods have significant potential to develop a new short-term risk predicting scheme with improved performance in detecting early abnormal symptom from the negative MGs. Breast density is considered a strong indicator of breast cancer risk [26, 27, 33, 34, 152]. Fonseca et al. [26, 152] explored an automatic breast composition classification work-flow based on CNN for feature extraction in combination with a support vector machines classifier. Similar approach was done by Becker [33] achieving an (AUC of 0.82) comparable to experienced radiologists (AUC of 0.79–0.87).
Li [153] trained a deep CNN to estimate a probability map of breast density (PMD) to classify mammographic pixels into fatty class or dense class. Kallenberg et al. [27] presented an unsupervised CNN for breast density segmentation and automatic texture scoring. The model learns features across multiple scales, then they are fed to a simple classifier that is specific to the task of interest yielding AUC of 0.59. Ahn et al. [34] used CNN for the task of automatic classification of mammographic breast tissues into dense and fatty tissues. Their CNN is configured to learn the local features from image patches while keeping the context information of the whole MG. Wu et al. [107] managed to train a MV deep CNN using a data-set of 201,179 MGs for breast density classification. Mohamed et al. [115] achieved AUC of 0.95 when using only the MLO view images. In comparison, the AUC is 0.88 when using only the CC view images. When both the MLO and CC view images were combined as a single data-set, the AUC is lowered to 0.92. The authors in [110] following their work in [115] achieved better AUC of 0.98 by fine-tuning a pre-trained network. Hang [148] achieved classification accuracy of 66% for classification of full images into normal, benign and malignant.
Lesion localization
For localization, the information about which category an image belongs to is already available and the task is to instead figure out where exactly the object is located in the image. Classification and localization can also be combined so that a fixed amount of lesions in an image will be classified and also located. This task, called multi-class localization. The following authors employed CNNs in the aim of lesions classification and then localization within these images [14, 31, 45, 52–54, 57, 61, 98, 154, 155], potentially enabling E2E training. Ben-Ari et al. [24] introduced the detection of AD using a supervised pre-trained region-based network (R-CNN). Ertosun and Rubin [53] developed an E2E dual CNN based visual search system for localization of mass(es) in MGs. Kisilev et al. [61] gave a semantic description for MGs. The authors presented a multi-task R-CNN approach for detection and semantic description of lesions in diagnostic images. Carneiro and Bradley [54] presented an automated supervised architecture composes of a multi-scale deep belief network that selects suspicious regions to be further processed by a two-level cascaded R-CNN. Akselrod et al. [45] integrated several cascaded segmentation modules into a modified cascaded R-CNN. Hwang et al. [51] proposed a self-transfer learning framework which enables training CNNs for object localization without neither any location information nor pre-trained models. Zhu et al. [57] introduced an E2E adversarial training for mammographic mass segmentation to learn robustly from scarce MGs. The authors highlighted the importance of pre-processing, augmentation, image enhancement, and normalization techniques. The authors stated that it is not feasible to use networks pre-trained on general images since ROI characteristics of medical images are thoroughly different from natural images. However, their opinion contradicts other researchers work.
The authors in [31, 52, 155] proposed a patch-based CNN to detect masses. Choukroun et al. [155] proposed a method that classifies MGs by detecting discriminative local information contained in patches through a deep CNN and then uses the local information to localize tumors. Dhungel et al. [52] used the output from a CNN as a complimentary potential function to a deep belief network (DBN) models for the localization of breast masses from MGs, using a small training data-set. A drawback of the patch-based approach in [31, 52] is that the input patches came from non-overlapping areas, which makes it difficult to preciously localize masses. Moreover, the size of the input patches in [31, 52] is very small that produces a difficulty in differentiating normal tissues from abnormal ones.
The authors in [14, 154] used the famous YOLO-based deep CNN [83] for breast mass classification and localization. The trained YOLO-based system localizes the masses and classifies their types into benign or malignant. The authors in [154] achieved a mass location with an overall accuracy of 96.33% and detection of benign and malignant lesions with an overall accuracy of 85.52%.
Image retrieval
Tasks like medical image retrieval using DL have been lately addressed in the medical field to facilitate the process of production and management of large medical image databases. Conventional methods for analyzing medical images have achieved limited success, as they are not capable to tackle the huge databases. The learned features and the classification results from training a CNN are used to retrieve medical images. Qayyum et al. [114] proposed a DL based framework for content based medical image retrieval (CBMIR) by training a deep CNN for the classification tasks using medical images for different body organs (e.g. MGs, lungs, brain, liver etc. Qayyum et al. [114] achieved an average classification accuracy of 99.77% for 24 classes of medical images. Similarly, Ahmad et al. [156] trained a deep CNN for CBMIR of different 193 classes for different body organs. Moreover, [156] applied TL and augmentation to increase the performance of their deep CNN.
Super resolution image reconstruction
The task of super resolution image reconstruction using CNN (SRCNN) is an E2E mapping between the low and high-resolution images for enhancing images [157]. The mapping is represented as a deep CNN that takes the low resolution image as the input and outputs the high resolution one. The study of Umehara et al. [158] shows that SRCNN can significantly outperform conventional interpolation methods for enhancing image resolution in digital mammography especially in dense breasts.
Research challenges and directions
In this section, we list the research challenges and directions that require further investigations by the community.
Localization of tumors
The patch-based CNNs, R-CNNs, Fast R-CNNs, Faster R-CNNs, and YOLO methods have recently become more popular for localization tasks for MGs. Faster R-CNN is the choice of most of the mammography researchers who aim to obtain high detection accuracy numbers. However, training a R-CNN and its variants faster versions is time-consuming and memory expensive. In contrast, for faster computations, less accurate detection, and limited memory computations, the YOLO method is the right choice. Finally, patch-based CNN methods are not recommended and result in many false positives. More research need to be done for better localization of tumors in MGs.
Limited data for learning
One of the challenging problems that face researchers while training CNNs is the size of the training data-set. As discussed in the best practice section, although several approaches such as data augmentation, TL, and drop out have been used to handle the problem of training the model with limited samples, this problem has remained challenging.
Imbalanced data-set
Another challenging problem is the imbalance ratio between positive and negative classes in the training data-sets. Training CNN models directly on imbalanced data-sets may bias the prediction towards the more common classes like normal. The effect of imbalanced data-set on the performance of a CNN for MGs has not been studied thoroughly. Some works used balanced data-set and some used imbalanced ones. Since in general less abnormal MGs are available compare to normal MGs it is very important to investigate the effect of using balanced and imbalanced data-sets on the accuracy of the CNN model.
Size of lesions
The size variation of lesions within MG images is another challenge for training CNNs in detecting cancer. Resizing a large MG to 224 ×224 or 227 ×227 (common choices among researchers) will likely make the ROI hard to detect and/or classify. To address this problem, several studies have proposed to train a CNN model using different scales of lesions [27, 54, 65]. More research is required to find lesions of different sizes.
Memory constrains
The classification of whole size MG images is challenging due to the memory constraints and increased feature space. Researchers in [22, 128] address this problem by resizing the images to smaller ones, however, this affects the accuracy of their model. More research should be done on how to overcome the memory constraints while training CNNs with full-size MG images.
Non-annotated data-set
Another challenging problem to researchers is how to train a CNN model using a non-annotated data-set. In non-annotated data-set, the input image to CNN model is binary labeled as normal or cancerous without any details about the location of the abnormalities. To address this problem, Lotter at el. [65] train a patch-level CNN classifier, which is then used as feature extractor to an image-level model. Training the CNNs for classification of non-annotated data-set is still an open area for research [20, 65, 129].
False positives reduction
Even though CNNs are very successful in providing better performance compared to traditional CADs, they still result in false positives. False positive results cause patients needless anxiety, additional testing, biopsies, and unnecessary costs. Several approaches have been proposed to improve false positive in CNNs such as using MV CNNs [15, 20–25, 95, 107, 108]. However, more research is required to integrate prior images with current screening to eliminate false positives.
Multiple detection
Current CNN models are trained to detect and/or localize mass(es) within MGs neglecting the existence of MCs. More research should be directed on detecting multiple abnormalities within the same breast.
Pre-processing filters
In FSM images, a significant number of abnormalities are misdiagnosed or missed due to the less visibility, low contrast, poor quality, and noisy nature of these images. Common pre-processing techniques (e.g. CLAHE, median filter) are proposed in [62, 89–91] to enhance image quality, image smoothing and noise reduction. However, choosing the proper pre-processing technique for MGs in order to improve the classification of CNNs is still an open problem.
Discussion and recommendations
We show a breakdown of the studies included in this survey grouped by their neural network task (see Additional file 1: Table S1). Figure 3, shows the percentage of studies employing some of the CNN best practices that are discussed in the previous section and are shown in Additional file 1: Table S1. 78 studies (out of 83) used common pre-processing techniques to enhance the quality of images, reduce or remove noise, and improve the contrast of MGs. That shows the importance of having a good separation between foreground and background pixels and not removing the important information from the images. Moreover, 59 studies used ROIs for more efficient computation, while 23 studies applied CNN to MG of full image size as in [20, 22, 26, 32, 45, 51, 57, 62]. Even for CNNs that are trained with full image size, the pre-processing is mandatory to remove marks, labels, pectoral muscle and black areas that can interfere in the post-processing of these images. Data augmentation has been recommended and employed by 52 studies. Data augmentation reduces overfitting by generating more instances of training data. TL is gaining more popularity for medical images, 32 studies have successfully applied it to pre-train their network. From 2015 until now, there is an increasing trend in using TL. 15 studies implemented a MV CNNs which lead to significant improvements in the performance of the single-view ones. It is a beneficial practice to use both CC and MLO views to detect abnormalities. 25 studies implemented an E2E CNN which may include segmentation, detection, and classification of lesions in MGs. We summarize the recommendations to significantly improve the performance of CNNs in detection and classification of breast cancer using MG images as follows:
Use pre-processing techniques such as CLAHE filter to improve the contrast of MGs, median filter to reduce noise, and un-sharp masking to smooth the images.
Apply cropping and down sampling for more efficient computation.
Use a suitable validation approach according to the size of the data-set available.
Use augmentation, drop-out, and TL to reduce overfitting and increase the generalization of the model.
Use suitable batch size if using ROIs.
Use multi-view (MV) CNNs to embed more information for better performance.
Use full resolution images if it is computationally practical.
Mix between FFDM and FSM images.
Use suitable activation function such as ReLU, be careful with initializing the learning rates and possibly monitor the fraction of dead neurons in the network.
Use large well labeled data-set if available.
Go deeper in layers if large data-set is available.
Use context and patient information in multi-modal models.
Use recently available libraries for implementing CNNs such as Tensorflow or Keras.
Conclusions
In this survey, we conducted a detailed review of the strengths, limitations, and performance of the most recent CNNs applications in analyzing mammogram (MG) images. This survey systematically compares recent approaches of CNNs in MG images, and show how the advances in DL methods give promising results that can aid radiologists and serve as a second eye for them. The potential role of CNN methods is to handle millions of routine imaging exams, presenting the potential cancers to the radiologists who perform follow-up procedures. We discuss the currently publicly available MG databases. We also give a deep insight into the architectures of CNNs used for various tasks in mammography.
This survey represents a valuable resource for the mammography research community since it can be utilized as a basis in their current and future studies. The given comparison among common publicly available MG repositories guides the community to select the most appropriate database for their application(s). Moreover, this survey lists the best practices that improve the performance of CNNs including the pre-processing of images and the use of multi-view images. In addition, other listed techniques like transfer learning (TL), data augmentation, batch normalization, and dropout are appealing solutions to reduce overfitting and increase the generalization of the CNN model. Finally, we identified research challenges and directions that require further investigations for mammography.
Additional file
Acknowledgements
Not applicable.
Funding
This study was supported by the Sheida Nabavi’s startup fund at the University of Connecticut and Dina Abdelhafiz’s scholarship from Egypt. Publication costs are funded by the Sheida Nabavi’s startup fund.
Availability of data and materials
The DDSM dataset is available online at http://www.eng.usf.edu/cvprg/Mammography/Database.html. The INbreast dataset can be requested online at http://medicalresearch.inescporto.pt/breastresearch/index.php/Get_INbreast_Database. The breast cancer digital repository (BCDR) dataset can be requested online at https://bcdr.eu. The MIAS database is available online at http://peipa.essex.ac.uk/info/mias.html.
About this supplement
This article has been published as part of BMC Bioinformatics Volume 20 Supplement 11, 2019: Selected articles from the 7th IEEE International Conference on Computational Advances in Bio and Medical Sciences (ICCABS 2017): bioinformatics. The full contents of the supplement are available online at https://bmcbioinformatics.biomedcentral.com/articles/supplements/volume-20-supplement-11.
Abbreviations
- ACC
Accuracy
- AUC
Area under the receiver operating curve
- AD
Architectural distortion
- BACs
Breast arterial calcifications
- BCDR
Breast cancer digital repository
- BN
Batch normalization
- CAD
Computer-aided detection
- CBMIR
Content based medical image retrieval
- CC
Craniocaudal
- CRF
Conditional random field
- CLAHE
Contrast limited adaptive histogram equalization
- CNNs
Deep convolutional neural networks
- DBT
Digital breast tomosynthesis
- DDMS
Digital database for screening mammography
- DL
Deep learning
- E2E
End-to-end
- FC
Fully connected layer
- FFDM
Digital mammography
- FN
False negative
- FP
False positive
- FPR
False positive rate
- FSMs
Film screen mammograms
- HGD
Histogram of the gradient divergence
- HOG
Histogram of oriented gradients
- IRMA
Image retrieval in medical applications
- MCs
Calcifications
- MG
Mammogram
- MIAS
Mammographic Image Analysis Society
- ML
Machine learning
- MLO
Mediolateral-oblique
- MLP
Multilayer perceptron
- PMD
Probability map of breast density
- Pool
Pooling layer
- R-CNN
Region-based network
- ReLU
Rectified liner unit
- ROC
Receiver operating characteristic curve
- ROIs
Region of interests
- SRCNN
Super resolution image reconstruction
- SSL
Semi-supervised learning
- SVM
Support vector machine
- TL
Transfer learning
- TN
True negative
- TP
True positive
- TPR
True positive rate
Authors’ contributions
DH and SN designed the study. DH performed all the analyses in this paper and interpreted the results. DH and SN wrote the manuscript. All authors read and approved the final version of the manuscript.
Ethics approval and consent to participate
Not applicable.
Consent for publication
Not applicable.
Competing interests
The authors declare that they have no competing interests.
Publisher’s Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Contributor Information
Dina Abdelhafiz, Email: dina.abdelhafiz@uconn.edu.
Clifford Yang, Email: cyang@uchc.edu.
Reda Ammar, Email: reda.ammar@uconn.edu.
Sheida Nabavi, Email: sheida.nabavi@uconn.edu.
References
- 1.Siegel RL, Miller KD, Jemal A. Cancer statistics, 2016. CA Cancer J Clin. 2016;66(1):7–30. doi: 10.3322/caac.21332. [DOI] [PubMed] [Google Scholar]
- 2.Li Y, Chen H, Cao L, Ma J. A survey of computer-aided detection of breast cancer with mammography. J Health Med Inf. 2016;4(7).
- 3.Feig SA. Screening mammography benefit controversies: sorting the evidence. Radiol Clin N Am. 2014;3(52):455–80. doi: 10.1016/j.rcl.2014.02.009. [DOI] [PubMed] [Google Scholar]
- 4.Welch HG, Passow HJ. Quantifying the benefits and harms of screening mammography. JAMA Intern Med. 2014;3(174):448–54. doi: 10.1001/jamainternmed.2013.13635. [DOI] [PubMed] [Google Scholar]
- 5.Lehman CD, Wellman RD, Buist DS, Kerlikowske K, Tosteson AN, Miglioretti DL. Diagnostic accuracy of digital screening mammography with and without-aided detection. JAMA Intern Med. 2015;175(11):1828–37. doi: 10.1001/jamainternmed.2015.5231. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 6.Hayward JH, Ray KM, Wisner DJ, Kornak J, Lin W, Joe BN, et al. Improving screening mammography outcomes through comparison with multiple prior mammograms. Am J Roentgenol. 2016;207(4):918–24. doi: 10.2214/AJR.15.15917. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 7.LeCun Y, Bengio Y, Hinton G. Deep learning. Nature. 2015;521(7553):436–44. doi: 10.1038/nature14539. [DOI] [PubMed] [Google Scholar]
- 8.Litjens G, Kooi T, Bejnordi BE, Setio AAA, Ciompi F, Ghafoorian M, et al.A survey on deep learning in medical image analysis. 2017. arXiv preprint arXiv:170205747. [DOI] [PubMed]
- 9.Zhao R, Yan R, Chen Z, Mao K, Wang P, Gao RX. Deep Learning and Its Applications to Machine Health Monitoring: A Survey. 2016. arXiv preprint arXiv:161207640.
- 10.Lee JG, Jun S, Cho YW, Lee H, Kim GB, Seo JB, et al. Deep Learning in Medical Imaging: General Overview. Korean J Radiol. 2017;4(18):570–84. doi: 10.3348/kjr.2017.18.4.570. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 11.Hedjazi MA, Kourbane I, Genc Y. On identifying leaves: A comparison of CNN with classical ML methods. In: Signal Processing and Communications Applications Conference (SIU) 2017 25th. IEEE: 2017. p. 1–4.
- 12.Kooi T, Gubern-Merida A, Mordang JJ, Mann R, Pijnappel R, Schuur K, et al.A comparison between a deep convolutional neuralnetwork and radiologists for classifying regions of interest in mammography. In: International Workshop on Digital Mammography. Springer: 2016. p. 51–6.
- 13.Chartrand G, Cheng PM, Vorontsov E, Drozdzal M, Turcotte S, Pal CJ, et al.Deep Learning: A Primer for Radiologist RadioGraphics. 2017; 7(37):2113–31. [DOI] [PubMed]
- 14.Platania R, Shams S, Yang S, Zhang J, Lee K, Park SJ. Automated Breast Cancer Diagnosis Using Deep Learning and Region of Interest Detection (BC-DROID). In: Proceedings of the 8th ACM International Conference on Bioinformatics, Computational Biology and Health Informatics. ACM: 2017. p. 536–43.
- 15.Wang J, Ding H, Azamian F, Zhou B, Iribarren C, Molloi S, et al.Detecting cardiovascular disease from mammograms with deep learning. IEEE Trans Med Imaging. 2017. [DOI] [PMC free article] [PubMed]
- 16.Krizhevsky A, Sutskever I, Hinton GE. Imagenet classification with deep convolutional neural networks. In: Advances in neural information processing systems: 2012. p. 1097–105.
- 17.Greenspan H, van Ginneken B, Summers RM. Guest editorial deep learning in medical imaging: Overview and future promise of an exciting new technique. IEEE Trans Med Imaging. 2016;35(5):1153–9. doi: 10.1109/TMI.2016.2553401. [DOI] [Google Scholar]
- 18.Weiss K, Khoshgoftaar TM, Wang D. A survey of transfer learning. J Big Data. 2016;3(1):9. doi: 10.1186/s40537-016-0043-6. [DOI] [Google Scholar]
- 19.Christoyianni I, Constantinou E, Dermatas E. Automatic detection of abnormal tissue in bilateral mammograms using neural networks. Methods Appl Artif Intell. 2004;:267–75.
- 20.Zhu W, Lou Q, Vang YS, Xie X. Deep multi-instance networks with sparse label assignment for whole mammogram classification. In: International Conference on Medical Image Computing and Computer-Assisted Intervention. Springer: 2017. p. 603–11.
- 21.Kooi T, Ginneken B, Karssemeijer N, Heeten A. Discriminating solitary cysts from soft tissue lesions in mammography using a pretrained deep convolutional neural network. Medical physics. 2017;44(3):1017–27. doi: 10.1002/mp.12110. [DOI] [PubMed] [Google Scholar]
- 22.Geras KJ, Wolfson S, Kim S, Moy L, Cho K. High-Resolution Breast Cancer Screening with Multi-View Deep Convolutional Neural Networks. 2017. arXiv preprint arXiv:170307047.
- 23.Yi D, Sawyer RL, Cohn III D, Dunnmon J, Lam C, Xiao X, et al.Optimizing and Visualizing Deep Learning for Benign/Malignant Classification in Breast Tumors. 2017. arXiv preprint arXiv:170506362.
- 24.Ben-Ari R, Akselrod-Ballin A, Karlinsky L, Hashoul S. Domain specific convolutional neural nets for detection of architectural distortion in mammograms. In: Biomedical Imaging (ISBI 2017) 2017 IEEE 14th International Symposium on IEEE.2017. p. 552–6.
- 25.Dhungel N, Carneiro G, Bradley AP. Fully automated classification of mammograms using deep residual neural networks. In: Biomedical Imaging (ISBI 2017) 2017 IEEE 14th International Symposium on IEEE.2017. p. 310–4.
- 26.Fonseca P, Mendoza J, Wainer J, Ferrer J, Pinto J, Guerrero J, et al.Automatic breast density classification using a convolutional neural network architecture search procedure. In: Proc of SPIE Vol: 2015. p. 941428–1.
- 27.Kallenberg M, Petersen K, Nielsen M, Ng AY, Diao P, Igel C, et al. Unsupervised deep learning applied to breast density segmentation and mammographic risk scoring. IEEE Trans Med Imaging. 2016;35(5):1322–31. doi: 10.1109/TMI.2016.2532122. [DOI] [PubMed] [Google Scholar]
- 28.Oustimov A, Gastounioti A, Hsieh MK, Pantalone L, Conant EF, Kontos D. Convolutional neural network approach for enhanced capture of breast parenchymal complexity patterns associated with breast cancer risk. In: SPIE Medical Imaging. International Society for Optics and Photonics: 2017. p. 101340S–101340S. [DOI] [PMC free article] [PubMed]
- 29.Petersen K, Nielsen M, Diao P, Karssemeijer N, Lillholm M. Breast tissue segmentation and mammographic risk scoring using deep learning. In: International Workshop on Digital Mammography. Springer: 2014. p. 88–94.
- 30.Qiu Y, Wang Y, Yan S, Tan M, Cheng S, Liu H, et al.An initial investigation on developing a new method to predict short-term breast cancer risk based on deep learning technology. In: SPIE Medical Imaging. International Society for Optics and Photonics.2016. p. 978521.
- 31.Sun W, Tseng TLB, Zheng B, Qian W. A preliminary study on breast cancer risk analysis using deep neural network. In: International Workshop on Digital Mammography. Springer: 2016. p. 385–91.
- 32.Carneiro G, Nascimento J, Bradley AP. Unregvistered multiview mammogram analysis with pre-trained deep learning models. In: International Conference on Medical Image Computing and Computer-Assisted Intervention. Springer: 2015. p. 652–60.
- 33.Becker AS, Marcon M, Ghafoor S, Wurnig MC, Frauenfelder T, Boss A. Deep learning in mammography: diagnostic accuracy of a multipurpose image analysis software in the detection of breast cancer. Investig Radiol. 2017;52(7):434–40. doi: 10.1097/RLI.0000000000000358. [DOI] [PubMed] [Google Scholar]
- 34.Ahn CK, Heo C, Jin H, Kim JH. A Novel Deep Learning-based Approach to High Accuracy Breast Density Estimation in Digital Mammography. In: SPIE Medical Imaging. International Society for Optics and Photonics.2017. p. 101342O–101342O.
- 35.Li H, Giger ML, Huynh BQ, Antropova NO. Deep learning in breast cancer risk assessment: evaluation of convolutional neural networks on a clinical dataset of full-field digital mammograms. J Med Imaging. 2017;4(4):041304. doi: 10.1117/1.JMI.4.4.041304. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 36.Suckling J, Parker J, Dance D, Astley S, Hutt I, Boggis C, et al.The mammographic image analysis society digital mammogram database. In: Exerpta Medica. International Congress Series.1994. p. 375–8.
- 37.Moreira IC, Amaral I, Domingues I, Cardoso A, Cardoso MJ, Cardoso JS. Inbreast: toward a full-field digital mammographic database. Acad Radiol. 2012;19(2):236–48. doi: 10.1016/j.acra.2011.09.014. [DOI] [PubMed] [Google Scholar]
- 38.Lopez MG, Posada N, Moura DC, Pollán RR, Valiente JMF, Ortega CS, et al.BCDR: a breast cancer digital repository. In: 15th International Conference on Experimental Mechanics: 2012.
- 39.Oliveiraa JE, Guelda MO, Araújob AdA, Ottc B, Desernoa TM. Towards a standard reference database for computer-aided mammography. In: Proc SPIE vol.2008. p. 69151Y.
- 40.Chan HP, Lo SCB, Sahiner B, Lam KL, Helvie MA. Computer-aided detection of mammographic microcalcifications: Pattern recognition with an artificial neural network. Med Phys. 1995;22(10):1555–67. doi: 10.1118/1.597428. [DOI] [PubMed] [Google Scholar]
- 41.Sahiner B, Chan HP, Petrick N, Wei D, Helvie MA, Adler DD, et al. Classification of mass and normal breast tissue: a convolution neural network classifier with spatial domain and texture images. IEEE Trans Med Imaging. 1996;5(15):598–610. doi: 10.1109/42.538937. [DOI] [PubMed] [Google Scholar]
- 42.Ge J, Hadjiiski LM, Sahiner B, Wei J, Helvie MA, Zhou C, et al. Computer-aided detection system for clustered microcalcifications: comparison of performance on full-field digital mammograms and digitized screen-film mammograms. Phys Med Biol. 2007;4(52):981. doi: 10.1088/0031-9155/52/4/008. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 43.Jamieson AR, Drukker K, Giger ML. Breast image feature learning with adaptive deconvolutional networks SPIE Medical Imaging. Strony. 2012;2012:831506–831506. [Google Scholar]
- 44.Mordang JJ, Janssen T, Bria A, Kooi T, Gubern-Mérida A, Karssemeijer N. Automatic microcalcification detection in multi-vendor mammography using convolutional neural networks. In: International Workshop on Digital Mammography. Springer: 2016. p. 35–42.
- 45.Akselrod-Ballin A, Karlinsky L, Alpert S, Hasoul S, Ben-Ari R, Barkan E. A region based convolutional network for tumor detection and classification in breast mammography. In: International Workshop on Large-Scale Annotation of Biomedical Data and Expert Label Synthesis. Springer: 2016. p. 197–205.
- 46.Huynh BQ, Li H, Giger ML. Digital mammographic tumor classification using transfer learning from deep convolutional neural networks. J Med Imaging. 2016;3(3):034501–034501. doi: 10.1117/1.JMI.3.3.034501. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 47.CBIS-DDSM. https://mcl.nci.nih.gov/science-data/cbis-ddsm-1. Accessed 3 Feb 2019.
- 48.Heath M, Bowyer K, Kopans D, Moore R, Kegelmeyer WP. The digital database for screening mammography. In: Proceedings of the 5th international workshop on digital mammography. Medical Physics Publishing: 2000. p. 212–8.
- 49.Abbas Q. DeepCAD: A Computer-Aided Diagnosis System for Mammographic Masses Using Deep Invariant Features. Computers. 2016;4(5):28. doi: 10.3390/computers5040028. [DOI] [Google Scholar]
- 50.Gallego-Posada J, Montoya-Zapata D, Quintero-Montoya O. Detection and Diagnosis of Breast Tumors using Deep Convolutional Neural Networks.
- 51.Hwang S, Kim HE. Self-transfer learning for fully weakly supervised object localization. 2016. arXiv preprint arXiv:160201625.
- 52.Dhungel N, Carneiro G, Bradley AP. Deep learning and structured prediction for the segmentation of mass in mammograms. In: International Conference on Medical Image Computing and Computer-Assisted Intervention. Springer: 2015. p. 605–612.
- 53.Ertosun MG, Rubin DL. Probabilistic visual search for masses within mammography images using deep learning. In: Bioinformatics and Biomedicine (BIBM), 2015 IEEE International Conference on. IEEE: 2015. p. 1310–5.
- 54.Carneiro NDG, Bradley AP. Automated Mass Detection from Mammograms using Deep Learning and Random Forest. 2016.
- 55.Jiao Z, Gao X, Wang Y, Li J. A deep feature based framework for breast masses classification. Neurocomputing. 2016;197:221–31. doi: 10.1016/j.neucom.2016.02.060. [DOI] [Google Scholar]
- 56.Bekker AJ, Greenspan H, Goldberger J. A multi-view deep learning architecture for classification of breast microcalcifications. In: Biomedical Imaging (ISBI) 2016 IEEE 13th International Symposium on. IEEE: 2016. p. 726–30.
- 57.Zhu W, Xie X. Adversarial deep structural networks for mammographic mass segmentation. 2016. arXiv preprint arXiv:161205970.
- 58.Lévy D, Jain A. Breast mass classification from mammograms using deep convolutional neural networks. 2016. arXiv preprint. arXiv:161200542.
- 59.Sharma K, Preet B. Classification of mammogram images by using CNN classifier; 2016. pp. 2743–9.
- 60.Suzuki S, Zhang X, Homma N, Ichiji K, Sugita N, Kawasumi Y, et al.Mass detection using deep convolutional neural network for mammographic computer-aided diagnosis. In: Society of Instrument and Control Engineers of Japan (SICE) 2016 55th Annual Conference of the. IEEE: 2016. p. 1382–6.
- 61.Kisilev P, Sason E, Barkan E, Hashoul S. Medical image description using multi-task-loss CNN. In: International Workshop on Large-Scale Annotation of Biomedical Data and Expert Label Synthesis. Springer: 2016. p. 121–9.
- 62.Teare P, Fishman M, Benzaquen O, Toledano E, Elnekave E. Malignancy Detection on Mammography Using Dual Deep Convolutional Neural Networks and Genetically Discovered False Color Input Enhancement. J Digit Imaging. 2017;4(30):499–505. doi: 10.1007/s10278-017-9993-2. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 63.Sert E, Ertekin S, Halici U. Ensemble of convolutional neural networks for classification of breast microcalcification from mammograms. In: Engineering in Medicine and Biology Society (EMBC) 2017 39th Annual International Conference of the IEEE. IEEE: 2017. p. 689–92. [DOI] [PubMed]
- 64.Jiao Z, Gao X, Wang Y, Li J. A parasitic metric learning net for breast mass classification based on mammography. Pattern Recogn. 2017.
- 65.Lotter W, Sorensen G, Cox D. A Multi-scale CNN and Curriculum Learning Strategy for Mammogram Classification. In: Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support. Springer: 2017. p. 169–177.
- 66.Jadoon MM, Zhang Q, Haq IU, Butt S, Jadoon A. Three-Class Mammogram Classification Based on Descriptive CNN Features. BioMed Res Int. 2017;:2017. [DOI] [PMC free article] [PubMed]
- 67.Domingues I, Cardoso J. Mass detection on mammogram images: a first assessment of deep learning techniques. In: 19th Portuguese Conference on Pattern Recognition (RECPAD).2013.
- 68.Domingues I, Sales E, Cardoso J, Pereira W. Inbreast-database masses characterization. XXIII CBEB. 2012.
- 69.Arevalo J, González FA, Ramos-Pollán R, Oliveira JL, Lopez MAG. Convolutional neural networks for mammography mass lesion classification. In: Engineering in Medicine and Biology Society (EMBC) 2015 37th Annual International Conference of the IEEE. IEEE: 2015. p. 797–800. [DOI] [PubMed]
- 70.Wichakam I, Vateekul P. Combining deep convolutional networks and SVMs for mass detection on digital mammograms. In: Knowledge and Smart Technology (KST) 2016 8th International Conference on. IEEE: 2016. p. 239–44.
- 71.Arevalo J, González FA, Ramos-Pollán R, Oliveira JL, Lopez MAG. Representation learning for mammography mass lesion classification with convolutional neural networks. Comput Methods Prog Biomed. 2016;127:248–57. doi: 10.1016/j.cmpb.2015.12.014. [DOI] [PubMed] [Google Scholar]
- 72.Jiang F, Liu H, Yu S, Xie Y. Breast mass lesion classification in mammograms by transfer learning. In: Proceedings of the 5th International Conference on Bioinformatics and Computational Biology. ACM: 2017. p. 59–62.
- 73.Chougrad H, Zouaki H, Alheyane O. Convolutional Neural Networks for Breast Cancer Screening: Transfer Learning with Exponential Decay. 2017. arXiv preprint arXiv:171110752.
- 74.Hepsağ PU, Özel SA, Yazıcı A. Using deep learning for mammography classification. In: Computer Science and Engineering (UBMK)2017. International Conference on IEEE.2017. p. 418—23.
- 75.Schmidhuber J. Deep learning in neural networks: An overview. Neural Netw. 2015;61:85–117. doi: 10.1016/j.neunet.2014.09.003. [DOI] [PubMed] [Google Scholar]
- 76.Wei D, Sahiner B, Chan HP, Petrick N. Detection of masses on mammograms using a convolution neural network. In: Acoustics Speech and Signal Processing 1995. ICASSP-95. 1995 International Conference on. vol. 5. IEEE: 1995. p. 3483–6.
- 77.Zeiler Matthew D., Fergus Rob. Computer Vision – ECCV 2014. Cham: Springer International Publishing; 2014. Visualizing and Understanding Convolutional Networks; pp. 818–833. [Google Scholar]
- 78.Szegedy C, Liu W, Jia Y, Sermanet P, Reed S, Anguelov D, et al.Going deeper with convolutions. In: Proceedings of the IEEE conference on computer vision and pattern recognition.2015. p. 1–9.
- 79.Simonyan K, Zisserman A. Very deep convolutional networks for large-scale image recognition. 2014. arXiv preprint. arXiv:14091556.
- 80.He K, Zhang X, Ren S, Sun J. Deep residual learning for image recognition. In: Proceedings of the IEEE conference on computer vision and pattern recognition.2016. p. 770–8.
- 81.Srinivas S, Sarvadevabhatla RK, Mopuri KR, Prabhu N, Kruthiventi SS, Babu RV. An Introduction to Deep Convolutional Neural Nets for Computer Vision. In: Deep Learning for Medical Image Analysis. Elsevier: 2017. p. 25–52.
- 82.Russakovsky O, Deng J, Su H, Krause J, Satheesh S, Ma S, et al. Imagenet large scale visual recognition challenge. Int J Comput Vis. 2015;3(115):211–52. doi: 10.1007/s11263-015-0816-y. [DOI] [Google Scholar]
- 83.Redmon J, Divvala S, Girshick R, Farhadi A. You only look once: Unified real-time object detection. In: Proceedings of the IEEE conference on computer vision and pattern recognition.2016. p. 779–88.
- 84.Al-masni MA, Al-antari MA, Park JM, Gi G, Kim TY, Rivera P, et al. Simultaneous detection and classification of breast masses in digital mammograms via a deep learning YOLO-based CAD system. Comput Methods Prog Biomed. 2018;157:85–94. doi: 10.1016/j.cmpb.2018.01.017. [DOI] [PubMed] [Google Scholar]
- 85.Mishkin D, Sergievskiy N, Matas J. Systematic evaluation of CNN advances on the ImageNet. 2016. arXiv preprint. arXiv:160602228.
- 86.Jifara W, Jiang F, Rho S, Cheng M, Liu S. Medical image denoising using convolutional neural network: a residual learning approach. J Supercomput. 2017:1–15.
- 87.Sharma J, Rai J, Tewari R. Identification of pre-processing technique for enhancement of mammogram images. In: Medical Imaging m-Health and Emerging Communication Systems (MedCom) 2014 International Conference on. IEEE: 2014. p. 115–9.
- 88.Bandyopadhyay SK. Pre-processing of Mammogram Images. Int J Eng Sci Technol. 2010;11(2):6753–8. [Google Scholar]
- 89.Kaur P, Kaur A. Review of Different Approaches in Mammography. 2016.
- 90.Bria A, Marrocco C, Galdran A, Campilho A, Marchesi A, Mordang JJ, et al.Spatial Enhancement by Dehazing for Detection of Microcalcifications with Convolutional Nets. In: International Conference on Image Analysis and Processing. Springer: 2017. p. 288–98.
- 91.Abdelhafiz D, Nabavi S, Ammar R, Yang C. The Effect of Pre-Processing on Breast Cancer Detection Using Convolutional Neural Networks. In: Poster session presented at the meeting of the IEEE International Symposium on Biomedical Imaging. Washington DC: 2018.
- 92.Abdelhafiz D, Nabavi S, Ammar R, Yang C. Survey on deep convolutional neural networks in mammography. In: Computational Advances in Bio and Medical Sciences (ICCABS), 2017 IEEE 7th International Conference on. IEEE. p. 1.
- 93.Ge J, Sahiner B, Hadjiiski LM, Chan HP, Wei J, Helvie MA, et al. Computer aided detection of clusters of microcalcifications on full field digital mammograms. Med Phys. 2006;33(8):2975–88. doi: 10.1118/1.2211710. [DOI] [PubMed] [Google Scholar]
- 94.Samala RK. Chan HP. Hadjiiski LM, Helvie MA, Cha KH. Richter CD. Multi-task transfer learning deep convolutional neural network: application to computer-aided diagnosis of breast cancer on mammograms. Phys Med Biol. 2017;23(62):8894. doi: 10.1088/1361-6560/aa93d4. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 95.Carneiro G, Nascimento J, Bradley AP. Automated Analysis of Unregistered Multi-View Mammograms With Deep Learning. IEEE Trans Med Imaging. 2017;11(36):2355–65. doi: 10.1109/TMI.2017.2751523. [DOI] [PubMed] [Google Scholar]
- 96.Kooi T, Litjens G, van Ginneken B, Gubern-Mérida A, Sánchez CI, Mann R, et al. Large scale deep learning for computer aided detection of mammographic lesions. Med Image Anal. 2017;35:303–12. doi: 10.1016/j.media.2016.07.007. [DOI] [PubMed] [Google Scholar]
- 97.Dhungel N, Carneiro G, Bradley AP. The automated learning of deep features for breast mass classification from mammograms. In: International Conference on Medical Image Computing and Computer-Assisted Intervention. Springer: 2016. p. 106–14.
- 98.Dhungel N, Carneiro G, Bradley AP. A deep learning approach for the analysis of masses in mammograms with minimal user intervention. Med Image Anal. 2017;37:114–28. doi: 10.1016/j.media.2017.01.009. [DOI] [PubMed] [Google Scholar]
- 99.Qiu Y, Yan S, Tan M, Cheng S, Liu H, Zheng B. Computer-aided classification of mammographic masses using the deep learning technology: a preliminary study. In: SPIE Medical Imaging. International Society for Optics and Photonics: 2016. p. 978520–978520.
- 100.Antropova N, Huynh BQ, Giger ML. A Deep Feature Fusion Methodology for Breast Cancer Diagnosis Demonstrated on Three Imaging Modality Datasets. Med Phys. 2017. [DOI] [PMC free article] [PubMed]
- 101.Wang D, Khosla A, Gargeya R, Irshad H, Beck AH. Deep learning for identifying metastatic breast cancer. 2016. arXiv preprint. arXiv:160605718.
- 102.Lo SCB, Li H, Wang Y, Kinnard L, Freedman MT. A multiple circular path convolution neural network system for detection of mammographic masses. IEEE Trans Med Imaging. 2002;21(2):150–8. doi: 10.1109/42.993133. [DOI] [PubMed] [Google Scholar]
- 103.Agrawal P, Vatsa M, Singh R. Saliency based mass detection from screening mammograms. Signal Process. 2014;99:29–47. doi: 10.1016/j.sigpro.2013.12.010. [DOI] [Google Scholar]
- 104.Dubrovina A, Kisilev P, Ginsburg B, Hashoul S, Kimmel R. Computational mammography using deep neural networks. Comput Methods Biomech Biomed Eng Imaging Vis. 2016:1–5.
- 105.Sun W, Tseng TLB, Zhang J, Qian W. Enhancing deep convolutional neural network scheme for breast cancer diagnosis with unlabeled data. Comput Med Imaging Graph. 2017;57:4–9. doi: 10.1016/j.compmedimag.2016.07.004. [DOI] [PubMed] [Google Scholar]
- 106.Samala RK, Chan HP, Hadjiiski L, Helvie MA, Wei J, Cha K. Mass detection in digital breast tomosynthesis: Deep convolutional neural network with transfer learning from mammography. Med Phys. 2016;43(12):6654–66. doi: 10.1118/1.4967345. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 107.Wu N, Geras KJ, Shen Y, Su J, Kim S, Kim E, et al.Breast density classification with deep convolutional neural networks. 2017. arXiv preprint arXiv:171103674.
- 108.Kooi T. Karssemeijer N. Classifying symmetrical differences and temporal change for the detection of malignant masses in mammography using deep neural networks. J Med Imaging. 2017;4(4):044501. doi: 10.1117/1.JMI.4.4.044501. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 109.Hadad O, Bakalo R, Ben-Ari R, Hashoul S, Amit G. Classification of breast lesions using cross-modal deep learning, Vol. 2017. IEEE. pp. 109–12.
- 110.Mohamed AA, Berg WA, Peng H, Luo Y, Jankowitz RC, Wu S. A deep learning method for classifying mammographic breast density categories. Med Phys. 2017. [DOI] [PMC free article] [PubMed]
- 111.Glorot X, Bordes A, Bengio Y. Deep sparse rectifier neural networks. In: Proceedings of the Fourteenth International Conference on Artificial Intelligence and Statistics.2011. p. 315–23.
- 112.Pedamonti D. Comparison of non-linear activation functions for deep neural networks on MNIST classification task. 2018. arXiv preprint arXiv:180402763.
- 113.Maas AL, Hannun AY, Ng AY. Rectifier nonlinearities improve neural network acoustic models. In: Proc icml. vol. 30.2013. p. 3.
- 114.Qayyum A, Anwar SM, Awais M, Majid M. Medical image retrieval using deep convolutional neural network. Neurocomputing. 2017.
- 115.Mohamed AA, Luo Y, Peng H, Jankowitz RC, Wu S. Understanding Clinical Mammographic Breast Density Assessment: a Deep Learning Perspective. J Digit Imaging. 2017:1–6. [DOI] [PMC free article] [PubMed]
- 116.Srivastava N, Hinton G, Krizhevsky A, Sutskever I, Salakhutdinov R. Dropout: a simple way to prevent neural networks from overfitting. J Mach Learn Res. 2014;15(1):1929–58. [Google Scholar]
- 117.Smirnov EA, Timoshenko DM, Andrianov SN. Comparison of regularization methods for imagenet classification with deep convolutional neural networks. Aasri Procedia. 2014;6:89—94. doi: 10.1016/j.aasri.2014.05.013. [DOI] [Google Scholar]
- 118.Ioffe S, Szegedy C. Batch normalization: Accelerating deep network training by reducing internal covariate shift. In: International Conference on Machine Learning.2015. p. 448–56.
- 119.Carneiro G, Nascimento J, Bradley AP. CHAPTER OUTLINE. Deep Learn Med Image Anal. 2017;:321.
- 120.Suzuki S, Zhang X, Homma N, Ichiji K, Kawasumi Y, Ishibashi T, et al. WE-DE-207B-02: Detection of Masses On Mammograms Using Deep Convolutional Neural Network. A Feasibility Study. Med Phys. 2016;43(6):3817–7. doi: 10.1118/1.4957862. [DOI] [Google Scholar]
- 121.Pang S, Yu Z, Orgun MA. A novel end-to-end classifier using domain transferred deep convolutional neural networks for biomedical images. Comput Methods Prog Biomed. 2017;140:283–93. doi: 10.1016/j.cmpb.2016.12.019. [DOI] [PubMed] [Google Scholar]
- 122.Tajbakhsh N, Shin JY, Gurudu SR, Hurst RT, Kendall CB, Gotway MB, et al. Convolutional neural networks for medical image analysis. Full training or fine tuning? IEEE Trans Med Imaging. 2016;35(5):1299–312. doi: 10.1109/TMI.2016.2535302. [DOI] [PubMed] [Google Scholar]
- 123.Oquab M, Bottou L, Laptev I, Sivic J. Learning and transferring mid-level image representations using convolutional neural networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition.2014. p. 1717–24.
- 124.Yosinski J, Clune J, Bengio Y, Lipson H. How transferable are features in deep neural networks? In: Advances in neural information processing systems.2014. p. 3320–8.
- 125.Tajbakhsh N, Shin JY, Gurudu SR, Hurst RT, Kendall CB, Gotway MB, et al.On the necessity of fine-tuned convolutional neural networks for medical imaging. In: Deep Learning and Convolutional Neural Networks for Medical Image Computing. Springer: 2017. p. 181–93.
- 126.Wei X, Chen J, Cai C. Using Deep Convolutional Neural Networks and Transfer Learning for Mammography Mass Lesion Classification. Journal of Computational and Theoretical Nanoscience. 2017;14(8):3802–06. doi: 10.1166/jctn.2017.6676. [DOI] [Google Scholar]
- 127.Xi P, Shu C, Goubran R. Abnormality Detection in Mammography using Deep Convolutional Neural Networks. 2018. arXiv preprint. arXiv:180301906.
- 128.Zhang X, Zhang Y, Han EY, Jacobs N, Han Q, Wang X, et al.Whole mammogram image classification with convolutional neural networks. In: Bioinformatics and Biomedicine (BIBM) 2017 IEEE International Conference on. IEEE: 2017. p. 700–4.
- 129.Shen L. End-to-end Training for Whole Image Breast Cancer Diagnosis using An All Convolutional Design. 2017. arXiv preprint. arXiv:170809427.
- 130.Hohman F, Kahng M, Pienta R, Chau DH. Visual Analytics in Deep Learning: An Interrogative Survey for the Next Frontiers. 2018. arXiv preprint arXiv:180106889. [DOI] [PMC free article] [PubMed]
- 131.Lee RS, Gimenez F, Hoogi A, Miyake KK, Gorovoy M, Rubin DL. A curated mammography data set for use in computer-aided detection and diagnosis research. Sci Data. 2017;4:170177. doi: 10.1038/sdata.2017.177. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 132.Qiu Y, Yan S, Gundreddy RR, Wang Y, Cheng S, Liu H, et al.A New Approach to Develop Computer-Aided Diagnosis Scheme of Breast Mass Classification Using Deep Learning Technology. J X-Ray Sci Technol (Preprint). 2017:1–13. [DOI] [PMC free article] [PubMed]
- 133.Thomaz RL, Carneiro PC, Patrocinio AC. Feature extraction using convolutional neural network for classifying breast density in mammographic images. In: Medical Imaging 2017: Computer-Aided Diagnosis. vol. 10134. International Society for Optics and Photonics.2017. p. 101342M.
- 134.Yadav S, Shukla S. Analysis of k-Fold Cross-Validation over Hold-Out Validation on Colossal Datasets for Quality Classification. In: Advanced Computing (IACC) 2016 IEEE 6th International Conference on. IEEE: 2016. p. 78–83.
- 135.Masko D. Hensman, P; 2015.
- 136.Chawla NV. Data mining for imbalanced datasets: An overview. In: Data mining and knowledge discovery handbook. Springer: 2009. p. 875–86.
- 137.Raman V, Sumari P. Then H Al-Omari SAK. Review on Mammogram Mass Detection by MachineLearning Techniques. Int J Comput Electr Eng. 2011;6(3):873. doi: 10.7763/IJCEE.2011.V3.436. [DOI] [Google Scholar]
- 138.El Atlas N, El Aroussi M, Wahbi M. Computer-aided breast cancer detection using mammograms: A review. In: Complex Systems (WCCS), 2014 Second World Conference on. IEEE: 2014. p. 626–31.
- 139.Erickson BJ, Korfiatis P, Akkus Z, Kline T, Philbrick K. Toolkits and Libraries for Deep Learning. J Digit Imaging. 2017;:1–6. [DOI] [PMC free article] [PubMed]
- 140.Sherkhane P, Vora D. Survey of deep learning software tools. In: Data Management Analytics and Innovation (ICDMAI) 2017 International Conference on. IEEE: 2017. p. 236–8.
- 141.Abadi M, Barham P, Chen J, Chen Z, Davis A, Dean J, et al.TensorFlow: A System for Large-Scale Machine Learning. In: OSDI. vol. 16.2016. p. 265–83.
- 142.Jia Y, Shelhamer E. Caffe model zoo. 2015.
- 143.Jia Y, Shelhamer E, Donahue J, Karayev S, Long J, Girshick R, et al.Caffe: Convolutional Architecture for Fast Feature Embedding. 2014. arXiv preprint arXiv:14085093.
- 144.Paszke A, Gross S, Chintala S, Chanan G, Yang E, DeVito Z, et al.Automatic differentiation in pytorch. 2017.
- 145.Vedaldi A, Lenc K. Matconvnet: Convolutional neural networks for matlab. In: Proceedings of the 23rd ACM international conference on Multimedia. ACM: 2015. p. 689–92.
- 146.Agarwal V, Carson C. Using Deep Convolutional Neural Networks to predict semantic features of lesions in mammograms. C231n Course Project Reports. 2015.
- 147.Jaffar MA. Deep Learning based Computer Aided Diagnosis System for Breast Mammograms. Int J Adv Comput Sci Appl. 2017;7(8):286–90. [Google Scholar]
- 148.Hang W, Liu Z, Hannun A. GlimpseNet: Attentional Methods for Full-Image Mammogram Diagnosis.
- 149.Kooi T, Mordang JJ, Karssemeijer N. Conditional Random Field Modelling of Interactions Between Findings in Mammography. In: SPIE Medical Imaging. International Society for Optics and Photonics.2017. p. 101341E–101341E.
- 150.Bakkouri I, Afdel K. Breast tumor classification based on deep convolutional neural networks. In: Advanced Technologies for Signal and Image Processing (ATSIP). International Conference on IEEE 2017.2017. p. 1–6.
- 151.Valvano G, Della Latta D, Martini N, Santini G, Gori A, Iacconi C, et al.Evaluation of a Deep Convolutional Neural Network method for the segmentation of breast microcalcifications in Mammography Imaging. In: EMBEC & NBC 2017. Springer: 2017. p. 438–41.
- 152.Fonseca P, Castañeda B, Valenzuela R, Wainer J. Breast Density Classification with Convolutional Neural Networks. In: Iberoamerican Congress on Pattern Recognition. Springer: 2016. p. 101–108.
- 153.Li S, Wei J, Chan HP, Helvie MA, Roubidoux MA, Lu Y, et al. Computer-aided assessment of breast density: comparison of learning supervised deep feature-based statistical learning. Phys Med Biol. 2018;63(2):025005. doi: 10.1088/1361-6560/aa9f87. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 154.Al-Masni M, Al-Antari M, Park J, Gi G, Kim T, Rivera P, et al.Detection and classification of the breast abnormalities in digital mammograms via regional Convolutional Neural Network. In: Engineering in Medicine and Biology Society (EMBC) 2017 39th Annual International Conference of the IEEE. IEEE: 2017. p. 1230–3. [DOI] [PubMed]
- 155.Choukroun Y, Bakalo R, Ben-Ari R, Akselrod-Ballin A, Barkan E, Kisilev P. Mammogram Classification and Abnormality Detection from Nonlocal Labels using Deep Multiple Instance Neural Network. 2017.
- 156.Ahmad J, Sajjad M, Mehmood I, Baik SW. SiNC: Saliency-injected neural codes for representation and efficient retrieval of medical radiographs. PLoS ONE. 2017;8(12):181707. doi: 10.1371/journal.pone.0181707. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 157.Dong C, Loy CC, He K, Tang X. Image super-resolution using deep convolutional networks. IEEE Trans Pattern Anal Mach Intell. 2016;2(38):295–307. doi: 10.1109/TPAMI.2015.2439281. [DOI] [PubMed] [Google Scholar]
- 158.Umehara K, Ota J, Ishida T. Super-Resolution Imaging of Mammograms Based on the Super-Resolution Convolutional Neural Network. Open J Med Imaging. 2017;7(04):180. doi: 10.4236/ojmi.2017.74018. [DOI] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Supplementary Materials
Data Availability Statement
The DDSM dataset is available online at http://www.eng.usf.edu/cvprg/Mammography/Database.html. The INbreast dataset can be requested online at http://medicalresearch.inescporto.pt/breastresearch/index.php/Get_INbreast_Database. The breast cancer digital repository (BCDR) dataset can be requested online at https://bcdr.eu. The MIAS database is available online at http://peipa.essex.ac.uk/info/mias.html.