Skip to main content
PLOS ONE logoLink to PLOS ONE
. 2021 Aug 26;16(8):e0256500. doi: 10.1371/journal.pone.0256500

Multi- class classification of breast cancer abnormalities using Deep Convolutional Neural Network (CNN)

Maleika Heenaye-Mamode Khan 1,*,#, Nazmeen Boodoo-Jahangeer 1,#, Wasiimah Dullull 1,#, Shaista Nathire 1,#, Xiaohong Gao 2,#, G R Sinha 3,#, Kapil Kumar Nagwanshi 4,#
Editor: Gulistan Raja5
PMCID: PMC8389446  PMID: 34437623

Abstract

The real cause of breast cancer is very challenging to determine and therefore early detection of the disease is necessary for reducing the death rate due to risks of breast cancer. Early detection of cancer boosts increasing the survival chance up to 8%. Primarily, breast images emanating from mammograms, X-Rays or MRI are analyzed by radiologists to detect abnormalities. However, even experienced radiologists face problems in identifying features like micro-calcifications, lumps and masses, leading to high false positive and high false negative. Recent advancement in image processing and deep learning create some hopes in devising more enhanced applications that can be used for the early detection of breast cancer. In this work, we have developed a Deep Convolutional Neural Network (CNN) to segment and classify the various types of breast abnormalities, such as calcifications, masses, asymmetry and carcinomas, unlike existing research work, which mainly classified the cancer into benign and malignant, leading to improved disease management. Firstly, a transfer learning was carried out on our dataset using the pre-trained model ResNet50. Along similar lines, we have developed an enhanced deep learning model, in which learning rate is considered as one of the most important attributes while training the neural network. The learning rate is set adaptively in our proposed model based on changes in error curves during the learning process involved. The proposed deep learning model has achieved a performance of 88% in the classification of these four types of breast cancer abnormalities such as, masses, calcifications, carcinomas and asymmetry mammograms.

Introduction

Breast cancer is known to be the second leading cause of death among women in the world. Breast cancer occurs when there is an abnormal growth of a mass of tissues of malignant cells. These cells originate from the milk glands of the breast. The classification of these cells is determined by the rate at which these unusual cells progress and the effect that they have on other normal cells that can eventually affect the whole body [1]. According to statistics of the World Health Organization (WHO), breast cancer is the most common type of cancer occurring in women globally, accounting for approximately 2.1 million new breast cancer cases [2]. In 2018, it is estimated that 627,000 women died from breast cancer–that is approximately 15% of all cancer deaths among women. There are research studies and work in the literature on detection and classification of breast cancers using some conventional methods [36] that have not used any specific type of machine learning.

It is reported that breast cancer is detected if specific symptoms appear. However, it was found that many women who have breast cancer do not have any symptoms. Primary prevention of breast cancer is not yet a reality except by undergoing prophylactic mastectomy for those who have the BRCA1 or BRCA2 gene mutation, known to cause breast cancer [7]. The effect of treatments administered to patients can also have an effect on the prognosis of the disease and the possibility of relapse in cancer patients. Surgery, chemotherapy, radiotherapy and hormone therapy are among the classical medical treatments used on breast cancer patients. Thus, screening is recommended as mandatory step so that the detection of breast cancer be made at early stage. If the cancer is detected at early stage then the lives can be saved [4]. Early detection of breast cancer helps in the early diagnosis and treatment. This is vital because prognosis is very important for long-term survival [7, 8]. Delays in the treatment of cancer contributes to the propagation of the illness and causes delays in the treatment. According to a study conducted by [9], it was noticed that those who have started their treatment within 3 months from the appearance of breast cancer symptoms have higher chance of survival and can reduce the proliferation of malignant cells from the body.

Advances in image processing especially medical image processing are bringing hopes in devising appropriate automated breast cancer detection and classification applications. The algorithms being deployed in deep learning have the capabilities of using layers of the neural networks to recognize patterns; and the computer algorithms are becoming very useful in the medical field. Despite continuous research in automated breast cancer applications, the correct identification or classification of breast abnormalities is still a challenge. In addition, deep learning requires large training data, which is very difficult to obtain in the medical domain. Thus, there are scopes in the further exploration of automated breast cancer detection applications to improve the accuracy of cancer screening. We have used a deep learning model for classification of abnormalities in this work. The main contribution of this research is summarized as follows:

  1. We have applied a pre-trained model ResNet50 and have devised ways to overcome overfitting models

  2. We have developed an enhanced deep CNN model by varying the learning rate, which is an important parameter, that influence performance. We have also considered ways of using the learning rate in adaptive manner during training. Our model has been able to differentiate between various types of abnormalities that cause breast cancer.

Related work

Considered as a screening procedure, breast self-examination is done by the individual themselves. By palming the breasts in different angles and under different pressures it is possible to detect any difference or changes in the breasts. However, the breast examination is the least reliable way to detect cancer. Mammography has emerged as an alternative and is being widely used in the medical field. However, relying only on mammograms has a high risk of false positives which often lead to unnecessary biopsies and surgeries [1].

With the development of automated medical applications, researchers have been developing automated breast cancer detection systems. Many applications have been deployed using machine learning techniques. In [1], the use of machine learning has been reported to determine the types of treatment that need to be administered to cancer patients. It has been reported by several authors [1012] that there has not been any significant improvement in accuracy with the early computer aided applications for breast cancer detection that were developed. In [13], GoogleTensorFlow is used to implement machine learning algorithms, that were applied on the breast features for classification. The Wisconsin datasets were used, where 569 data points were considered, out of which 212 malignant cases and 357 benign cases were seen. Features that were used were mainly the details of the nuclei of the image which were directly computed from the digital image obtained from the fine needle aspirate (FNA) of the masses from breast images. The mainly used features are smoothness, concavity, texture, compactness, perimeter, fractal metrics and few more interrelated attributes. The support vector machine (SVM) and Gated Recurrent Unit (GRU) were utilized. The GRU is actually a variant of recurrent neural network (RNN). The networks were used for binary classification in which training and testing were performed. The dataset includes 30% of testing data and 70% of training dataset. In the training process, a number of metrics were used such as true positive rate (TPR), number of epochs, accuracy, data points, true positive rate (TPR), false negative rate (FNR) and false positive rate (FPR). The proposed machine learning method provided classification performance of 93.75%. In addition, K-fold cross validation was also used to confirm the validity of the results.

In [14], the authors have used the gray level co-occurrence matrix (GLCM) to extract the statistical texture features from mammograms namely: Entropy, Angular Second Moment, Contrast, Mean and Difference Moment from each density function. In [1], numerous ways were studied to detect breast cancer while emphasizing the use of machine learning. A hybrid model was proposed with a combination of SVM, Artificial Neural Network (ANN), K Nearest Neighbour (KNN), and Decision Tree (DT) algorithms. With varying combinations of these techniques, SVM was found as the most popular method that resulted an accuracy of 99.8% irrespective of the fact that it was used alone or as a hybrid. On the other hand, authors in [15] used thresholding techniques to segment the out boundaries of the breast images and applied the Discrete Wavelet Transform (DWT) to extract the features. The success of this method has been noted to be of a rate higher than 90%.

In another work reported in [16], histopathological images were used to detect breast cancer. Histopathological images show observations obtained from biopsy. These images contain both local and hidden patterns. In order to determine the hidden patterns, unsupervised techniques namely Convolutional Neural Network (CNN), a Long-Short-Term-Memory (LSTM), and a combination of the CNN and LSTM models. SVM was then used to classify the images. In [17], authors have used a feed-forward back propagation network having number of hidden layers as 3. In these hidden layers, number of neurons were 50, 10 and 50 respectively in first, second and third hidden layers. In training process, suitable features were extracted which were used for diagnosis of the images. In this research, individual stage of cancer was linked through the neural network. In [18], authors worked on a CAD to distinguish between the different regions of breast tissues. Using both the standard MIAS and DDSM database, they enhanced the peripheral region of the mammograms and extracted 422 features from the ROI. They used selection techniques such as t-test and Sequential Backward selection. The four classifiers: K Nearest- Neighbour (KNN), Support Vector Machine (SVM), Linear Discriminant Analysis (LDA) and Quadratic discriminant analysis (QDA) were applied and compared.

Despite the ongoing development of machine learning techniques, there are no significant improvements in the performance of these applications. In the meantime, deep learning, which learns representations from data which promotes the learning of successive layers of increasingly meaningful representations, have been successful in visual object detection and classification in many domains. Eventually there is a growing interest in the exploration of deep learning techniques for breast cancer detection applications. Recent studies on the application of deep learning on breast cancer have provided good preliminary output and showed scopes for further exploration [19, 20]. Despite the numerous research and application of neural network, none of the work developed so far could classify the breast abnormalities, clearly showing a research gap in this area.

Material and methods

This section highlights about the methodology used in implementing the deep learning for classification.

Architecture of the proposed solution

Fig 1 shows various components of the proposed automated breast cancer detection and its application. A customized dataset is firstly constructed and then the images are pre-processed to remove blurs and noises. A deep convolution neural network has been developed in addition to the application of an existing pre-trained deep learning model. The enhanced model has been tested and evaluated.

Fig 1. Architecture of the proposed model.

Fig 1

Dataset preparation

In this research work, several abnormalities that potentially determine breast cancer are considered. The main concern is that the available online datasets do not cater for all the abnormalities that can occur in the breast. Thus, a customized dataset using CBIS-DDSM (available at: https://wiki.cancerimagingarchive.net/display/Public/CBIS-DDSM) and UPMC (available at: https://cancerregistrynetwork.upmc.com/upmc-network-cancer-registry/standardized-data-set-2/) has been created to contain images having masses, calcifications, carcinomas and asymmetry mammograms. The DDSM is a publicly available dataset that has been created by South Florida University and the classification of the breast abnormality has been done by expert radiologists. The dataset was created for research-based applications and studies for breast cancer detection. The computer-aided diagnosis (CAD) operates over the dataset. The image dimension of the breast images was taken as 3000 X 4800 pixels having the resolution up to 42 microns and depth of 16 bits. Around 2620 mammograms were included in the DDSM dataset which were scanned film images. In each of the cases, 4 different vies of breast images were taken so that the classification is satisfactory. There were two Mediolateral Oblique (MLO) views in each of the breast images and other two were Cranio-Caudal views. The CBIS-DDSM dataset is a subset of the DDSM one and it consists of 6775 studies. This subset is however curated by a trained mammography expert. Thus, it is recognized as an updated and standardized version of DDSM. The format of the images in DICOM is a decompressed version of the images and represents abnormalities like calcifications, mass, asymmetry and carcinomas. The UPMC is a dataset which contains tomo-synthesis images, where asymmetry breast abnormalities and mass images were selected. Fig 2 shows one sample of each of the different types of abnormalities considered namely: asymmetry, calcification, carcinoma and mass.

Fig 2.

Fig 2

(a) Asymmetry (b) Calcifications (c) Carcinoma (d) Mass.

Data pre-processing

This step is used to enhance the quality of the image and increase the chance for a better abnormality classification. The intensity between objects is improved to better highlight the breast tissue and noise introduced at image capture are also eliminated. Several filtering techniques have been analyzed and applied on the images to select the most appropriate one in this context. One popular filter namely the Wiener filter has been applied which is normally used to curb the amount of noise in the image. The noise signal is attenuated and brought down to some minimum level, rather than complete elimination which is not possible. It eliminates the additive noise and using inverse filtering, we can recover blurred image. However, as discussed in [21], blurry sharp contents or edges of images such as edges, lines and other fine details are messed up after the application of this filter. Median filter is yet another technique that is popularly used as filter technique and works by replacing each pixel value with the centroid or median value of the surrounding neighborhood of the pixel under the rectangular region it is working on. Eventually, it can reduce the amount of intensity variation between one pixel and the other pixels while keeping the sharpness of image edges [22]. A detailed analysis was conducted in [23] to show the complexities of the applications of pre-processing techniques in different types of medical images and the need for this process. In this work, Contrast limited adaptive histogram equalization (CLAHE) has been chosen as it operates on small region of the breast image rather than the entire image and apply equalization on each of them. In addition, the breast images had less broken lines. The images are in Dicom format and has been converted into jpg. Convolution Neural Network (CNN), which is one model that automate the process of feature construction is widely being adopted. Though these models have great potential in analyzing images, large datasets are required for the training. Since there are limited data currently available in the medical field, data augmentation was used to expand the dataset. Eventually the breast images were augmented by a factor of 10 by applying data augmentation techniques. The created dataset has 10,200 images, after applying data augmentation namely flipping, rotation, cropping and translation on each type of abnormalities. The breast images were then segmented from the background to eliminate unwanted background information. Fig 3 shows a pre-processed image using CLAHE.

Fig 3.

Fig 3

(a) Original Image (b) Image after CLAHE.

Transfer learning using pre-trained ResNet50

Transfer learning is based on a single approach of using already learned features from one task and applying to another task without the need to learn from scratch. This is normally conducted on already built Convolutional Neural Network (CNN) models which have been trained on the most commonly used and well-known datasets [24]. The CNN model analyses an input image and assigns weights to various aspects of the image with the aim of differentiating one image from other images. It consists of several layers, as shown in Fig 4.

Fig 4. Layers in CNN.

Fig 4

CNN consists of multiple sets of convolutional, pooling and fully-connected layers. The convolutional layers use the images to extract its features [25], and then the pooling layers are used for selecting two samples and discard the next one [26]. As a result, less data is sent in the layers which reduces the complexity of the computations. The pooling layers have another impact which is it can also initiate translation invariance. The fully connected layers are actually the units of the top and final layers being connected together. We have applied ResNet50, which is an improved version of ResNet model. It consists of 50 layers of shortcuts connections with the residual blocks. The shortcuts make use of less computation and as well as provide rich combination features. ResNet50 consists of one convolutional layer, then a set of normalization layers, and lastly 16 residual modules found in between two pooling layers [27]. ResNet50, which uses 3x3 filters to perform spatial convolution, was introduced for the classification of abnormalities of the breast images. The weights were first set and the convolution base which was used as the feature extractor was freezed. The spatial size of the feature map is reduced by adopting Maxpooling method. Our dataset was split into 60% for the training set and 40% for the testing set.

Development of the enhanced CNN model

The CNN model is used for training the data very often in medical image diagnosis, analysis and its applications. In fact, the medical imaging in CAD system becomes successful because of CNN use [28, 29]. This substantiates the use of CNN for our application in breast cancer detection, of course with some improvement or enhancement. Generally, a CNN comprises of hundreds of neurons structured in different layers, forming the kernels. The size of kernels is chosen small as the depth of input image has to be same. A receptive field which is actually a small region, is connected with neurons. This is done because it is very difficult to connect all neurons to the previous outputs, especially when the input has high dimensional images. Let us understand by an example: For an image of size 100 x 100 having 10000 pixels, there are 100 neurons which will require or produce 1000000 parameters. Now, each neuron does not need to have weight of full input dimension, rather the neuron will hold the weight corresponding to the kernel dimension since the kernels of small dimension have been connected with neurons. The kernel needs to slide over the image both in height and width so that high level features are extracted and two-dimensional (2D) activation map is produced. The rate of sliding can also be controlled using a suitable parameter. Finally, the output of convolution layer is stacked and activation map defines the input for next layer to be propagated.

Our architecture is based on the Sequential model, which allows the model to stack sequential layers of the network in order from input to output. The convolution layer uses filters that perform convolution operations as it is scanning the input image with respect to its dimensions. Its hyper-parameters include the filter size and stride, which represents the step size between consecutive receptive filters. The resulting output is called feature map or activation map. First 2D convolutional layer is added to process the input breast images. The first argument passed to the convolution layer function is the number of output channels and in our case, we have used 16 output channels. In our model, we have used 3x3 filter kernel, with stride 1, which slide across the width and height of the image, to perform spatial convolution. During the experiments we have also experience with different kernel size (from 1 to 7) to analyze the differences. We have noticed that the small kernel size 1×1 acts as a bottleneck. We have considered padding so that the input image gets fully covered by the filter and stride specified in this model. We have applied the rectified linear unit (RELU) activation function, which is defined mathematically as: f(x) = (0,x), where the input is represented by x. Max pooling is used to reduce the spatial size of the feature map. The Maxpooling function downsamples the input representation by taking the maximum value over the window defined by pool_size for each dimension along the feature’s axis. Similarly, the procedure is repeated with two more convolutional layers with 32 and 64 output channels. Initially, we have used a filter of 2x2 max pooling. Inspired from the work conducted in [30] for addressing the degradation problem, we have used the same concept of deep residual learning framework. It works on repeated units of convolution layers having the filter size 1x1,3x3 and 1x1. Global average pooling was used to compute the activation of each feature map.

Results and performance evaluation

To evaluate the models, the datasets were divided into three sets namely training, validation and testing set. The training set, which represents the sample data, is used to fit the model. The latter learns from this data, which is 60% taken from the dataset. To ensure that the four classes (mass, carcinoma, asymmetry and calcifications) are well represented, 15% from each category were considered. To avoid unbiassed evaluation on the training set, a validation set is chosen to tune the parameters. Out of the remaining 40%, 20% was allocated to the evaluation set and 20% to the test set. The main role of the validation set is to choose the best combination of parameters for our model. As for the test set, it is being used for the final evaluation of the model. Thus, the metrics adopted are training accuracy, validation accuracy and testing accuracy. Initially, during testing, the training accuracy that was achieved was 96.8% and a validation accuracy of 36.7%.

The initial results clearly show that the model is overfitting and is not performing well on generalized datasets. To address this problem, drop-out regularization was considered. In fact, drop-out regularization is known to ignore both hidden and visible neurons which are selected at random during the training process. It removes inputs to a layer which might be the features in the image or activation from a preceding layer. Eventually, this stimulates a different network structure which forces the model to learn more robust features, thus, increasing the generalizing power of the model. The datasets were also re-constructed using the raw data and the augmented images. Initially, attention was not given to the proportion of images for the three classes of images. While experimenting with individual classes, it was seen that there were too few images in one class, and this may be affecting the validation performance. Images were augmented and more images were considered under each class. After drop-down regularization and the generation of the new dataset, we have achieved a training accuracy of 92.8% and validation accuracy of 86.7%. As reported by [31], overfitting hinders the generalization of models. The solutions investigated are early stopping, removal of noise in the data, data augmentation to fine tune parameters and regularization. Likewise, we have investigated the application of drop-out regularization and use augmented datasets to handle overfitting leading to better accuracy.

The model is further enhanced after experimenting with learning rate. After several trials, it was deduced that the model could preserve primitive features of the bottom layers. However, it was observed that more training was required to learn the higher order features. In fact, we updated the weights of network at each epoch and at certain value of learning rate during training process. It is very important to choose an optimal learning rate to minimize error. One potential solution is to choose a constant learning rate. But still, the system will cause fluctuations in the curve if the value is larger than the optimal value. On the other hand, the model will reduce the convergence rate if the chosen value is less than the optimal value. In addition, choosing an optimal value of learning rate is a challenging task as can be seen in [32]; and thus, we have used two ways to make the learning rate adaptive in entire training process during the experimentation [33]. In first approach, we have considered learning curve which is checked at each step of implementation. If the curve is found to be increasing, then the rate of learning is reduced by multiplication factor which is done by a variable coefficient whose value is less than one. If the increasing trend continues the learning rate is further increased, again by multiplying by a variable coefficient having its value more than one. In the similar manner, if we find constant error curve then we can maintain constant value of rate of learning. In Eq (1), we set the learning rate at each step, in which η(t) indicates the learning rate value after doing nth epoch. The values of variables λ1(t) and λ2(t) respectively in Eq (2) and Eq (3) are coefficients over time that vary at each iteration of training process. The value of λ1(t) is less than one and value of λ2(t) is kept greater than one. This helps to decrease and increase the rate of learning in comparison with previous step used. The coefficients α1 and α2 should be evaluated between the interval λ(0) and λ(0) ± γ, of λ(t) value. In this manner, the values are chosen that can also be seen in Eq (4), where ‘m’ represents the maximum value of epochs used in training; γ indicates the interval for choosing the desired changes in λ and β. This assists in determining the speed while the values in the interval tends to decrease. We can clearly see that smaller is the value of β, faster is the change in value of the λ by either increasing or decreasing to the final value.

η(t+1)={η(t)*λ1(t)if(E(t)>E(t+1)η(t)*λ2(t)if(E(t)<E(t+1)η(t)if(E(t)=E(t+1) (1)
λ1(t)=λ1(t1)α1etβ (2)
λ2(t)=λ2(t1)+α2etβ (3)
αe1β+αe2β++αemβ=γ (4)
α=γ(1e1β)e1β(1emβ) (5)

On the other hand, Adaptive Moment Estimation (ADAM) optimization algorithm has also been applied and analyzed. Adam attempts to compute adaptive learning rates for each parameter. This is very useful in complex network structures because different parts of the network have different sensitivity to weight adjustment. At times, in breast images, the abnormalities cover a very small region and cannot be detected easily. A very sensitive part usually requires a smaller learning rate. Adam maintains two additional variables mt and vt for each variable to be trained:

mt=β1mt1+(1β1)gt (6)
vt=β2vt1+(1β2)gt2 (7)

where t represents tth iteration and gt is the calculated gradient. mt and vt are moving averages of the gradient and the squared gradient. From the statistical perspective, mt and vt are values of the estimates of the first and second moment respectively. The first moment here is the mean value and the second moment corresponds to the uncentered variance of the gradients. Now, when the values of mt and vt are initialized as vectors of 0’s, then these are biased towards 0, especially during the initial steps, and especially when β1 and β2 are close to 1. To solve this problem, m^t and v^t are being introduced.

m^t=mt1β1t (8)
v^t=vt1β2t (9)

ADAM’s weight are being updated as follows:

wt+1=wtηv^t+ϵm^t (10)

In practice, ADAM converges quickly. Other parameters do not need adjustment. To ensure that the performance of the model does not remain stagnant, the ReduceLROnPlateau is being used. It monitors the accuracy of the model and decreases the learning rate by a certain factor once the accuracy has stop improving. The fluctuation was seen in learning rate error curve while using the constant rate in the training. The ADAM optimizer was also adopted by [34] and have achieved positive results.

The refined model was then tested on the testing set. The enhanced model has achieved a training accuracy of 96.7% on the training set and 83.3% on the validation set. This shows that the enhanced CNN performs better and could learn more general features to correctly classify the breast abnormalities. Fig 5 shows the loss on the training data and the loss on the validation data which is converging towards zero.

Fig 5. Loss on training and validation data.

Fig 5

Actually, the training loss is actually higher than the validation loss. This clearly shows we are controlling overfitting data in our model. The model was then applied to the testing set and the performance is summarized in Table 1.

Table 1. Performance of the enhanced CNN model.

True Labels vs Predicted Labels Calcifications Mass Asymmetry Carcinomas
Calcifications 0.91 0.06 0.08 0.03
Mass 0.01 0.85 0.06 0.04
Asymmetry 0.03 0.04 0.82 0.03
Carcinomas 0.04 0.05 0.04 0.90

An overall performance of 88% was achieved for the multi-class classification of the abnormalities.

The ResNet50 and our proposed model have been refined by minimizing the loss function and fine tuning all the parameters using the training accuracy and validation accuracy. In the testing phase, ResNet50 has achieved an overall accuracy of 81.5% while our enhanced model has reached 88%. Table 2 summarizes the results attained.

Table 2. Summary of performance of the models developed.

Model Testing Accuracy (Overall)
RestNet50 Model 81.5%
Enhanced CNN Model 88%

Comparison with existing state-of-the art

Up to our knowledge, no work conducted so far have achieved the four classification of abnormalities: masses, calcifications, carcinomas and asymmetry. In one work, [35] have used CNN to classify carcinoma, which is one type of abnormality used to detect the possibility of breast cancer. The authors have achieved an accuracy of 77.8%. In order to avoid overfitting, the authors have applied data augmentation for the different patches. Likewise, CNN have been trained on histopathological images and fusion of breast images was investigated [36]. In this work, it was reported that CNN performed better compared to hand-crafted techniques. It was concluded that CNN should be further investigated as well as ways to improve the accuracy. We analyzed the genes expressions that helped in detecting and classifying the abnormalities in the breast cancer images using CNN. In line, [37] have explored CNN to classify breast cancer using genes expressions. The results have demonstrated scopes in using CNN. However, the work has not explored multi-class classification which is more challenging task. In a work conducted by [38], CNN was used to classify breast cancer into benign and malignant. To detect masses, feature fusion was applied using CNN deep features. Extreme learning machine (ELM) clustering was then applied to classify masses into benign and malignant. Most of the work was conducted so far focusses on ordinary classification of cancer tissues as either malignant or benign using CNN. The preliminary results provide scopes for the further exploration of CNN. In an analysis conducted by [39], it was reported that there is an urgent need to improve imaging analysis and explore detection techniques using AI for breast cancer detection. Breast cancer is determined by several abnormalities and thus, this work contributes in detecting and classifying these abnormalities. Our work has achieved an overall accuracy of 88%. Table 3 compares our proposed solution with existing research work.

Table 3. Summary of comparison of proposed model with existing works.

Existing Work Research Conducted Accuracy Discussion
Araújo et al. (2017) [35] Classification of only 1 type of abnormality, that is, carcinoma into normal, in situ carcinoma and invasive carcinoma An Accuracy of 77.8% was achieved for four classes In this work, only one type of abnormality has been considered. However, for this abnormality, the classification was done on 4 different classes.
An Accuracy of 77.8% was achieved for four class
Spanhol et al (2016) [36] Investigation on the different layers of CNN to investigate on the parameters and performance An accuracy of around 85% for the fusion of images In this work, the different layers of CNN were analyzed. Fusion was done by combining the different patches results with the whole image. It was mentioned that more research should be carried out to determine ways of achieving better accuracy.
Elbashir et al. (2019) [37] Classification of breast cancer using genes expressions. An accuracy of 98.76% to determine whether a breast is cancerous or not This work used the gene expressions to classify cancers. Research have shown that CNN is a potential technique that can be used to investigate patterns. However, the work has not explored multi-class classification which is more challenging.
Wang et al. (2019) [38] Detection of masses using CNN through feature fusion An overall accuracy of 87% CNN was used to detect masses. The ELM was used to classify masses into benign and malignant. Multi-class classification was not investigated
Proposed Work Detection and Classification of breast cancer abnormalities An overall accuracy of 88% In our work, we have been able to detect masses, calcifications, carcinomas and asymmetry mammograms. The detection of these abnormalities helps in the early detection and eventually diagnosis of breast cancer

Conclusion

The diagnosis of breast cancer in an early stage can help in the reduction of the mortality caused by breast cancer. The remarkable success of deep learning and image processing has spurred the development of automated medical applications. There are various abnormalities in the breast like carcinomas, masses, lumps, calcification and asymmetry, which can indicate a potential breast cancer. A deep learning model has been developed to help the radiologist in reading a breast image. Firstly, a pre-trained model, namely the Resnet 50 has been applied to our datasets, created from CBIS-DDSM and UPMC. Then, data Augmentation was applied on the datasets to obtain a varied set of images. To further improve the model, an enhanced CNN was formulated by monitoring the learning rate. It was deduced that constant learning rate does not allow the model to learn new parameters. In the initial testing phase, the model shows that it was overfitting and thus, the parameters were improved. Our enhanced model achieved a recognition rate of 88% for the multiclass classification of the various abnormalities namely mass, calcification, asymmetry and carcinomas. More breast image datasets taken under varied conditions should be made available to the research community to improve the diagnosis of breast cancer. Followed by image processing, artificial intelligence techniques can be applied to determine the most appropriate imaging technology to be adopted. Alongside, techniques should be improved further to detect and classify the abnormality in the breast at an early stage, thus, determining the most appropriate treatment to be administered to the patient to reduce deaths caused by breast cancer.

Acknowledgments

We would like to thank Dr Sunilduth Baichoo and Mr Somveer Kishnah for their valuable feedback.

Data Availability

All relevant data are within the manuscript and its Supporting Information files.

Funding Statement

The author(s) received no specific funding for this work.

References

  • 1.Tahmooresi M, Afshar A, Rad BB, Nowshath KB, Bamiah MA. Early detection of breast cancer using machine learning techniques. Journal of Telecommunication, Electronic and Computer Engineering (JTEC). 2018Sep26;10(3–2):21–7. [Google Scholar]
  • 2.World Health Organisation. 2018 [cited 10 March 2020]. Available from: https://www.who.int/cancer/prevention/diagnosis-screening/breast-cancer/en/
  • 3.Patel B. C., Sinha G. R. Abnormality Detection and Classification in Computer-aided Diagnosis CAD) of Breast Cancer Images. Journal of Medical Imaging and Health Informatics. 2014. 4(6): 881–885. [Google Scholar]
  • 4.Sinha G R. Gray level clustering and contrast enhancement (GLC–CE) of mammographic breast cancer images. CSI Transactions of Springer. 2015. 2(4): 279–286. [Google Scholar]
  • 5.Sinha G. R., Patel BC. Medical Image Processing: Concepts and Applications. Prentice Hall of India. 2014. First Edition. [Google Scholar]
  • 6.Nwe Ni K, Myat, MNW, Phyu MT, Sinha G.R., Phyu PK, Moe MH. Breast Cancer Research using Novel Segmentation Methods. 2019 Joint International Conference on Science, Engineering and Innovation. 2019. Sept. 16; 270–274.
  • 7.Caplan L. Delay in breast cancer: implications for stage at diagnosis and survival. Frontiers in public health. 2014Jul29;2:87. doi: 10.3389/fpubh.2014.00087 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.Kirubakaran R, Jia TC, Aris NM. Awareness of breast cancer among surgical patients in a tertiary hospital in Malaysia. Asian Pacific journal of cancer prevention: APJCP. 2017;18(1):115. doi: 10.22034/APJCP.2017.18.1.115 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9.Richards M.A., Westcombe A.M., Love S.B., Littlejohns P. and Ramirez A.J., 1999. Influence of delay on survival in patients with breast cancer: a systematic review. The Lancet, 353(9159), pp.1119–1126. [DOI] [PubMed] [Google Scholar]
  • 10.Fenton JJ, Taplin SH, Carney PA, Abraham L, Sickles EA, D’Orsi C, et al. Influence of computer-aided detection on performance of screening mammography. New England Journal of Medicine. 2007Apr5;356(14):1399–409. doi: 10.1056/NEJMoa066099 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11.Cole EB, Zhang Z, Marques HS, Edward Hendrick R, Yaffe MJ, Pisano ED. Impact of computer-aided detection systems on radiologist accuracy with digital mammography. American Journal of Roentgenology. 2014Oct;203(4):909–16. doi: 10.2214/AJR.12.10187 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12.Lehman CD, Wellman RD, Buist DS, Kerlikowske K, Tosteson AN, Miglioretti DL, Breast Cancer Surveillance Consortium. Diagnostic accuracy of digital screening mammography with and without computer-aided detection. JAMA internal medicine. 2015Nov1;175(11):1828–37. doi: 10.1001/jamainternmed.2015.5231 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.Agarap AF. On breast cancer detection: an application of machine learning algorithms on the wisconsin diagnostic dataset. InProceedings of the 2nd international conference on machine learning and soft computing 2018 Feb 2 (pp. 5–9).
  • 14.Jog NV, Mahadik SR. Implementation of classification technique for mammogram image. IOSR Journal of Electronics and Communication Engineering (IOSR-JECE). 2014;9(1):76–8. [Google Scholar]
  • 15.Hamad YA, Simonov K, Naeem MB. Breast cancer detection and classification using artificial neural networks. In2018 1st Annual International Conference on Information and Sciences (AiCIS) 2018 Nov 20 (pp. 51–57). IEEE.
  • 16.Nahid AA, Mehrabi MA, Kong Y. Histopathological breast cancer image classification by deep neural network techniques guided by local clustering. BioMed research international. 2018Mar7;2018. doi: 10.1155/2018/2362108 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17.Moh’d Rasoul A, Al-Gawagzeh MY, Alsaaidah BA. Solving mammography problems of breast cancer detection using artificial neural networks and image processing techniques. Indian journal of science and technology. 2012Apr;5(4):2520–8. [Google Scholar]
  • 18.Elmanna ME, Kadah YM. Implementation of practical computer aided diagnosis system for classification of masses in digital mammograms. In 2015 International Conference on Computing, Control, Networking, Electronics and Embedded Systems Engineering (ICCNEEE) 2015 Sep 7 (pp. 336–341). IEEE.
  • 19.Rodríguez-Ruiz A, Krupinski E, Mordang JJ, Schilling K, Heywang-Köbrunner SH, Sechopoulos I, et al. Detection of breast cancer with mammography: effect of an artificial intelligence support system. Radiology. 2019Feb;290(2):305–14. doi: 10.1148/radiol.2018181371 [DOI] [PubMed] [Google Scholar]
  • 20.Sasaki M, Tozaki M, Rodríguez-Ruiz A, Yotsumoto D, Ichiki Y, Terawaki A, et al. Artificial intelligence for breast cancer detection in mammography: experience of use of the ScreenPoint Medical Transpara system in 310 Japanese women. Breast Cancer. 2020Feb12:1–0. doi: 10.2147/BCTT.S232021 [DOI] [PubMed] [Google Scholar]
  • 21.Joseph AM, John MG, Dhas AS. Mammogram image denoising filters: A comparative study. In2017 Conference on emerging devices and smart systems (ICEDSS) 2017 Mar 3 (pp. 184–189). IEEE.
  • 22.Athira P, Fasna KK, Krishnan A. An overview of mammogram noise and denoising techniques. International Journal of Engineering Research and General Science. 2016Mar;4(2). [Google Scholar]
  • 23.Liu X, Song L, Liu S, Zhang Y. A review of deep-learning-based medical image segmentation methods. Sustainability. 2021Jan;13(3):1224. [Google Scholar]
  • 24.Maheshwari S, Kanhangad V, Pachori RB. Cnn-based approach for glaucoma diagnosis using transfer learning and lbp-based data augmentation. arXiv preprint arXiv:2002.08013. 2020 Feb 19.
  • 25.Nawaz M, Sewissy AA, Soliman TH. Multi-class breast cancer classification using deep learning convolutional neural network. Int. J. Adv. Comput. Sci. Appl. 2018Jun1;9(6):316–32. [Google Scholar]
  • 26.Kabir HM, Alam SB, Azam MI, Hussain MA, Sazzad AR, Sakib MN, et al. Non-linear down-sampling and signal reconstruction, without folding. In 2010 Fourth UKSim European Symposium on Computer Modeling and Simulation 2010 Nov 17 (pp. 142–146). IEEE.
  • 27.Chougrad H, Zouaki H, Alheyane O. Deep convolutional neural networks for breast cancer screening. Computer methods and programs in biomedicine. 2018Apr1;157:19–30. doi: 10.1016/j.cmpb.2018.01.011 [DOI] [PubMed] [Google Scholar]
  • 28.Yadav SS, Jadhav SM. Deep convolutional neural network based medical image classification for disease diagnosis. Journal of Big Data. 2019Dec;6(1):1–8. [Google Scholar]
  • 29.Lundervold AS, Lundervold A. An overview of deep learning in medical imaging focusing on MRI. Zeitschrift für Medizinische Physik. 2019May1;29(2):102–27. doi: 10.1016/j.zemedi.2018.11.002 [DOI] [PubMed] [Google Scholar]
  • 30.He K, Zhang X, Ren S, Sun J. Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition 2016 (pp. 770–778).
  • 31.Ying X. An overview of overfitting and its solutions. In Journal of Physics: Conference Series 2019 Feb 1 (Vol. 1168, No. 2, p. 022022). IOP Publishing.
  • 32.Hamid NA, Nawi NM, Ghazali R, Salleh MN. Accelerating learning performance of back propagation algorithm by using adaptive gain together with adaptive momentum and adaptive learning rate on classification problems. In International Conference on Ubiquitous Computing and Multimedia Applications 2011 Apr 13 (pp. 559–570). Springer, Berlin, Heidelberg.
  • 33.Amiri Z, Hassanpour H, Khan NM, Khan MH. Improving the performance of multilayer backpropagation neural networks with adaptive leaning rate. In 2018 International Conference on Advances in Big Data, Computing and Data Communication Systems (icABCD) 2018 Aug 6 (pp. 1–4). IEEE.
  • 34.Adeshina SA, Adedigba AP, Adeniyi AA, Aibinu AM. Breast cancer histopathology image classification with deep convolutional neural networks. In 2018 14th international conference on electronics computer and computation (ICECCO) 2018 (pp. 206–212). IEEE.
  • 35.Araújo T, Aresta G, Castro E, Rouco J, Aguiar P, Eloy C, et al. Classification of breast cancer histology images using convolutional neural networks. PloS one. 2017Jun1;12(6):e0177544. doi: 10.1371/journal.pone.0177544 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 36.Spanhol FA, Oliveira LS, Petitjean C, Heutte L. Breast cancer histopathological image classification using convolutional neural networks. In 2016 international joint conference on neural networks (IJCNN) 2016 Jul 24 (pp. 2560–2567). IEEE.
  • 37.Elbashir MK, Ezz M, Mohammed M, Saloum SS. Lightweight convolutional neural network for breast cancer classification using RNA-seq gene expression data. IEEE Access. 2019Dec18;7:185338–48. [Google Scholar]
  • 38.Wang Z, Li M, Wang H, Jiang H, Yao Y, Zhang H, et al. Breast cancer detection using extreme learning machine based on feature fusion with CNN deep features. IEEE Access. 2019Jan16;7:105146–58. [Google Scholar]
  • 39.de Figueiredo GN, Ingrisch M, Fallenberg EM. Digital Analysis in Breast Imaging. Breast Care. 2019;14(3):142–50. doi: 10.1159/000501099 [DOI] [PMC free article] [PubMed] [Google Scholar]

Decision Letter 0

Gulistan Raja

13 Jul 2021

PONE-D-21-18653

Multi- class classification of breast cancer abnormalities using Deep convolutional neural network (CNN)

PLOS ONE

Dear Dr. Heenaye- Mamode Khan,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

The manuscript had been reviewed by 2 reviewers. Reviewer 1 was of the view that manuscript partly describes a technically sound piece of scientific research and recommended major revision. Reviewer 2 was of the opinion that your manuscript describes a technically piece of scientific research, however, he had made certain observations and also recommended major revision. 

After thorough consideration of comments of Reviewer 1 and Reviewer 2, my decision is "major revision". Please incorporate comments raised by both reviewers. 

Please submit your revised manuscript by Aug 27 2021 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: http://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols.

We look forward to receiving your revised manuscript.

Kind regards,

Gulistan Raja

Academic Editor

PLOS ONE

Journal Requirements:

When submitting your revision, we need you to address these additional requirements.

1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at

https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and

https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf

2. Please amend your Methods section to provide links to all of the online data sets used in the study.

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Partly

Reviewer #2: Yes

**********

2. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

Reviewer #2: Yes

**********

3. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: Yes

**********

4. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

**********

5. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: The research paper contributes towards classification of breast cancer abnormalities using Deep convolutional neural network. The four types of breast cancer abnormalities considered in the work are masses, calcifications, carcinomas and asymmetry mammograms.

Introduction part is well written. Introduction/Related work must conclude with clear research gap.

Data set description is clearly mentioned in the proposed work. Authors have used CLAHE for image pre-processing. Need to analyze this image pre-processing technique on various parameters, namely, Illumination corrections, Blur and focus corrections, Filtering and noise removal, Thresholding Edge enhancements, Morphology, Segmentation, Region processing and filters, Point processing, Color space conversions (Whichever is appropriate for given data set).

Proposed architecture is well defined.

Results section need to be improvised. Authors may implement classification using other deep learning models and compare results in terms of validation accuracy and loss, with enhanced CNN model.

Authors are suggested to include more analysis graphs in the results section. Authors may analyze the output by considering various epochs in the training/testing phase. Authors need to mention various parameters used during the implementation of CNN. The typical parameters could be Activation Function, Optimizer, Loss function, Metric (Whichever is suitable for proposed work).

Authors are suggested to include Future work in this area.

Overall evaluation:

The researchers have attempted to give a solution to the classification of breast cancer abnormalities using Deep convolutional neural network. The paper needs a revision by incorporating the changes mentioned above before it is accepted for publication.

Reviewer #2: In this work, author have developed a Deep Convolutional Neural Network (CNN) to segment and classify the various types of breast abnormalities, such as calcifications, masses, asymmetry and carcinomas unlike existing research

work. The authors are encouraged to consider the following issues to further improve the quality of the paper.

1.The language expression needs to be more concise. There are many grammatical mistakes in the paper, so I suggest author revising revise them.

2.Data pre-processing, Results and Performance Evaluation,

The Development of the Enhanced CNN Model should not too much describe the previous work, but focus on the work of this paper.

3.The Related Work part mentioned that 30% test set and 70% training set were carried out, and after combining CBIS-DDSM Dataset, It is also divided into 60% training set and 40% test set. The Results and Performance Evaluation section also mentions having a validation set, How exactly did you divide the data set, please explain in detail.

4.The format of the table needs to be adjusted so that the words are on the same line as possible for easy reading.

5.Formula labels are duplicated and there are two different formulas (5).  Please check carefully

6.The reference is suggested to be in a unified format, and the number of reference 8 is wrong.

7.Please redesign pictures 4, 5 and 6 to ensure the sharpness of the pictures

8.Figure 5 should not take a random output as the result, but should look at the overall accuracy and the change of loss.

9.Try to keep the form to one page.

10.The paper should refer to the latest research work.

A Review of Deep-Learning-Based Medical Image Segmentation Methods, Sustainability, 2021, 13(3), 1224.

An image enhancement algorithm of video surveillance scene based on deep learning, IET Image Processing, 2021, https://doi.org/10.1049/ipr2.12286

**********

6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: Yes: Shuai Liu

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

PLoS One. 2021 Aug 26;16(8):e0256500. doi: 10.1371/journal.pone.0256500.r002

Author response to Decision Letter 0


22 Jul 2021

First of all, I would like to thank you for the reviews. The comments/ suggestions from the reviewers were valuable and has allowed us to improve our paper.

I have attached a file "Response to Reviewer- Kindly access same"

Thank you

Attachment

Submitted filename: Response to Reviewers.docx

Decision Letter 1

Gulistan Raja

9 Aug 2021

Multi- class classification of breast cancer abnormalities using Deep convolutional neural network (CNN)

PONE-D-21-18653R1

Dear Dr. Heenaye- Mamode Khan,

We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements.

Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication.

An invoice for payment will follow shortly after the formal acceptance. To ensure an efficient process, please log into Editorial Manager at http://www.editorialmanager.com/pone/, click the 'Update My Information' link at the top of the page, and double check that your user information is up-to-date. If you have any billing related questions, please contact our Author Billing department directly at authorbilling@plos.org.

If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

Kind regards,

Gulistan Raja

Academic Editor

PLOS ONE

Additional Editor Comments (optional):

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation.

Reviewer #1: All comments have been addressed

Reviewer #2: All comments have been addressed

**********

2. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

Reviewer #2: Yes

**********

3. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

Reviewer #2: Yes

**********

4. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: Yes

**********

5. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

**********

6. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: Authors have addressed all the comments given. The following changes are incorporated by the authors:

Research gap has been mentioned in the related work.

Authors have given clarification for the preprocessing of experimental images. The comparison of ResNet50 and CNN techniques is included in the results section.

Future work has been mentioned.

I appreciate and congratulate all the authors for their efforts in preparing this research article and wish them all the best for their future work.

The paper may be accepted in the current state.

Reviewer #2: The revision seems good enoguh. I have no other comments since all previous comments from me were addressed.

**********

7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

Acceptance letter

Gulistan Raja

16 Aug 2021

PONE-D-21-18653R1

Multi- class classification of breast cancer abnormalities using Deep Convolutional Neural Network (CNN)

Dear Dr. Heenaye- Mamode Khan:

I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now with our production department.

If your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information please contact onepress@plos.org.

If we can help with anything else, please email us at plosone@plos.org.

Thank you for submitting your work to PLOS ONE and supporting open access.

Kind regards,

PLOS ONE Editorial Office Staff

on behalf of

Dr. Gulistan Raja

Academic Editor

PLOS ONE

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    Attachment

    Submitted filename: Response to Reviewers.docx

    Data Availability Statement

    All relevant data are within the manuscript and its Supporting Information files.


    Articles from PLoS ONE are provided here courtesy of PLOS

    RESOURCES