Skip to main content
Journal of Digital Imaging logoLink to Journal of Digital Imaging
. 2023 Jul 25;36(6):2461–2479. doi: 10.1007/s10278-023-00878-x

ExpHBA Deep-IoT: Exponential Honey Badger Optimized Deep Learning For Breast Cancer Detection in IoT Healthcare System

R Rajeswari 1,, G V Sriramakrishnan 2, ChVidyadhari 3, K V Kanimozhi 4
PMCID: PMC10584750  PMID: 37491544

Abstract

Breast cancer (BC) is the most widely found disease among women in the world. The early detection of BC can frequently lessen the mortality rate as well as progress the probability of providing proper treatment. Hence, this paper focuses on devising the Exponential Honey Badger Optimization-based Deep Covolutional Neural Network (EHBO-based DCNN) for early identification of BC in the Internet of Things (IoT). Here, the Honey Badger Optimization (HBO) and Exponential Weighted Moving Average (EWMA) algorithms have been combined to create the EHBO. The EHBO is created to transfer the acquired medical data to the base station (BS) by choosing the best cluster heads to categorize the BC. Then, the statistical and texture features are extracted. Further, data augmentation is performed. Finally, the BC classification is done by DCNN. Thus, the observational outcome reveals that the EHBO-based DCNN algorithm attained outstanding performance concerning the testing accuracy, sensitivity, and specificity of 0.9051, 0.8971, and 0.9029, correspondingly. The accuracy of the proposed method is 7.23%, 6.62%, 5.39%, and 3.45% higher than the methods, such as multi-layer perceptron (MLP) classifier, deep learning, support vector machine (SVM), and ensemble-based classifier.

Keywords: Deep Covolutional Neural Network, Exponential Weighted Moving Average, Honey Badger Optimization, Local Gabor Binary Pattern, Local Optimal Oriented Pattern

Introduction

IoT is one of the developed networks, which can collect and exchange information through the sensors in the healthcare environment [1]. In the medical industry, IoT is deliberated as a proficient application system. Moreover, it is used to investigate the physical constraints of patients through the sensor nodes correlated with the body of the patient using smart portable devices [2]. Some of the factors, such as energy transmission, network topology, and computation capacity is effectively handled by IoT devices while transmitting the information throughout the network [3].

Cancer is the abnormal progression of cells in a specified region, which damages the linked tissues in the human body. Generally, the tumor can be classified into two types, such as benign and malignant. The malignant affects the various parts of the body, whereas benign affects the nearest cells of the affected region. BC is the second most widely spread cancer in the globe, which is largely affected by Malaysian women. Every year, more than 3500 women are affected by the BC [46]. The early detection and active prevention of BC have diminished the chances of death rate [6, 7].

Microscopic or X-ray imaging is used for identifying whether the tumor is malignant or not. However, this process is time-consuming [8]. During the analysis of BC detection, 3D and 2D mammogram images have been considered as input, which is generated by the 3D mammogram machines. However, the recognition of BC from these images is challenging for radiologists [8, 9]. ML is chosen as an effective diagnostic tool, which lowers the expenses of computing and physical practice [10]. DL is derived from machine learning, which can be used to train and identify the features from the BC dataset [11]. Most of the investigators have analyzed BC detection using BC benchmark data with WDBC [12] and WPBCC [13].

The purpose of this research is to categorize the BC using the EHBO-based DL technique. In this research, the devised EHBO-based DL algorithm performs two functions, one is the efficient routing and the other one is the training of DCNN. The main contribution of this paper is as follows:

  • EHBO-based DCNN: The BC classification is completed by DCNN where the training process is done by EHBO. The devised EHBO algorithm is utilized to feed the information to the BS through the best path. The EWMA is formed by combing EWMA and HBO.

The structure of this essay is as follows. The literature review and problems with the BC classification model are portrayed in the “Literature Review” section, the system model is described in the “System Model” section, the implemented scheme is shown in the “Proposed EHBA-Based DCNN for BC Classification” section, the results and analysis of the devised scheme are shown in the “Results and Discussion” section, and the conclusion of this paper is shown in the “Conclusion” section.

Literature Review

Gopal et al. [13] implemented a machine learning approach for the prediction of BC with IoT devices. Although the MLP classifier offered the maximum accuracy with less error rate, the computation cost of this approach was high. Savitha et al. [14] modeled the OKM-ANFIS for predicting BC in an IoT environment. This method achieved more accuracy and security, but it failed to authenticate the attack during the data transfer. Zahir et al. [8] devised the DL model for detecting BC in IoT using histopathological images. The developed approach was built on a Raspberry Pi, which could run on a small and portable processor. However, the computational complexity was very high. Chokka and Rani [15] introduced the AdaBoost algorithm with feature selection to detect BC in IoT. This method was simple, rapid, and accurate in all tumor genotypes, but the implementation time was high. To reduce the processing time, Khamparia et al. [9] modeled the hybrid transfer learning scheme for diagnosing BC with mammography. This method decreased the false negative as well as false positive rates, which progresses the efficacy of mammogram analysis. However, it failed to categorize the affected cells. Memon et al. [16] designed the SVM with a recursive feature selection scheme for detecting the BC. Here, the devised system performance was outstanding, due to the assortment of excess suitable features. Khan et al. [6] devised an ensemble-based classifier for detecting and classifying the BC in IoT using breast cytology images. This method minimized the computational complexity, but deep learning approaches were not used, which affected the classification accuracy. Suresh et al. [17] designed a hybrid classifier for monitoring the health condition of patients to identify BC. The proposed classifier effectively resolved the difficulty of misclassified malignant, but the training time of this approach was extreme.

System Model

IoT network is modeled by interconnecting various hardware devices, such as sensors, actuators, transmitters, and receivers. Here, all the devices that exist in the network are linked via the wireless communication medium with a particular explicit communication range. Moreover, the IoT network comprises three nodes, namely, CH, BS, and normal nodes. In addition, all the nodes are associated with each other to form the cluster. The system model is conveyed in Fig. 1.

Fig. 1.

Fig. 1

IoT system model

Energy Model

For performing every operation in the communication medium, like transmitting, receiving, and forwarding the packet over the chosen path, each node spends energy [18]. During transmission, the energy of the node is computed as

Obu=Onu+Otxu+Orxu+Oplu 1
Obu=anuHn+Htx+HrxKL+apluHplQ 2

where Onu, Otxu, Orxu, and Oplu indicate the energy consumed by the nodes while listening, broadcasting, receiving, as well as sleeping, consistently. HnHtx, Hrx, and Hpl signify the listening, transmission, reception, as well as sleep mode, consistently; Q specifies the battery voltage in nodes; K specifies the packet length; and L signifies the data rate. Additionally, aplu and anu are calculated as

aplu=Ib-Os=2Bo-2SF 3
anu=Ib-atxu+arxu+aplu+aCCA 4

where Bo signifies the order of Beacon and SF demonstrates the order of superframe. Assume Prxu=0 or arxi=0 when u is nominated as the source node, and then, Otxu=0 or atxi=0 when u is nominated as the target node. For revising Eq. (3),

Obu=anuHn+HtxKL+apluHplQifuisTxanuHn+HrxKL+apluHplQifuisRx 5

Thus, the residual node energy is demonstrated as

Oeu=Oou-Obu+Ocu 6

where Oou specifies the initial energy and Ocu represents the harvested energy.

Mobility Model

The mobility model [19] is intended to describe how mobile nodes move as well as how their location, velocity, and acceleration alter over time. Meanwhile, the mobility patterns play a crucial role in deciding how the routing algorithm is presented. The mobility model is set up to evaluate the sensor node’s motion based on its position, acceleration, and velocity at a certain time instant. Assume that the node Bf and node Bx are located at (δ,κ) and (δ,κ) location at time a=0. In this model, the nodes Bf and Bx are transmitted to the new place with velocities depending on two dissimilar angles Ωb,h and Ωb,t. Additionally, the node bh moves a space Gb,x, and node Bx transmits through the distance Gb,x at time a=0. Assume δy,κy and δx,κx portrayed the updated localities utilized by nodes By and Bx at the time a=1. At the specific time instant a, the distance is calculated among nodes at By position δy,κy, and δx,κx is represented by

G(by,bx,a)=|δy-δx|2+|κy-κx|2 7

where the updated position of nodes Bx and By is articulated as δy,yy and δx,κx.

Proposed EHBA-Based DCNN For BC Classification

The main goal of this research is to create the EHBA-based DCNN, which is a BC detection method in an IoT healthcare system. Routing and detection are the processing steps involved in this research. IoT nodes are first simulated, and after that, the simulated nodes collect images from the database. Then, using an EHBA-based DCNN algorithm created based on fitness parameters such as distance, energy, throughput, delay, and connection quality, the gathered images are provided to the BS for BC detection through the optimal path. The steps below are used in the BS to detect BC. The obtained mammography pictures are first supplied to the pre-processing module, which uses Region of Interest (RoI) extraction. Once the pre-processing is done, then the affected region is segmented using the U-net model. Then, the texture features, such as LGBP [20] feature, GLCM feature [21], and LOOP feature [22] are extracted initially from the segmented region. At that time, the statistical features are not extracted in this stage since the statistical features do not provide a better classification outcome before augmenting the segmented image. For that, the statistical feature is extracted after the augmentation. The selected features are augmented by rotation, flipping, cropping, and zooming to enlarge the size of features for attaining better results. Then, to get better results, statistical features including area, energy, homogeneity, mean, variance, entropy, kurtosis, skewness, and HOG [23] properties are extracted. The BC classification is finally completed using Deep CNN [24], whose weight will be learned using the created EHBA algorithm. Figure 2 displays the block diagram of the EHBO-based DCNN.

Fig. 2.

Fig. 2

Block diagram of proposed EHBO-based DCNN for BC classification

Routing Using Proposed EHBO Algorithm

This section explains the routing process based on the proposed EHBO algorithm, which is modeled by the assimilation of the EWMA [25] and HBO [26] algorithms. The EWMA algorithm was designed by Roberts, who discovered it by adapting the average run length aspects of EWMA control approaches. Moreover, the EWMA algorithm is useful for effectively recognizing the small variations in the target value. Likewise, the HBO algorithm was modeled by considering the digging aspects of honey bees. It resolves the optimization issues with complex search space. Moreover, the convergence speed as well as the exploration rate is extraordinary. Thus, the EHBO is modeled by taking advantage of both optimization algorithms.

Encoding of Agent’s Location

This section describes the encoding location of agents, which is illustrated in Fig. 3. Solution encoding is utilized to pick the optimal CH according to the fitness function. Here, each part indicates the nodes’ index, and the solution dimension is 1×n, where n depicts the number of required nodes. In addition, the selected CH does not surpass the total node count, and the selected path is from 2810.

Fig. 3.

Fig. 3

Solution encoding

Fitness Function

The optimum solution is chosen using the fitness function. In this case, the fitness is calculated using variables including distance, energy, link quality, throughput, and delay.

Fuvmin=1h2u=1hv=u+1hGuv+guv+1-Quv+E(hu)+(1-Muv) 8

where Guv specifies the distance, guv indicates the delay, Quv specifies the link quality, Muv indicates the throughput, and E(hu) specifies the consumed energy.

The fitness parameters are explained below.

  1. Distance: It is the distance between the two specified nodes and is portrayed as

Guv=Pu-PvNF 9

where Pu-Pv indicates the distance among uth node and vth node and NF specifies the normalization factor.

  • b)

    Delay: The term delay is defined as the time taken to transfer data from source to destination. The formula is given as

guv=1Pu=1nhu 10

where P demonstrates the overall node count and n specifies the node count in a path.

  • iii)

    Link quality: It is expressed as the ratio between power consumed by the node u and v, and the distance among the nodes u and v, which is portrayed as,

Quv=XuvGuv 11

where Xuv specifies the power consumed by the uth and vth node and Guv specifies the distance among the node u and v.

  • iv)

    Throughput: The term throughput is expressed as the successfully transmitted packets to the destination and is expressed in Eq. (13).

Muv=u=1hEudsDp 12

where Eu represents the energy of the node u, ds specifies the data packet size, and Dp signifies the data packet.

  • e)

    Energy consumption: Energy consumption, which is described in the “Energy Model” section, is used to calculate how much energy is used by each node.

Algorithmic Steps of EHBO+

  • i)

    Initialization: The count of a honey badger with the population size as Z and the respective locations are expressed as

fd=LBd+k1×UBd-LBd 13

where k1 designates the random number amid 0 and 1, fd denotes the placement of dth honey badger, and UBd and LBd denote the lower bound and upper bound.

  • b)

    Intensity evaluation: The definition of intensity takes into account both the prey’s level of attention and their proximity to the honey badger. If any of the locations have a high smell, then the motion of prey is faster and is portrayed as

Jd=k2×W4πsd2 14
W=fd-fd+12 15
sd=fprey-fd 16

where W signifies the strength of concentration and sd denotes the spacing between prey as well as dth badger.

  • iii)

    Update the density factor: The density factor is renewed based on the time-changing randomization. Thus, the decreasing value of Ξ decreases the iteration count and is expressed as

Ξ=K×exp-rrmax 17

where K specifies the constant, Ξ specifies the density factor, and rmax specifies the maximum iteration, which is greater than or equal to one.

  • iv)

    Escape from local optimum: This process is done using two phases, namely, digging and honey phase. This algorithm utilizes the flag D, for altering the exploration direction to carefully study the search space.

  • v)

    Renew the agent position: The location of the agent is updated based on two phases, like digging and honey phases.

Digging phase: The honey badger updates its location in a cardioid fashion throughout this phase, and its mobility is described as

fnew=fprey+D×ω×R×fprey+D×k3×μ×sd×Cos(2πk4)×1-Cos(2πk5 18

where

sd=fprey(r)-fd(r) 19

Let us assume that

fnew=fd(r+1) 20
fprey=fprey(r) 21
fd=fd(r) 22

Substitute Eqs. (19), (20), (21), and (22) in Eq. (18), then the expression is

fd(r+1)=fprey(r)+D×ω×R×fprey(r)+D×k3×μ×Cos(2πk4)×1-Cos(2πk5×fprey(r)-fd(r) 23
fd(r+1)=fprey(r)+D×ω×R×fprey(r)+D×k3×μ×Cos(2πk4)×1-Cos(2πk5×fprey(r)-D×k3×μ×Cos(2πk4)×1-Cos(2πk5fd(r) 24
fd(r+1)=fprey(r)1+D×ω×R+D×k3×μ×Cos(2πk4)×1-Cos(2πk5-D×k3×μ×Cos(2πk4)×1-Cos(2πk5fd(r) 25

From EWMA [25],

fdE(r)=σfd(r)+1-σfdE(r-1) 26
fd(r)=1σfdE(r)-1-σfdE(r-1) 27

Applying Eq. (27) in Eq. (25), then the equation is

fd(r+1)=fprey(r)1+D×ω×R+D×k3×μ×Cos(2πk4)×1-Cos(2πk5-D×k3×μ×Cos(2πk4)×1-Cos(2πk51σfdE(r)-1-σfdE(r-1) 28

where ω1, k4, k4, k5, and k7 denote the random number between 0 and 1, fdE(r) shows the location of the honey badger using EWMA at rth iteration, ω1 indicates the honey badger’s capability to catch the food, and fdE(r-1) indicates the location of honey badger using EWMA at r-1th iteration.

D=1ifk70.5-1else 29

While performing the digging phase, the badger accepts the disturbance D, which permits the badger to evaluate the better location of prey.

Honey phase: Here, the honey badger obeys the rules of the honey bird for approaching the beehive, and this is articulated as

fnew=fprey+D×k7×μ×sd 30

where

sd=fprey-fd 31

Applying Eq. (31) in (30), then the expression can be portrayed as

fnew=fprey+D×k7×μ×fprey-fd 32
fnew=fprey1+D×k7×μ-D×k7×μ×fd 33

Apply Eqs. (20), (21), and (22) on Eq. (33), then the expression is changed to

fd(r+1)=fprey(r)1+D×k7×μ-D×k7×μ×fd(r) 34

Apply Eq. (27) in Eq. (34), then it is modified as

fd(r+1)=fprey(r)1+D×k7×μ-D×k7×μ×1σfdE(r)-1-σfdE(r-1) 35

where

μ=B×exprrmax 36

where rmax specifies the highest iteration count, B indicates the constant, which is greater than 1, sd indicates the distance between prey and dth honey badger, and fprey(r) denotes the location of prey.

  • f)

    Evaluation of reliability: In this method, the best solution is acquired by Eq. (8), and the minimized value of the fitness function is portrayed as an optimal solution.

  • g)

    Termination: The aforesaid processes are achieved repeatedly till the highest solution is achieved.

In this case, the established EHBO algorithm is used to transport information among the nodes via the most efficient route. Nodes are chosen in this case depending on the fitness function.

BC Detection at the BS

The BS runs the BC detection system, which is an EHBO-based DCNN. Here, ROI extraction is used for pre-processing, and feature extraction is used to isolate the important features, such as statistical features and texture features. The processes, such as rotation, flipping, cropping, and zooming are utilized to perform the data augmentation. After that, the BC classification is performed using DCNN, where the weights and bias are tuned with devised EHBO algorithm.

Let us assume the dataset W with d images, which are mathematically portrayed as

W=W1,W2,...,Wz,...,Wd 37

where Wz indicates the zth images and d indicates the overall image count. Wz is an input of RoI extraction.

Pre-processing

The pre-processed image speeds up the classifier’s classification performance while reducing computing time. ROI extraction refers to the process of extracting or filtering the specified region from images. It eliminates the noise that exists in the image. Moreover, the outcome of RoI extraction is represented as XR, which is processed under segmentation.

Segmentation

The input of BC segmentation is XR, and the segmentation is done with U-net. In this step, the segmented regions are separated from XR. The advantage of the U-net model [27] is that it has less training time and the efficiency of segmentation is maximal. The structure of the U-net is designated in the next section.

U-net Structure

U-net architecture [27] is formed as in U-shape so that it is called as U-net model. The U-net architecture contains two paths, namely, contrasting path and expansive path. The contrasting path is placed on the left lobes, whereas the expansive path is placed on the right lobes. Here, the contrasting path comprises two conv layers, pooling layer, and ReLU with 2 strides for performing downsampling. In every step of downsampling, the feature channel counts are twined. In expansive path, each step comprises upsampling-dependent feature map with conv layers and ReLU layers, which reduces the feature channel to be halved. This method achieved the loss of border pixel, such that the process of cropping is performed at every convolution. Lastly, the conv layer is exploited to map all the feature vectors to the requisite class count. The BC segmented region gotten from the U-net model is characterized as Ms. Figure 4 depicts the design of the U-net model.

Fig. 4.

Fig. 4

U-net architecture

Feature Extraction

The process of feature excavation is considered a technique for extracting the feature from a segmented image Ms. The dimension of the image is large; hence, it contains multiple numbers of features. Nonetheless, all the features do not have huge information about the image. Thus, only the texture features with huge information about the image is extracted in this step.

Texture feature extraction: The texture features, such as LGBP, GLCM, and LOOP, are excavated from the pre-processed image. After extracting the texture features, the extracted features are processed under data augmentation for augmenting the accuracy of classification. Then, the statistical features are excerpted from the augmented image. The mathematical expansion of LGBP, GLCM, and LOOP features are explained below.

LGBP: LGBP operator [20] is attained by encoding the magnitude values of the LBP operator. Let us assume the center value as hm and the neighboring pixel as hp, then the computation of the LGBP operator is expressed as

hp-hm=1,hphm0,hp<hm 38

Moreover, the LBP pattern of the pixel value is computed by assigning the binomial factor p for every hp-hm and is expressed as

LBP=p=07hp-hm2p 39

Equation (39) signifies the spatial pattern of local image texture. Moreover, the LGBP feature is indicated as n1.

GLCM feature: GLCM [21] is a statistical method that analyzes the texture feature from an image based on the spatial relationship among pixels. GLCM is in the form of a matrix where the rows and columns of the matrix are similar to the gray level count. In this paper, numerous features are excerpted from GLCM, which is mathematically indicated asau(l)=t=0C-1Al,t and 

av(t)=l=0C-1Al,t 40

αu=l=0G-1lAu(l) and

αv=t=0G-1tav(t) 41
βu2=l=0G-1Au(l)-αu(l)2 42
βv2=t=0G-1Av(t)-αv(t)2 43

where C depicts the gray level count, α signifies the mean value of A, αu, αv, βu, and βv demonstrates the mean and standard deviation of Au and Av. Thus, the GLCM feature is represented as n2.

LOOP features: The rotation invariance is encoded into the leading formulation by the LOOP descriptor [22]. The LOOP algorithm also bargains the empirical conveyance of extraneous parameters. Thus, the LOOP value of the pixel ud,vd is portrayed as

LOOPud,vd=p=17Z(ep-em).2rp 44
Z(u)=1ifu00Otherwise 45

where ep and em are the pixel intensities and Z(u) be the Kirsch function. The LOOP feature is represented as n3. Thus, the final texture-based feature vector is formed by joining all the texture features, which is formulated as

N=n1,n2,n3 46

where n1 specifies the LGBP feature, n2 specifies the GLCM feature, and n3 denotes the LOOP feature. In this step, the texture features are only extracted. Then, the segmented image is then expanded using data augmentation. Therefore, the texture features are extracted before augmentation. Further, the statistical features are extracted after the data augmentation.

Data Augmentation

Before classifying the BC, data augmentation [28] is necessary to improve the classification. The extracted texture features, such as n1, n2, and n3 are presented to the data augmentation. Here, the processes, such as rotation, flipping, cropping, and zooming are performed separately on each feature, thereby a total of 12 images are attained in the data augmentation process. Moreover, a brief description of the data augmentation process is given below.

Rotation: This step shifts the image to a specified degree around the center with mapping of each pixel of an image.

Flipping: The process of flipping is performed across the vertical axis, which is computationally effective and easy to execute.

Cropping: This step crops or extracts the significant features from the corner or center of an image.

Zooming: This process is used to zoom-in or zoom-out the processed image.

Thus, the data-augmented outcome is portrayed as DA.

Statistical feature extraction: The statistical features [29], like area, energy, homogeneity, mean, variance, entropy, kurtosis, skewness, and HOG, are extracted from the augmented outcome DA. Here, the statistical features are computed for every augmented image. Moreover, the expression for statistical features is portrayed as

  • i)

    Area: Area is one of the significant statistical features, which is expressed as the number of pixels in a boundary surrounded by the affected area in the segmented image quantified as area. Thus, the notation for the area is y1 with dimension 1×1.

  • ii)

    Energy: Based on energy, also known as the angular second moment, the mathematical form of energy is calculated and is written as

Energy=u,v=0A(u,v)2 47

where A(u,v) depicts the area of the pixel location. Thus, the energy output is denoted as y2 with size 1×1.

  • iii)

    Homogeneity: Homogeneity refers to the local information, which portrays the regularity of computed area and is expressed as

Homogeneity=u,v=0A(u,v)1+u-v 48

Thus, the homogeneity outcome is y3 with dimension 1×1.

  • iv)

    Mean: The term mean refers to the average intensity values of every pixel and is demonstrated as

Mean=u=1nVun 49

where Vu indicates the sum of all pixel values. Moreover, the output for the mean is y4 with dimension 1×1.

  • xxii)

    Variance: The term variance refers to the dissimilarities of intensities among the pixels and is indicated as

Variance=u=0N-1α(u)-α22N 50

Here, the term y5 is the output of variance with dimension 1×1.

  • f)

    Entropy: The term entropy is used to express the texture of an image, and the expression becomes

Entropy=-u=0N-1v=0N-1A(u,v)log2A(u) 51

The entropy output is denoted as y6 with dimension 1×1.

  • vii)

    Kurtosis: It characterizes the data distribution of the peak of the signal created as compared with normal distribution and is indicated as

Kurtosis=κ4ς4 52

where κ4 depicts the fourth moment of mean, ς indicates the standard deviation. Here, the kurtosis is indicated as y7 with dimension is 1×1.

  • viii)

    Skewness: Skewness is utilized to compute the abnormality of information around the average of the sample and is illustrated as

skewness=u=1nVu-V¯n-13 53

where the skewness is indicated as y8 with dimension 1×1.

  • i)

    HOG: HOG feature is a descriptor, which is calculated by partitioning the region of the frame into cells. For every cell, a local 1D histogram of the gradient is applied on every pixel of a cell. The HOG feature is represented as y9.

Thus, the final vector is generated by joining all the statistical features and is expressed as

F=y1,......,y9 54

where y1 denotes the area feature, y2 signifies the energy, y3 denotes the homogeneity, y4 signifies the mean, y5 portrays the variance, y6 specifies the entropy, y7 denotes the kurtosis, y8 indicates the skewness, and y9 specifies the HOG. Following the generation of statistical characteristics, Deep CNN was trained using a hybrid optimization algorithm to classify BC.

BC Classification

The extracted feature F is fed to the input of Deep CNN to categorize the image into either normal or abnormal. Deep CNN [24] is a DL classifier, which is widely utilized for classification purposes. This method has multiple conv layers, which enhances the prediction accuracy.

Structure of Deep CNN

Deep CNN [24] contains three layers, such as convolutional layer, FC layer, and pooling layer; each of these layers performs a separate operation. The conv layer performs the feature extraction process, pool layer performs the sub-sampling process, and the FC layer performs the classification process. Deep CNN contains numerous quantity of conv layers, which progresses the detection efficiency. The structure of Deep CNN is portrayed in Fig. 5 in which its input vector is indicated as F.

Fig. 5.

Fig. 5

Structure of Deep CNN

Conv layer: The conv layer contains the multiple counts of succeeding layers and is the most important layer in Deep CNN. Additionally, the conv layer count affects how accurately predictions are made. Utilizing accessible fields, the conv layers’ neurons are coupled to the trainable weights.

Cbqw,z=xbqw,z+o=1g1b-1i=-j1bj1bc=-j2bj2brq,obi,cCbq-1w+i,z+c 55

where Cbqw,z characterizes the feature map of bth conv layers, xbq shows the bias at bth conv layers, and rq,obi,c and Cbq-1w+i,z+c describe the kernel function in bth conv layers.

ReLU layer: ReLU is considered as an activation function, which is used to eliminate the negative values to progress the effectiveness. Moreover, the outcome of the ReLU layer is portrayed as

Cbq=fun(Czq-1) 56

where fun() indicates the activation function of the conv layer b.

Pooling layers: Pooling layers perform the sub-sampling process by minimizing the feature maps for improving the invariance features to the input data distortions. In this layer, there is no bias or weight as it is a non-parameterized layer.

FC layer: The FC layer is used to perform the classification or prediction process. In this paper, the FC layer is used to classify the BC disease, and this layer is portrayed as

Tbq=ρLbqwithLbq=o=1g1b-1i=-j1bj1bc=-j2bj2brq,obi,cCbq-1w+i,z+c 57

Hence, the output attained by DCNN is articulated as Sx.

Training of Deep CNN Using EHBO Algorithm

This sector denotes the Deep CNN training procedure employing the created EHBO algorithm. In addition, the fitness function of the EHBO for training the Deep CNN is MSE, which is portrayed as

Fitness=1cx=1cSx-Sx2 58

where c indicates the overall sample count, Sx indicates the expected outcome, and Sx specifies the predicted result of Deep CNN.

Results and Discussion

This section describes the outcomes and a discussion of the new EHBO algorithm for BC classification.

Experimental Setup

The testing of devised EHBO algorithm is carried out on MATLAB tool, intel i3 core processor with Windows 10 OS. Table 1 displays the experimental setup of the implemented system.

Table 1.

Experimental setup

Parameters Values
Epoch 20
Batch size 32
Learning rate 0.001
Optimization iteration 100
Upper bound 30
Lower bound 0

Dataset Description

The newly devised EHBO algorithm is implemented using the MIAS and the DDSM dataset [28].

MIAS dataset: This dataset comprises 322 digitized films, which persist on 2.3 GB 8-mm tape. It also contains the “truth” of the radiologist and the location of any irregularities that persist in the image.

DDSM dataset: DDSM is another resource of mammographic images for the research community. Here, the dataset contains two images of both breasts and their correlated patient information. Moreover, the images comprise affected areas with information about localities and types of suspicious areas.

Evaluation Metrics

The evaluation metrics are described as follows:

  • i)

     Testing accuracy: The testing accuracy is expressed as the ratio of the number of appropriately classified samples to the overall tested samples. The testing accuracy formula is depicted below:

A=Hp+InHp+Hn+Ip+In 59

where Hp shows the true positive, Hn denotes the true negative, Ip represents false positive, and In means false negative.

  • b)

    Sensitivity: The sensitivity is defined as the probability of producing a positive result and is expressed as

E=HpHp+In 60
  • iii)

    Specificity: Specificity is defined as the probability of producing a negative result and is illustrated as

K=HnHn+Ip 61

Observational Outcome

The observational outcome of the EHBO algorithm is specified in Fig. 6. Here, the input image is given in Fig. 6a, the segmented image is specified in Fig. 6b, the outcome of the HOG feature is portrayed in Fig. 6c, the extracted LGBP is explained in Fig. 6d, the extracted LOOP features are given in Fig. 6e, the output image of flipped features are given in Fig. 6f, the outcome of rotated image is given in Fig. 6g, the outcome of zoomed image is given in Fig. 6h, and the outcome of cropped image is given in Fig. 6i.

Fig. 6.

Fig. 6

Experimental outcome. a Input image, b segmented image, c HOG feature, d LGBP feature, e LOOP feature, f flipped image, g rotated image, h zoomed image, and i cropped image

The screenshots of the experiment performed is specified in Fig. 7.

Fig. 7.

Fig. 7

Screenshot of the experiment performed

Comparative Methods

The effectiveness of the devised algorithm is analyzed with other alternative CN classification algorithms, such as MLP classifier [13], deep learning [8], SVM [16], and ensemble-based classifier [6].

For routing, the various algorithms, such as MACO_QCR protocol [30], hierarchical cluster-based routing protocol [31], energy efficient routing protocol [32], and chain cluster-based routing [33] are utilized to compare the routing performance with EHBO-based routing.

The description of comparative methods is depicted in Table 2.

Table 2.

Description of comparative methods

Method Description
Classification methods
Multi-layer perceptron (MLP) classifier [13] In this research, the BC is identified in its early stage, and the prediction performance was improved using MLP classifier, by reducing the Relative Absolute Error rate.
Deep learning [8] The CNN-based deep model was used for BC detection using histopathology images. Here, the graphics processing unit was used to train the deep learning model to attain better outcomes.
Support vector machine (SVM) [16] Here, the best features were selected using recursive feature selection, and based on that, the classification was done by SVM.
Ensemble-based classifier [6] In this approach, the malignant cells were detected based on extracting textured-based features and shapes, and the classification was done by the ensemble-based classifier.
Routing methods
Multi-objective ant-colony-optimization-based QoS-aware cross-layer routing (MACO_QCR) protocol [30] This protocol was proposed to do the communication in WSN. Here, the fuzzy membership function was used as the fitness function.
Hierarchical cluster-based routing protocol [31] In this approach, the hierarchical cluster-based routing protocol was proposed for reducing the consumption of energy and increasing the lifetime of the network. Here, the optimal solution was identified using the convex function.
Energy efficient routing protocol [32] In this research, an energy-efficient protocol, named, ring routing was developed for reducing the overhead in WSN.
Chain cluster-based routing [33] This method had the advantages of both LEACH and PEGASIS. It reduced the energy consumption and network delay.

Comparative Assessment

The comparative evaluation of devised EHBO-based DCNN is analyzed by modifying the training data % utilizing MIAS and DDSM datasets.

Comparative Assessment Using DDSM

The achieved values of sensitivity, accuracy, and specificity according to the devised scheme are given in Fig. 8. Figure 8a shows the accuracy analysis. Here, the accuracy of 0.9051 is attained by the devised scheme, whereas the accuracy of 0.8396, 0.8451, 0.8563, and 0.8738 is gotten by the other conventional BC classification techniques, like MLP classifier, deep learning, SVM, and ensemble-based classifier for the training data is 90%. Moreover, the improved percentages of projected approach with conventional techniques are 7.23%, 6.62%, 5.39%, and 3.45%. Figure 8b demonstrates the sensitivity of the projected BC classification technique by altering the training data from 50 to 90%. When the training data is maximum, then the sensitivity is 0.8971 for the new EHBO-based DCNN, and the sensitivity is 0.8268 for the MLP classifier, 0.8303 for deep learning, 0.8415 for SVM, and 0.8797 for the ensemble-based classifier. In addition, the devised EHBO algorithm acquired improved performance with the conventional scheme in terms of percentages 7.83%, 7.44%, 6.19%, and 1.93%. The specificity analysis is given in Fig. 8c. The specificity of EHBO-based DCNN is 0.9029, MLP classifier is 0.822, deep learning is 0.8417, SVM is 0.8526, and ensemble classifier is 0.8669, while 90% of training data is processed. The performance improvement of EHBO-based DCNN is 8.96%, 6.77%, 5.57%, and 3.98%. Figure 8d demonstrates the ROC analysis of EHBO-based BC classification. If the FPR value is 90, then the TPR of EHBO-based BC classification is 0.9365, whereas the conventional approaches are 0.8975, 0.8803, 0.8990, and 0.8616.

Fig. 8.

Fig. 8

Comparative analysis of EHBO-based DCNN in terms of a testing accuracy, b sensitivity, c specificity, and d ROC analysis

The observational outcome of the projected scheme based on energy and distance is demonstrated in Fig. 9. The energy values of the developed technique using the DDSM dataset are shown in Fig. 9a. The MACO QCR protocol, the hierarchical cluster-based routing protocol, the energy efficient routing protocol, and the chain cluster-based routing all acquired energy values of 0.101 J, 0.097 J, 0.102 J, and 0.107 J, respectively, when the round is 501, while the newly developed EHBO-based routing is 0.112. Figure 9b demonstrates the distance value measured by the various BC classification techniques. When the number of rounds is 401, then the existing and devised BC classification technique measured the distance of 39.30 km, 36.65 km, 38.02 km, 42.04 km, and 31.65 km, respectively.

Fig. 9.

Fig. 9

Comparative assessment based on a energy and b distance

Comparative Assessment Using MIAS

The measured values of sensitivity, accuracy, and specificity according to the devised scheme are given in Fig. 10. Figure 10a illustrates the testing accuracy of the projected BC classification technique by adjusting the training data from 50 to 90%. When the training data is 90%, then the testing accuracy is 0.9015 for the EHBO-based DCNN and the testing accuracy is 0.8271 for the MLP classifier, 0.8326 for deep learning, 0.8525 for SVM, and 0.8813 for the ensemble-based classifier. In addition, the devised EHBO algorithm acquired improved performance with the conventional scheme in terms of percentages 8.25%, 7.64%, 5.43%, and 2.24%. Figure 10b shows the EHBO-based DCNN’s sensitivity. While 90% of the training data is processed, the sensitivity of the EHBO-based DCNN is 0.9058, the MLP classifier is 0.8314, deep learning is 0.8369, SVM is 0.8568, and ensemble classifier is 0.8856. In addition, the improved percentage of EHBO-based DCNN is 8.21%, 7.60%, 5.40%, and 2.23%. Figure 10c shows the specificity analysis. Here, for the training data is 90%, the specificity attained by the devised scheme is 0.9029, whereas the specificity is 0.822, 0.8417, 0.8526, and 0.8669 for the techniques, like MLP classifier, deep learning, SVM, and ensemble-based classifier. Figure 10d demonstrates the ROC analysis. For 90% FPR, the TPR of EHBO-based BC classification is 0.9532, whereas the MLP classifier, deep learning, SVM, and ensemble-based classifier have the TPR of 0.9084, 0.8803, 0.8990, and 0.8634, respectively.

Fig. 10.

Fig. 10

Comparative analysis of EHBO-based DCNN in terms of a testing accuracy, b sensitivity, c specificity, and d ROC analysis

The outcome of the projected scheme based on energy and distance is shown in Fig. 11. Figure 11a portrays the energy values attained by the devised scheme. For round 501, then the existing routing methods acquired the energy values of 0.101 J, 0.097 J, 0.102 J, and 0.107 J, whereas the projected EHBO-based routing attained the energy of 0.111 J. Figure 11b provides the distance value recorded by the several BC classification approaches. As the round is 401, then the existing and devised BC classification technique achieved the distance of 38.94 km, 36.89 km, 33.70 km, 36.92 km, and 31.36 km, correspondingly.

Fig. 11.

Fig. 11

Comparative assessment based on a energy and b distance

Comparative Discussion

Table 3 displays a comparison of the developed EHBO-based DCNN method for classifying BC. In this study, two datasets, the DDSM and MIAS datasets, are used to classify BC. From Table 2, the projected EHBO-based DCNN attained a better performance correlating to the DDSM dataset. The recorded values of testing accuracy, sensitivity, and specificity according to the devised EHBO-based DCNN are 0.9051, 0.8971, and 0.9029, correspondingly. Similarly, the test accuracy of the MLP classifier, deep learning, SVM, and ensemble-based classifier is 0.8396, 0.8451, 0.8563, and 0.8738, sensitivity of 0.8268, 0.8303, 0.8415, and 0.8797, and the specificity is 0.822, 0.8417, 0.8526, and 0.8669, correspondingly.

Table 3.

Comparative discussion

Database DDSM MIAS
Testing accuracy Sensitivity Specificity Testing accuracy Sensitivity Specificity
MLP classifier 0.8396 0.8268 0.822 0.8271 0.8314 0.8327
Deep learning 0.8451 0.8303 0.8417 0.8326 0.8369 0.8382
SVM 0.8563 0.8415 0.8526 0.8525 0.8568 0.8581
Ensemble-based classifier 0.8738 0.8797 0.8669 0.8813 0.8856 0.8869
Proposed EHBO-based DCNN 0.9051 0.8971 0.9029 0.9015 0.9058 0.9071

Due to the extraction of useful features, the developed approach performed better in this study than the traditional techniques. Generally, the segmented region contains excess features, but all those features are not significant. Thus, the developed technique extracts only the significant features, such as area, energy, homogeneity, mean, variance, entropy, kurtosis, skewness, and HOG, and the texture features, which improves the efficacy of predicted outcomes in terms of evaluation metrics.

Conclusion

This paper presents the new EHBO-based DL algorithm for BC classification using MIAS and DDSM datasets. The devised method attained improved classification accuracy and high processing speed, due to the integral use of effective segmentation and feature extraction. The tumor region from the processed image is isolated by the U-net model. From the segmented region, the significant features, like area, energy, homogeneity, mean, variance, entropy, kurtosis, skewness, and HOG, and the texture features have been extracted. Moreover, the devised scheme gathers the record of every referred patient to attain better classification accuracy. The gathered information is transmitted to the BS using devised EHBO algorithm for classifying the BC, and the classification process is done with DCNN. Here, the EHBO is a combination of the EWMA and HBO algorithms, which perform two functions, such as routing and tuning the weights of DCNN. The observational outcome reveals that the projected EHBO-based DCNN algorithm attained outstanding performance based on the test accuracy, sensitivity, and specificity of 0.9051, 0.8971, and 0.9029, correspondingly. To further improve the BC diagnosis system’s performance, deep learning classification methods, optimization, and other characteristic selection algorithms will be applied in the future.

Acknowledgements

I would like to express my very great appreciation to the co-authors of this manuscript for their valuable and constructive suggestions during the planning and development of this research work.

Nomenclature

BC

Breast cancer

EHBO

Exponential Honey Badger Optimization

DCNN

Deep Covolutional Neural Network

IoT

Internet of Things

HBO

Honey Badger Optimization

EWMA

Exponential Weighted Moving Average

BS

Base station

HOG

Histogram of oriented gradients

LGBP

Local Gabor Binary Pattern

GLCM

Gray-Level Co-occurrence Matrix

LOOP

Local Optimal Oriented Pattern

MLP

Multi-layer perceptron

DL

Deep learning

SVM

Support vector machine

WDBC

Wisconsin Diagnostic BC

WPBCC

Wisconsin Prognostic BC Chemotherapy

ML

Machine learning

OKM-ANFIS

Optimized K-means-based adaptive neuro-fuzzy inference system

CH

Cluster head

conv

Convolutional

ReLU

Rectified linear unit

FC

Fully connected

MIAS

Mammographic Image Analysis Society

CNN

Convolutional Neural Network

DDSM

Digital Database for Screening Mammography

IMS

Intelligent Monitoring System

MACO_QCR

Multi-objective ant-colony-optimization based QoS-aware cross-layer routing

PEGASIS

Power efficient gathering in sensor information systems

WSN

Wireless Sensor Network

LEACH

Low-energy adaptive clustering hierarchy

Author Contributions

All authors have made substantial contributions to conception and design, revising the manuscript, and the final approval of the version to be published.

Data Availability

The data underlying this article are available in MIAS and DDSM datasets, at https://www.mammoimage.org/databases/.

Declarations

Ethical Approval

Not applicable.

Informed Consent

Not applicable.

Conflict of Interest

The authors declare no competing interests.

Disclaimer

Also, all authors agreed to be accountable for all aspects of the work in ensuring that questions related to the accuracy or integrity of any part of the work are appropriately investigated and resolved.

Footnotes

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

References

  • 1.Al-Turjman F, Alturjman S. Context-sensitive access in industrial internet of things (IIoT) healthcare applications. IEEE Transactions on Industrial Informatics. 2018;14(6):2736–2744. doi: 10.1109/TII.2018.2808190. [DOI] [Google Scholar]
  • 2.Deebak BD, Al-Turjman F, Aloqaily M, Alfandi O. An authentic-based privacy preservation protocol for smart e-healthcare systems in IoT. IEEE Access. 2019;7:135632–135649. doi: 10.1109/ACCESS.2019.2941575. [DOI] [Google Scholar]
  • 3.Al-Turjman F, Zahmatkesh H, Mostarda L. Quantifying uncertainty in internet of medical things and big-data services using intelligence and deep learning. IEEE Access. 2019;7:115749–115759. doi: 10.1109/ACCESS.2019.2931637. [DOI] [Google Scholar]
  • 4.Manogaran G, Lopez D, Chilamkurti N. In-Mapper combiner based MapReduce algorithm for processing of big climate data. Future Generation Computer Systems. 2018;86:433–445. doi: 10.1016/j.future.2018.02.048. [DOI] [Google Scholar]
  • 5.Kharya S, Soni S. Weighted naive bayes classifier: a predictive model for breast cancer detection. International Journal of Computer Applications. 2016;133(9):32–37. doi: 10.5120/ijca2016908023. [DOI] [Google Scholar]
  • 6.Khan SU, Islam N, Jan Z, Din IU, Khan A, Faheem Y. An e-Health care services framework for the detection and classification of breast cancer in breast cytology images as an IoMT application. Future Generation Computer Systems. 2019;98:286–296. doi: 10.1016/j.future.2019.01.033. [DOI] [Google Scholar]
  • 7.Anand P, Kunnumakara AB, Sundaram C, Harikumar KB, Tharakan ST, Lai OS, Sung B, Aggarwal BB. Cancer is a preventable disease that requires major lifestyle changes. Pharmaceutical research. 2008;25(9):2097–2116. doi: 10.1007/s11095-008-9661-9. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.Zahir S, Amir A, Zahri NAH, Ang WC. Applying the deep learning model on an IoT board for breast cancer detection based on histopathological images. In Journal of Physics: Conference Series. 2021;1755(1):012026. [Google Scholar]
  • 9.Khamparia A, Bharati S, Podder P, Gupta D, Khanna A, Phung TK, Thanh DN. Diagnosis of breast cancer based on modern mammography using hybrid transfer learning. Multidimensional systems and signal processing. 2021;32(2):747–765. doi: 10.1007/s11045-020-00756-7. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10.Barracliffe, L., Arandjelovic, O. and Humphris, G., “A pilot study of breast cancer patients: can machine learning predict healthcare professionals’ responses to patient emotions”, In Proceedings of the International Conference on Bioinformatics and Computational Biology, Honolulu, HI, USA, pp. 20-22, 2017.
  • 11.Alzubi JA, Manikandan R, Alzubi OA, Qiqieh I, Rahim R, Gupta D, Khanna A. Hashed Needham Schroeder industrial IoT based cost optimized deep secured data transmission in cloud. Measurement. 2019;150:107077. doi: 10.1016/j.measurement.2019.107077. [DOI] [Google Scholar]
  • 12.Sharma, A., Kulshrestha, S. and Daniel, S., “Machine learning approaches for breast cancer diagnosis and prognosis”, In 2017 International Conference on Soft Computing and Its Engineering Applications (icSoftComp), pp. 1-5, 2017.
  • 13.Gopal VN, Al-Turjman F, Kumar R, Anand L, Rajesh M. Feature selection and classification in breast cancer prediction using IoT and machine learning. Measurement. 2021;178:109442. doi: 10.1016/j.measurement.2021.109442. [DOI] [Google Scholar]
  • 14.Savitha V, Karthikeyan N, Karthik S, Sabitha R. A distributed key authentication and OKM-ANFIS scheme based breast cancer prediction system in the IoT environment. Journal of Ambient Intelligence and Humanized Computing. 2021;12(2):1757–1769. doi: 10.1007/s12652-020-02249-8. [DOI] [Google Scholar]
  • 15.Chokka, A. and Rani, K.S., “AdaBoost with feature selection using IoT to bring the paths for somatic mutations evaluation in cancer”, In Internet of Things and Personalized Healthcare Systems, pp. 51-63, 2019.
  • 16.Memon, M.H., Li, J.P., Haq, A.U., Memon, M.H. and Zhou, W., “Breast cancer detection in the IOT health environment using modified recursive feature selection”, wireless communications and mobile computing, 2019.
  • 17.Suresh A, Udendhran R, Balamurgan M, Varatharajan R. A novel internet of things framework integrated with real time monitoring for intelligent healthcare environment. Journal of medical systems. 2019;43(6):1–10. doi: 10.1007/s10916-019-1302-9. [DOI] [PubMed] [Google Scholar]
  • 18.Nguyen, T.D., Khan, J.Y. and Ngo, D.T., “An effective energy-harvesting-aware routing algorithm for WSN-based IoT applications”, In proceedings of 2017 IEEE International Conference on Communications (ICC), pp. 1-6, 2017.
  • 19.Yadav AK, Tripathi S. QMRPRNS: design of QoS multicast routing protocol using reliable node selection scheme for MANETs. Peer-to-Peer Networking and Applications. 2017;10(4):897–909. doi: 10.1007/s12083-016-0441-8. [DOI] [Google Scholar]
  • 20.Zhang, W., Shan, S., Gao, W., Chen, X. and Zhang, H., “Local gabor binary pattern histogram sequence (lgbphs): a novel non-statistical model for face representation and recognition”, In Tenth IEEE International Conference on Computer Vision (ICCV'05) Volume 1, Vol. 1, pp. 786-791, 2005.
  • 21.Zulpe N, Pawar V. GLCM textural features for brain tumor classification. International Journal of Computer Science Issues (IJCSI) 2012;9(3):354. [Google Scholar]
  • 22.Chakraborti, T., McCane, B., Mills, S. and Pal, U., “LOOP descriptor: encoding repeated local patterns for fine-grained visual identification of Lepidoptera”, Computer Vision and Pattern Recognition, 2017.
  • 23.Lakshmi, N.D., Latha, Y.M. and Damodaram, A., “Silhouette extraction of a human body based on fusion of HOG and graph-cut segmentation in dynamic backgrounds”, 2013.
  • 24.Sugave S, Jagdale B. Monarch-EWA: Monarch-earthworm-based secure routing protocol in IoT. The Computer Journal. 2020;63(6):817–831. doi: 10.1093/comjnl/bxz135. [DOI] [Google Scholar]
  • 25.Saccucci MS, Amin RW, Lucas JM. Exponentially weighted moving average control schemes with variable sampling intervals. Communications in Statistics-simulation and Computation. 1992;21(3):627–657. doi: 10.1080/03610919208813040. [DOI] [Google Scholar]
  • 26.Hashim FA, Houssein EH, Hussain K, Mabrouk MS, Al-Atabany W. Honey Badger Algorithm: new metaheuristic algorithm for solving optimization problems. Mathematics and Computers in Simulation. 2022;192:84–110. doi: 10.1016/j.matcom.2021.08.013. [DOI] [Google Scholar]
  • 27.Ronneberger, O., Fischer, P. and Brox, T., “U-net: convolutional networks for biomedical image segmentation”, In Proceedings of International Conference on Medical Image Computing and Computer-Assisted Intervention, pp. 234-241, 2015.
  • 28.MIAS and DDSM dataset will be gathered from, https://www.mammoimage.org/databases/, accessed on January 2022.
  • 29.L. Malliga, “A novel statistical based methodology for the feature extraction of both MRI and CT images”, International Journal of Engineering and Advanced Technology (IJEAT), Volume-8, Issue-6S3, 2019.
  • 30.Kaur T, Kumar D. MACO-QCR: multi-objective ACO-based QoS-aware cross-layer routing protocols in WSN. IEEE Sensors Journal. 2020;21(5):6775–6783. doi: 10.1109/JSEN.2020.3038241. [DOI] [Google Scholar]
  • 31.Ke W, Yangrui O, Hong J, Heli Z, Xi L. Energy aware hierarchical cluster-based routing protocol for WSNs. The Journal of China Universities of Posts and Telecommunications. 2016;23(4):46–52. doi: 10.1016/S1005-8885(16)60044-4. [DOI] [Google Scholar]
  • 32.Tunca C, Isik S, Donmez MY, Ersoy C. Ring routing: an energy-efficient routing protocol for wireless sensor networks with a mobile sink. IEEE Transactions on Mobile Computing. 2014;14(9):1947–1960. doi: 10.1109/TMC.2014.2366776. [DOI] [Google Scholar]
  • 33.Tang, F., You, I., Guo, S., Guo, M. and Ma, Y., “A chain-cluster based routing algorithm for wireless sensor networks”, journal of intelligent manufacturing, vol.23, no.4, pp.1305-1313, 2012.

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Data Availability Statement

The data underlying this article are available in MIAS and DDSM datasets, at https://www.mammoimage.org/databases/.


Articles from Journal of Digital Imaging are provided here courtesy of Springer

RESOURCES