Skip to main content
IEEE - PMC COVID-19 Collection logoLink to IEEE - PMC COVID-19 Collection
. 2020 Sep 30;8:179317–179335. doi: 10.1109/ACCESS.2020.3028012

Novel Feature Selection and Voting Classifier Algorithms for COVID-19 Classification in CT Images

El-Sayed M El-Kenawy 1, Abdelhameed Ibrahim 2,, Seyedali Mirjalili 3,4, Marwa Metwally Eid 1, Sherif E Hussein 2
PMCID: PMC8545288  PMID: 34976558

Abstract

Diagnosis is a critical preventive step in Coronavirus research which has similar manifestations with other types of pneumonia. CT scans and X-rays play an important role in that direction. However, processing chest CT images and using them to accurately diagnose COVID-19 is a computationally expensive task. Machine Learning techniques have the potential to overcome this challenge. This article proposes two optimization algorithms for feature selection and classification of COVID-19. The proposed framework has three cascaded phases. Firstly, the features are extracted from the CT scans using a Convolutional Neural Network (CNN) named AlexNet. Secondly, a proposed features selection algorithm, Guided Whale Optimization Algorithm (Guided WOA) based on Stochastic Fractal Search (SFS), is then applied followed by balancing the selected features. Finally, a proposed voting classifier, Guided WOA based on Particle Swarm Optimization (PSO), aggregates different classifiers’ predictions to choose the most voted class. This increases the chance that individual classifiers, e.g. Support Vector Machine (SVM), Neural Networks (NN), k-Nearest Neighbor (KNN), and Decision Trees (DT), to show significant discrepancies. Two datasets are used to test the proposed model: CT images containing clinical findings of positive COVID-19 and CT images negative COVID-19. The proposed feature selection algorithm (SFS-Guided WOA) is compared with other optimization algorithms widely used in recent literature to validate its efficiency. The proposed voting classifier (PSO-Guided-WOA) achieved AUC (area under the curve) of 0.995 that is superior to other voting classifiers in terms of performance metrics. Wilcoxon rank-sum, ANOVA, and T-test statistical tests are applied to statistically assess the quality of the proposed algorithms as well.

Keywords: COVID-19, CT scans, convolutional neural network, guided whale optimization algorithm, features selection, voting ensemble

I. Introduction

Coronavirus (COVID-19) is a virus infection, named Severe Acute Respiratory Syndrome-Corona Virus-2 (SARS-CoV-2), which appeared in Wuhan toward the end of 2019 [1], [2]. Due to the outbreak, COVID-19 has emerged as a pandemic that threatened human lives and caused devastating economic consequences that arose since that time. Therefore, a significant number of researches were instantiated to discover a solution to control the spread and mortality. Due to COVID-19 implication, many research proposals were conducted to assess the presence and severity of pneumonia caused by COVID-19. Such studies are centered around the screening process to discover early-stage patients, the proposed treatment protocol, and the assessment for various stages and recovery of treated patients. The image modalities including Chest X-ray and Computed Tomography (CT) are non-invasive and are widely used in hospitals to detect both the presence and severity of COVID-19 pneumonia [3], [4]. Compared to CT, even though X-ray is more accessible in hospitals around the world, X-ray images can be considered less sensitive than CT scans for the investigation of COVID-19 patients. [3] reported that X-ray was diagnosed to be normal in both early and mild stages. On the other hand, CT images enable the non-destructive 3D visualization of internal structures and are considered as a powerful analysis tool [5], [6] that has been applied widely to clinical diagnosis [7] and biomedical imaging [8]. In addition, CT has always aimed to achieve improved scanning efficiency in both time and radiation dose [9]. The development of Multi-slice CT (MSCT) has been successful to improve the efficiency of scanning by simultaneously increasing the number of scanned slices [10]. Moreover, dual-source CT managed to achieve a larger temporal resolution improvement, [11].

Machine learning algorithms have been gaining momentum over the last decades for medical applications such as computer-aided diagnosis to help physicians for an early diagnosis, which can lead to better-personalized therapies and enhancement of the medical care offered to patients [12], [13]. Convolutional neural networks (CNN), as a subset of machine learning algorithms, is a unique structure of synthetic neural networks used for image classification. There are several CNN models including AlexNet [14], VGG-Net [15], GoogLeNet [16], and ResNet [17]. In the CNN models, classification accuracy correlates with the extended number of convolution layers [18].

Optimization is the process by which the best possible solution is found for a particular problem from all the available solutions [19]. One of the most powerful methods to solve applications in radiology problems are Meta-heuristic algorithms. The inspiration of most of these algorithms is from physical algorithms’ logical behavior found in nature. The acceptable solutions found these optimization techniques are typically obtained with less computational effort in a reasonable time, [20]. The early diagnosis of coronavirus can significantly limit its wide-spreading and therefore increases the patients’ recovery rates. So, several artificial intelligence (AI) techniques have been proposed for the early detection of COVID-19 in the literature.

In this article, a framework for COVID-19 classification is proposed based on three cascaded phases. The first phase automatically extracts features from the training CT images by a CNN model named AlexNet. Then, a proposed feature selection algorithm, using Stochastic Fractal Search (SFS) and Guided Whale Optimization Algorithm (Guided WOA) techniques, is applied to properly select the valuable features. The LSH-SMOTE (Locality Sensitive Hashing Synthetic Minority Oversampling Technique) is used in the second phase to balance the extracted features. The last phase classifies the selected features by a proposed voting classifier, using Particle Swarm Optimization (PSO) and Guided WOA techniques, by aggregating the Support Vector Machine (SVM) [21], Neural Networks (NN) [22], k-Nearest Neighbor (KNN) [23], and Decision Trees (DT) [24] classifiers to improve the ensemble’s accuracy.

Two kinds of CT datasets are used in the experiments to test the proposed framework. The first dataset has COVID-19 CT images, while the second dataset has extra CT images with clinical cases that have no COVID-19. For feature selection, the proposed (SFS-Guided WOA) algorithm is compared in experiments with binary versions of the original WOA [25], Grey Wolf Optimizer (GWO) [26], Genetic Algorithm (GA) [27], PSO [28], hybrid of PSO and GWO (GWO-PSO) [29], hybrid of GA and GWO (GWO-GA), Bat Algorithm (BA) [30], Biogeography-Based Optimizer (BBO) [31], Multiverse Optimization (MVO) [32], Bowerbird Optimizer (SBO) [33], and Firefly Algorithm (FA) [34] in terms of average error, average select size, average (mean) fitness, best fitness, worst fitness, and standard deviation fitness. Lastly, the proposed voting classifier (PSO-Guided WOA) result of 0.995 is compared with voting WOA, voting GWO, voting GA, and Voting PSO in terms of Area Under The Curve (AUC) and the Mean Square Error (MSE). The main contributions of this article are as follow:

  • A COVID-19 classification framework based on proposed algorithms for feature selection and classification is developed.

  • A novel feature selection algorithm based on SFS and Guided WOA techniques is proposed.

  • A novel voting classifier based on PSO and Guided WOA techniques is proposed.

  • The proposed framework can classify the input CT images to COVID-19 or non-COVID-19 effectively.

  • The proposed framework is evaluated using two datasets of COVID-19 CT images and non-COVID-19 CT images.

  • Statistical tests of Wilcoxon rank-sum, ANOVA, and T-test are carried out to ensure the quality of the proposed algorithms.

  • This framework can be generalized to the applications of biomedical imaging diagnoses.

This article contains the following sections. Related work and the problem definition are discussed in Section II. Section III introduces the materials and methods employed in this research. Section IV presents the model and the proposed algorithms in detail. Section V shows the designed scenarios and results. Section VI discusses the experimental results. The conclusions and future work are shown in Section VII. See Table 1 for a list of abbreviations.

TABLE 1. List of Abbreviations.

Abbreviation Explanation
AI Artificial Intelligence
ANN artificial neural network
AUC Area Under the Curve
BA Bat Algorithm
BBO Biogeography-Based Optimizer
CNN Convolution Neural Network
CT Computed Tomography
DLA Diffusion Limited Aggregation
DT Decision Trees
FA Firefly Algorithm
FS Fractal Search
GA Genetic Algorithm
GWO Grey Wolf Optimizer
KNN k-Nearest Neighbor
LSH-SMOTE Locality Sensitive Hashing SMOTE
MLP Multilayer Perceptron
MSE Mean Square Error
MVO Multiverse Optimization
NN Neural Networks
PCR Polymerase Chain Reaction
PSO Particle Swarm Optimization
ROC Receiver Operating Characteristics
SBO Bowerbird Optimizer
SFS Stochastic Fractal Search
SMOTE Synthetic Minority Over-sampling Technique
SVM Support Vector Machine
WOA Whale Optimization Algorithm

II. Related Work

In this section, the recent literature utilizing the CT scans for diagnosing COVID-19 patients will be summarized. Then, the recent evaluation of Artificial Intelligence (AI) against COVID-19 based on the CT scans will be discussed as well.

A. COVID-19 and CT Scans

Recent study proposed several COVID-19 detection paradigms. In [35], Li et al. proposed a methodology to recognize the infection rate using the coronal and axial view of lung CT scans. The proposed work achieved a specificity of 100%, AUC of 0.918, and sensitivity of 82.6%. Another study by [36] evaluated COVID-19 disease using visual inspection. They claimed that visual inspection can help to correctly identify the infection. In [37], Panwar et al. proposed a scheme to evaluate the lung CT scans and implemented visual inspection-based detection. Their scheme could achieve Specificity of 94%., AUC of 0.892, and Sensitivity of 83.3%. In [2], Wang et al. investigated 90 patients’ lung CT scans. Their investigation managed to detect the severity based on the time since the patient got infected. In [38], in addition a diagnostic methodology was proposed based on the CT scans image features. They concluded that the combination of both image features evaluation and clinical findings can early detect the presence of COVID-19. In [39], Bai et al. investigated the patient’s information and considered the CT scans and RT-PCR for the examination. They achieved a specificity of 100% and a sensitivity of 93%. In a similar study [40], authors clinically evaluated patients with both CT scans and real-time RT-PCR with an early detection accuracy of 90%.

B. Artificial Intelligence for COVID-19

Recent works show that the CT scans are mainly utilized to offer fast diagnostic methods to prevent and control the spread of COVID-19 and assist physicians and radiologists to correctly manage patients in high workload. Authors in [41] developed a method based on deep learning to accurately assist radiologists to identify COVID-19 patients using CT images. They used deep learning to train a neural network to screen COVID-19 patients based on their CT images. The proposed method achieved a specificity of 61.5%, sensitivity of 81.1%, AUC of 0.819, and accuracy of 76%. In [42], Ardakani et al. proposed a method to diagnose COVID-19 using an AI technique based on CT slices and ten convolutional neural network models to correctly diagnose COVID-19 from non-COVID-19 groups. The authors found that both ResNet-101 and Xception have achieved the best performance. Moreover, ResNet-101 managed to detect COVID-19 cases with a specificity of 99.02%, Sensitivity of 100%, AUC of 0.994, and Accuracy of 99.51%. On the other hand, Xception achieved a Specificity of 100%, Sensitivity of 98.04%, AUC of 0.994, and Accuracy of 99.02%. The authors recommended the use of ResNet-101 to characterize and diagnose COVID-19 infections due to its higher sensitivity.

Another study in [43] used a large CT dataset to develop an AI method that can diagnose COVID-19 and differentiate it from normal controls and other types of pneumonia. The authors investigated the significance of identifying important clinical markers using the convolutional neural network ResNet-18 model. Their proposed method achieved a Specificity of 91.13%, Sensitivity of 94.93%, AUC of 0.981, and Accuracy of 92.49% for COVID-19. In [44], the authors proposed a deep learning neural network-based method named nCOVnet for detecting the COVID-19 based on analyzing the patients’ X-ray images. Their nCOVnet method achieved a Specificity of 89.13%, Sensitivity of 97.62%, AUC of 0.881, and Accuracy of 88.10% for COVID-19. Butt et al. [45] used a special type of CNN, namely ResNet-18 to classify CT samples with COVID-19, normal subjects, and Influenza viral pneumonia. They achieved an accuracy of 86.7% with 98.2% sensitivity, 92.2% specificity, and AUC value of 0.996.

Chua et al. [46] proposed a model based on the CNN architecture model that was trained from scratch. Their model consisted of five convolution layers utilized as a deep feature extractor. K-nearest neighbor, SVM, and decision tree were fed using the extracted deep discriminative features. The superiority of the SVM classifier was demonstrated with an accuracy of 98.97%, a sensitivity of 89.39%, and a specificity of 99.75%. Another study by Wu et al. [47] proposed a weakly supervised CNN that could achieve an accuracy of 96.2% with 94.5% sensitivity, 95.3% specificity, and AUC value of 0.970. A ML-method is proposed in [48] to classify the chest x-ray images into COVID-19 or non-COVID-19 patients. A Fractional Multichannel Exponent Moments (FrMEMs) method is used for feature extraction. A modified Manta-Ray Foraging Optimization based on differential evolution is then used to select the most significant features. The authors’ proposed method is evaluated using two COVID-19 x-ray datasets. The recent AI research for COVID-19 is summarized in Table 2.

TABLE 2. Recent AI Research for COVID-19.

Reference Methods # of samples # of classes Type of Images Sensitivity Specificity Accuracy AUC
X. Wu et al. (2020) [41] Multi-view deep learning model (ResNet50 based) 495 2 CT images 81.1% 61.5% 76% 0.819
A. A. Ardakani et al. (2020) [42] Deep learning technique (ResNet-101 based) 1020 2 CT images 100%, 99.02%, 99.51% 0.994
Deep learning technique (Xception based) 1020 2 CT images 98.04%, 100% 99.02% 0.994
K. Zhang et al. (2020) [43] AI system (ResNet-18 based) 3,777 3 CT images 94.93% 91.13% 92.49% 0.981
H. Panwar et al. (2020) [44] nCOVnet, transfer learning, deep CNN 337 2 X-ray images 97.62% 89.13% 88.10% 0.881
C. Butt et al. (2020) [45] Multiple CNN models (ResNet-18 based) 618 3 CT images 98.2% 92.2% 86.7% 0.996
M. Nour et al. (2020) [46] Training CNN model, deep feature extraction, SVM 2,905 3 X-ray images 89.39% 99.75% 98.97%, 0.994
X. Wang et al. (2020) [47] Weakly supervised deep learning framework 450 3 CT images 94.5% 95.3% 96.2% 0.970

The importance of the AI techniques in the early evaluation of COVID-19 and the areas where AI can contribute to the battle against COVID-19 are discussed in [50]. The authors concluded that AI is not fully utilized in COVID-19 because of the possible lack of data or excessive data. To overcome these constraints careful balance must be made between public health, data privacy, and the right utilization of the AI techniques. Furthermore, the need for an extensive gathering of diagnostic data will be extremely crucial to train AI, save lives, and limit the associated economic damages.

Most of the above-discussed studies mainly applied statistical analysis and visual inspection techniques to correctly diagnose COVID-19 infection. A lesser number of applied researches used transfer learning and CNN with CT datasets of coronavirus pneumonia patients, non-corona virus pneumonia patients, and healthy subjects. Therefore, more study needs to be conducted that utilizes AI with properly optimized performance metrics. As per the literature review of this work, it is recommended to use the CT images as a fast method to diagnose patients with COVID-19. The proposed paradigms need to be both reproducible and easily validated to can be quickly integrated into the arsenal of battling the COVID-19 pandemic.

III. Materials and Methods

This section discuss data sets and methodologies of this research. The datasets, dataset balancing, and the optimization methods of WOA, PSO, and SFS are discussed. The CNN models, classification methods, and ensemble learning techniques are also explained.

A. Datasets

Data collection is considered as the first and main step in COVID-19 applications. Recently, it has been reported that several data collection works were done on COVID-19. The authors have used two datasets to apply the proposed paradigm. The first is the COVID-19-dataset which has 334 CT images containing clinical findings of COVID-19. While the second is the non-COVID-19-dataset that has extra 794 CT images with clinical cases that have no COVID-19. Figure 1 shows samples of the COVID-19 and the non-COVID-19 cases. The images are collected from COVID19-related articles from medRxiv, bioRxiv, NEJM, JAMA, and Lancet.CTs containing COVID-19 abnormalities were selected by reading through the papers’ figures captions [49]. All patients’ images in the datasets were high-resolution Multi-Detector Computerized Tomography (MDCT) Axial images. The Axial images show bilateral scattered ground-glass opacities with air space consolidation, mainly posterior segments of lower lung lobes with peripheral and subpleural distribution; the picture of atypical pneumonia caused by COVID-19 that is clinically proved by Polymerase Chain Reaction (PCR). PCR is a process that replicates a small segment of DNA, a large number of times, to create enough samples for analysis.

FIGURE 1.

FIGURE 1.

Original images from the dataset for COVID-19 and Non-COVID-19 cases [49].

B. Dataset Balancing

The extracted features from the utilized datasets may suffer from a class imbalance problem. Therefore, several algorithms were investigated to solve that type of problems. Some of the recent algorithms are the SMOTE and the LSH-SMOTE [51], [52]. The SMOTE technique finds its k-nearest minority class neighbors for a selected minority class instance Inline graphic at random. Then, it randomly chooses another k-nearest neighbor Inline graphic to be connected with Inline graphic to form a line segment in the feature space. Euclidean distance is used to sort the instances while selecting the k-nearest neighbors. Finally, a list of k-nearest neighbor’s instances is returned to the main SMOTE class for generating the synthetic instances. LSH-SMOTE was first introduced by [52] to improve the performance of the feature selection SMOTE based optimization techniques. The algorithm starts with hashing and dividing the dataset into buckets by assigning similar items with similar hash codes to the same bucket. That, in turn, can increase the matching probability between similar items leading to a simplified search for the k-nearest neighbors.

C. Convolutional Neural Network (CNN)

CNN is of the most well-regarded machine learning methods in the literature. One of the reasons of its popularity is due to the automatic hierarchical feature representation in recognizing objects and patters in images [42]. CNNs reduce the parameters of a given problem using spatial relationships between them. This makes them a more practical classifier specially in image processing where we deal with a large number of parameters (pixels), rotation, translation, and scale of images. In fact, CNNs alleviate the drawbacks of Feel Forward Neural networks and Multi-Layer Perceptons by using an alternative to matrix multiplication. We use this powerful method in this study due to the nature of COVID-19 diagnosis from CT images and its high-dimensional nature.

D. Whale Optimization Algorithm

In the WOA algorithm, the inspiration is from the foraging behaviour of whales, in which bubbles are used to trap the prey by forcing them to the surface in a spiral-shaped [25], [53]. Mathematically, the first mechanism by this optimizer is based on the following equation:

D.

where vector Inline graphic represents a solution at iteration Inline graphic and vector Inline graphic represents the position of the prey. the “.” indicates pairwise multiplication and Inline graphic represent the updated position for the solution [54], [55]. The two vectors of Inline graphic and Inline graphic are updated in each iteration by Inline graphic and Inline graphic for vector Inline graphic changes from 2 to 0 linearly and Inline graphic and Inline graphic are random values in [0, 1].

The second mechanism includes a shrinking encircling, which decreases the values of Inline graphic and Inline graphic vectors, and a spiral process for updating the positions as follows

D.

where Inline graphic represents Inline graphicth whales and the best one distance. Parameter Inline graphic is a constant, represents the spiral’s shape, and Inline graphic is a random value in [−1, 1]. The WOA mechanism can be simulated by the following equation

D.

where Inline graphic represents a random value in [0, 1].

The last mechanism can be achieved based on the Inline graphic vector. The position of search agent is updating based on a random whale Inline graphic to allow a global search by the following equation

D.

Thus, the exploitation and exploration are controlled by Inline graphic, and the spiral or circular movement is controlled by Inline graphic. The WOA algorithm is shown step by step in Algorithm 1.

Algorithm 1 Original WOA Pseudo-Code

  • 1:

    Initialize WOA population Inline graphic with size Inline graphic, maximum iterations Inline graphic, fitness function Inline graphic.

  • 2:

    Initialize WOA parameters Inline graphic, Inline graphic, Inline graphic, Inline graphic, Inline graphic, Inline graphic, Inline graphic

  • 3:

    Initialize t as the iteration counter

  • 4:

    Calculate fitness function Inline graphic for each Inline graphic

  • 5:

    Find best individual Inline graphic

  • 6:

    while Inline graphic do

  • 7:

    for (Inline graphic) do

  • 8:

    if (Inline graphic) then

  • 9:

    if (Inline graphic) then

  • 10:

    Update current search agent position using Eq. 1

  • 11:

    else

  • 12:

    Select a random search agent Inline graphic

  • 13:

    Update current search agent position by Eq. 4

  • 14:

    end if

  • 15:

    else

  • 16:

    Update current search agent position by Eq. 2

  • 17:

    end if

  • 18:

    end for

  • 19:

    Update Inline graphic, Inline graphic, Inline graphic, Inline graphic, Inline graphic

  • 20:

    Calculate fitness function Inline graphic for each Inline graphic

  • 21:

    Find best individual Inline graphic

  • 22:

    Set t = t +1. (increase counter).

  • 23:

    end while

  • 24:

    return Inline graphic

E. Stochastic Fractal Search

The Stochastic Fractal Search (SFS) technique was proposed by [56] in which the fractal mathematical concept was used as a property of objects’ self-similarity. The Fractal Search (FS) algorithm depending on the Diffusion Limited Aggregation (DLA)that generates the objects’ fractal-shaped. Figure 2 presents a random fractal sample. The SFS technique uses diffusion and two kinds of updating processes to outperform the original FS technique. Figure 2 shows the diffusion process of the SFS technique in a graphical form for a solution. For the best solution Inline graphic, a list of solutions Inline graphic, and Inline graphic can be listed around this best solution [57].

FIGURE 2.

FIGURE 2.

SFS algorithm; Random fractal sample with diffusion around the best solution.

F. Particle Swarm Optimization

PSO algorithm is based on the swarming pattern of flocks in nature [58], [59]. PSO algorithm simulates an animal’s social behavior such as birds. The swarms searching for food by changing their positions according to the updated velocity. PSO has several particles and each particle has the following parameters:

  • Position (Inline graphic), which indicated a point in Inline graphic search space. The fitness function is used to evaluate the particles’ current positions.

  • Velocity or rate of position change, (Inline graphic),

  • Last best positions (Inline graphic), which store better positions’ values of the particles.

During the algorithm iterations, the positions and velocity of all particles are changing. The particles’ positions are updated as follows:

F.

where Inline graphic is the new particle position, and the updated velocity of each particle Inline graphic can be calculated as

F.

where Inline graphic is the inertia weight, Inline graphic and Inline graphic represent cognition learning factor and the social learning factor. Parameter Inline graphic is the global best position and Inline graphic and Inline graphic are random numbers in [0; 1].

G. Classification Methods

SVM can perform classification, regression, and outlier detection [21]. SVMs are suited for the classification of complex datasets. The classification of the SVM technique is based on transforming the features dimension space that is nonlinearly separable into a higher dimension space in which a hyperplane can easily separate the different classes. That can be done using a kernel trick in which linear, polynomial, or Gaussian RBF kernel can be used to decrease the computational complexity associated with the calculations of added features. The margin between classes depends on dataset instances called support vectors. While the kernel hyperparameters are those parameters that determine the margin of separation between classes and the tolerance for permitting margin violation. Even though SVM is a binary classifier, it can be easily extended to be used in multiclass classification.

KNN method can also be used for classification and regression [23] purposes. As a classifier, this algorithm considers Inline graphic closest training examples in the feature space. The output in this algorithm is a class membership. DT [24] is also a machine learning capable of doing both classification and regression.

MLP is a class of feedforward ANN [22]. There are three layers in MLP: input, hidden, and output layers. Such architecture with three layers is mostly suited to small or medium datasets. In addition, the dataset complexity can be accommodated using suitable activation functions and/or a suitable number of perceptrons in the hidden layers. However, large datasets can be more complex to be accommodated by only three layers of nodes. Therefore, architectures with more than three layers are common while suitable training techniques for them are usually called deep learning. That architecture can capture the complex relations associated with the large dataset they try to model or classify. The problem might arise when a small dataset with a large number of attributes needs to be used in MLP of complex architectures of many layers.

H. Ensemble Learning

Ensemble Learning is the aggregation of a group of predictors (such as classifiers), which can often achieve better predictions. It is recommended to use diverse, independent classifiers in such methods to get the best outcome [60]. One way to achieve this is to use different learning algorithms.

To create a better classifier, the predictions of each classifier can be aggregated and then determine the class with the most votes. This is called the majority-vote classifier which is considered a hard-voting classifier. Using this approach will raise the chance that the individual classifiers will make very different types of errors to improve the ensemble’s accuracy. Another way is to use the same algorithms with different data subsets such as the Random forest. In that ensemble classifier, “forest” is an analogy that refers to creating decision trees that is trained by “bagging” method.

In bagging, a similar learning algorithm is used for all the predictors. To get the most reliable income, however, it is recommended to train them on different random subsets of the training set while sampling is performed with replacement. The general idea of this method is to increase the overall result accuracy due to the soft-computing nature of all methods in this area. Another type of ensemble classification is AdaBoost [61] in which the output of the weak learners, other learning algorithms, is collected into a weighted sum and this represents the boosted classifier final output.

IV. Proposed Framework

The proposed framework has three phases. The first phase has a feature engineering process which includes the CNN training techniques. The second phase represents the proposed SFS-Guided WOA for feature selection and then applying the LSH-SMOTE method for balancing the selected features. The last phase, phase three, applies the proposed voting classifier algorithm (PSO-Guided WOA) for the selected features from the second phase to classify the infected cases.

A. First Phase

In the first phase of the proposed framework, CNN is used. As dsicussed above, CNN reduce the parameters of a given problem using spatial relationships between them, which makes them a more practical classifier specially in image processing where we deal with a large number of parameters (pixels), rotation, translation, and scale of images.

Several CNN models including AlexNet [14], VGG-Net (VGG16Net and VGG19Net) [15], GoogLeNet [16], and ResNet-50 [17] are involved in this phase as shown in Fig. 4. In the CNN models, classification accuracy correlates with the extended number of convolution layers. The pre-trained CNN models are employed in this phase.

FIGURE 3.

FIGURE 3.

How to move a particle in the PSO algorithm.

FIGURE 4.

FIGURE 4.

First and second phases of the proposed framework for COVID-19 patient classification.

To understand the CT images in the datasets, a Radiology Registrar at the Typical Medical complex in Riyadh and a Fellow of The Royal College of Radiologists in UK help the authors. They guided the authors to deal with COVID-19 CT images of the infected cases to differentiate them from the non-infected cases. The preprocessing step makes the data ready for the machine learning models. Based on the problem of COVID-19 and the available dataset, some data processing tasks are required before feeding the images to the learning model.

To feed the current dataset of images to the convolutional network, they must be resized to have the same size. All the CT images have been resized to Inline graphic by the Nearest Neighbour interpolation function which is a simple and commonly used. The learning model can be applied in this stage for salient features extraction from CT images by altering the nodes in the fully connected layer and performing a fine-tuning using the input dataset. Then, the Min-Max-Scalar is employed for the Inline graphicth input image Inline graphic normalization to be within [0, 1] by applying the following form

A.

where Inline graphic is the resized image.

The data augmentation technique is applied in this research on the existing data to create new training data artificially. Image augmentation, as a type of data augmentation, creates versions of the images in the training dataset. Image transformations include horizontal and vertical shift, horizontal and vertical flip, random rotation, and random zoom are applied to the input dataset. The shift augmentation moves all pixels of the CT image in horizontal or vertical direction and keeps the image at the same dimensions. The flip process reverses all pixels rows and columns for a horizontal flip or vertical flip. The rotation augmentation rotates the CT image randomly clockwise from 0 to 360 degrees. Finally, the zoom augmentation zooms the CT image randomly by a factor range [0.9, 1.1]. The image augmentation algorithm is shown in (Algorithm 2).

Algorithm 2 Image Augmentation Algorithm

  • 1:

    Input Resized CT images Inline graphic, where Inline graphic is the number of images and Inline graphic input image is denoted as Inline graphic

  • 2:

    Initialize Inline graphic random [0:360] and Inline graphic random [0.9:1.1]

  • 3:

    for (Inline graphic) do

  • 4:

    Get Inline graphic Vshift (Inline graphic)

  • 5:

    Get Inline graphic Hshift (Inline graphic)

  • 6:

    Get Inline graphic Vflip (Inline graphic)

  • 7:

    Get Inline graphic Hflip (Inline graphic)

  • 8:

    Get Inline graphic Rotation (Inline graphic)

  • 9:

    Get Inline graphic Zoom (Inline graphic)

  • 10:

    end for

  • 11:

    Output Inline graphic, (Image transformations)

B. Second Phase

One of the most powerful methods to solve applications in radiology problems are Meta-heuristic algorithms. Optimization is the process by which the best possible solution is found for a particular problem from all the available solutions. The acceptable solutions are provided by these optimization techniques with less computational effort in a reasonable time. This section describes the proposed (SFS-Guided WOA) algorithm for feature selection. The numerical features that are extracted from the first phase of the CNN model are the input to the second phase for the proposed algorithm as shown in Fig. 4. The SMOTE and LSH-SMOTE methods are then applied for balancing the selected features for improving the accuracy of COVID-19 classification at the last phase.

1). Guided WOA

The Guided WOA is a modified version of the original WOA. To overcome the drawback of this method, the search strategy for one random whale can be replaced with an advanced strategy that can move the whales rapidly toward the best solution or prey. From the original WOA, Eq. 4 forces whales to move around each other randomly which is similar to the global search. In the modified WOA (Guided WOA), to enhance exploration performance, a whale can follow three random whales instead of one. This can force whales for more exploration and not being affected by the leader position by replacing Eq. 4 with the following equation

1).

where Inline graphic, Inline graphic, and Inline graphic are three random solutions. Inline graphic is random value between [0, 0.5]. Inline graphic and Inline graphic are two random values between [0, 1]. Inline graphic decreases exponentially instead of linearly to smoothly change between exploitation and exploration and calculated as

1).

where Inline graphic represents iteration number and Inline graphic indicates maximum number of iterations. The proposed SFS-Guided WOA algorithm is shown in (Algorithm 3).

Algorithm 3 Pseudo-Code of Proposed SFS-Guided WOA
  • 1:

    Initialize WOA population Inline graphic with size Inline graphic, maximum iterations Inline graphic, fitness function Inline graphic.

  • 2:

    Initialize WOA parameters Inline graphic, Inline graphic, Inline graphic, Inline graphic, Inline graphic, Inline graphic, Inline graphic

  • 3:

    Initialize Guided WOA parameters Inline graphic, Inline graphic, Inline graphic

  • 4:

    Set t = 1

  • 5:

    Convert solution to binary [0 or 1].

  • 6:

    Calculate fitness function Inline graphic for each Inline graphic

  • 7:

    Find best individual Inline graphic

  • 8:

    while Inline graphic (Termination condition) do

  • 9:

    for (Inline graphic) do

  • 10:

    if (Inline graphic) then

  • 11:

    if (Inline graphic) then

  • 12:

    Update position of current search agent as Inline graphic

  • 13:

    else

  • 14:

    Select three random search agents Inline graphic, Inline graphic, and Inline graphic

  • 15:

    Update (Inline graphic) by the exponential form of Inline graphic

  • 16:

    Update position of current search agent as Inline graphic

  • 17:

    end if

  • 18:

    else

  • 19:

    Update position of current search agent as Inline graphic

  • 20:

    end if

  • 21:

    end for

  • 22:

    for (Inline graphic) do

  • 23:

    Calculate Inline graphic

  • 24:

    end for

  • 25:

    Update Inline graphic, Inline graphic, Inline graphic, Inline graphic, Inline graphic

  • 26:

    Convert updated solution to binary by Eq. 11.

  • 27:

    Calculate fitness function Inline graphic for each Inline graphic

  • 28:

    Find best individual Inline graphic

  • 29:

    Set t = t +1

  • 30:

    end while

  • 31:

    return Inline graphic

2). Diffusion Procedure of SFS

Based on the diffusion procedure of the SFS algorithm, a series of random walks around the best solution can be created. This increases the exploration capability of the Guided WOA using this diffusion process for getting the best solution. The Gaussian random walks as a part of the diffusion process around the updated best position Inline graphic is calculated as

2).

where Inline graphic is the updated best solution based on the diffusion process. The parameters of Inline graphic and Inline graphic are random numbers in [0, 1]. Inline graphic and Inline graphic indicate the best point position and the Inline graphicth point in the surrounding group. Inline graphic is equal to Inline graphic and Inline graphic is equal to Inline graphic since the number of generation around the best solution decreases.

3). Binary Optimizer

For the feature selection, the solution is converted to a binary solution of 0 or 1. The following sigmoid function is applied to convert the continues solution to a binary one

3).

where Inline graphic is the best position at iteration Inline graphic. The role of the Inline graphic function is to scale the continuous values between 0 and 1. The condition of Inline graphic is used to decide whether the value of the dimension will be 0 or 1.

4). Selected Features Balance

The LSH-SMOTE technique is employed in this research to balance the selected features by the proposed SFS-Guided WOA algorithm to improve the performance of the classification algorithm. The LSH-SMOTE technique consists of the following steps:

  • 1)

    LSH-SMOTE initialization,

  • 2)

    converting the minority class instances into vectors,

  • 3)

    creating Hash Codes by using Hash Functions then creating Hash Tables,

  • 4)

    creating the nearest Neighbors List,

  • 5)

    Synthetic instances generation using the SMOTE algorithm.

5). Computational Complexity Analysis

The SFS-Guided WOA algorithm’ computational complexity according to Algorithm (3) will be discussed. Let Inline graphic as number of population; Inline graphic as total number of iterations. For each part of the algorithm, the time complexity can be defined as:

  • Population initialization: Inline graphic (1).

  • Parameters initialization: Inline graphic, Inline graphic, Inline graphic, Inline graphic, Inline graphic, Inline graphic, Inline graphic, Inline graphic, Inline graphic, Inline graphic: Inline graphic (1).

  • Iteration counter initialization: Inline graphic (1).

  • Binary conversion: Inline graphic (Inline graphic).

  • Objective function evaluation: Inline graphic (Inline graphic).

  • Finding the best individual: Inline graphic (Inline graphic).

  • Position updating: Inline graphic (Inline graphic).

  • Diffusion process calculation: Inline graphic (Inline graphic).

  • Updating Inline graphic by the exponential form: Inline graphic (Inline graphic).

  • Updating parameters Inline graphic, Inline graphic, Inline graphic, Inline graphic, Inline graphic: Inline graphic (Inline graphic).

  • Converting updated solution to binary: Inline graphic (Inline graphic).

  • Objective function evaluation: Inline graphic (Inline graphic).

  • Best individual update: Inline graphic (Inline graphic).

  • Iteration counter increment: Inline graphic (Inline graphic).

As per the above complexities, the overall complexity of the proposed SFS-Guided WOA algorithm is Inline graphic (Inline graphic). Considering the number of variables as Inline graphic, the final computational complexity of the algorithm will be Inline graphic (Inline graphic).

C. Third Phase

The third and last phase is the classification of infected patients. Figure 5 shows the third phase of the proposed framework for COVID-19 patient classification. In this section, a voting classifier is proposed based on PSO and Guided WOA algorithms as shown in Algorithm 4. The PSO-Guided WOA aggregates the SVM, NN, KNN, and DT classifiers to improve the ensemble’s accuracy. After balancing the selected features by the SMOTE or LSH-SMOTE algorithms, the classifiers are trained to get the optimal weights. The PSO-Guided WOA starts to optimize theses weights.

Algorithm 4 Pseudo-Code of Proposed PSO-Guided WOA

  • 1:

    Initialize WOA population Inline graphic with size Inline graphic, maximum iterations Inline graphic, fitness function Inline graphic.

  • 2:

    Initialize WOA parameters Inline graphic, Inline graphic, Inline graphic, Inline graphic, Inline graphic, Inline graphic, Inline graphic

  • 3:

    Initialize Guided WOA parameters Inline graphic, Inline graphic, Inline graphic

  • 4:

    Set t = 1

  • 5:

    Calculate fitness function Inline graphic for each Inline graphic

  • 6:

    Find best individual Inline graphic

  • 7:

    while Inline graphic (Termination condition) do

  • 8:

    if (Inline graphic) then

  • 9:

    for (Inline graphic) do

  • 10:

    if (Inline graphic) then

  • 11:

    if (Inline graphic) then

  • 12:

    Update position of current search agent as Inline graphic

  • 13:

    else

  • 14:

    Select three random search agents Inline graphic, Inline graphic, and Inline graphic

  • 15:

    Update (Inline graphic) by the exponential form of

  • 16:

    Inline graphic Update position of current search agent as Inline graphic

  • 17:

    end if

  • 18:

    else

  • 19:

    Update position of current search agent as Inline graphic

  • 20:

    end if

  • 21:

    end for

  • 22:

    Calculate fitness function Inline graphic for each Inline graphic from Guided WOA

  • 23:

    else

  • 24:

    Calculate fitness function Inline graphic for each Inline graphic from PSO

  • 25:

    end if

  • 26:

    Update Inline graphic, Inline graphic, Inline graphic, Inline graphic, Inline graphic

  • 27:

    Find best individual Inline graphic

  • 28:

    Set t = t +1

  • 29:

    end while

  • 30:

    return Inline graphic

FIGURE 5.

FIGURE 5.

Third phase of the proposed framework for COVID-19 patient classification.

For the proposed Algorithm 4, the guided WOA in section IV-B1 is employed in the algorithm development. After the initialization of the WOA algorithm and find the first best solution Inline graphic (Lines from 1 to 6), the iteration number Inline graphic starts to divide the calculation of the fitness function from the guided WOA or from the PSO algorithm. If Inline graphic (Line 8), then the algorithm goes through the updating positions and calculating the fitness function Inline graphic for the updated solutions from the guided WOA (Lines from 9 to 22). Otherwise, the fitness function Inline graphic will be calculated based on The PSO algorithm (Line 24).

1). Computational Complexity Analysis

The proposed PSO-Guided WOA algorithm’ computational complexity will be discussed here according to Algorithm (4).Let Inline graphic as number of population; Inline graphic as number of iterations. For each part of the algorithm, the time complexity can be defined as:

  • Population initialization: Inline graphic (1).

  • Parameters initialization Inline graphic, Inline graphic, Inline graphic, Inline graphic, Inline graphic, Inline graphic, Inline graphic, Inline graphic, Inline graphic, Inline graphic: Inline graphic (1).

  • Iteration counter initialization: Inline graphic (1).

  • Objective function evaluation: Inline graphic (Inline graphic).

  • Determining the best solution: Inline graphic (Inline graphic).

  • Position updating: Inline graphic (Inline graphic).

  • Objective function evaluation for each individual from Guided WOA: Inline graphic (Inline graphic).

  • Fitness function calculation for each individual from PSO: Inline graphic (Inline graphic).

  • Updating parameters Inline graphic, Inline graphic, Inline graphic, Inline graphic, Inline graphic: Inline graphic (Inline graphic).

  • Best solution update: Inline graphic (Inline graphic).

  • Iteration counter increament: Inline graphic (Inline graphic).

Thus, the overal complexity PSO-Guided WOA algorithm is Inline graphic (Inline graphic). Considering a problem with Inline graphic variables, the final computational complexity of the algorithm will be Inline graphic (Inline graphic).

D. Objective Function

Objective functions are used to evaluate the solutions in an optimization algorithm. The function is depending on two parameters of the classification error rate and the number of selected features. The solution is good if the subset of features gives a lower number of selected features and a lower classification error rate. The following equation is used to get the quality of each solution

D.

where Inline graphic is the rate of error for the optimizer, Inline graphic indicates the number of selected features, Inline graphic indicates the total number of features and Inline graphic manage the importance of the number of the selected feature for population with size Inline graphic and the classification error rate.

V. Experimental Results

The experiments section in this article is divided into three scenarios. The first scenario is based on the first phase of the proposed model. This experiment shows the effectiveness of different CNN models for classifying the COVID-19 cases and interns show the importance of extracting features for the next phase. In the second scenario, the proposed feature selection algorithm (SFS-Guided WOA) is tested and compared to other algorithms to show its performance. The third scenario is designed to investigate the ability of the proposed voting optimizer (PSO-Guided WOA) for improving the classification accuracy of the COVID-19 cases. Finally, Wilcoxon’s rank-sum test and t-test are performed to verify the superiority of the proposed algorithms in a statistical way. The CT images datasets, [49], are separated randomly in the experiment of the first scenario into (60%, 20%, 20%) images for the training, validation, and testing processes.

A. First Scenario: Model’s First Phase

The first experiment is designed to investigate the classification accuracy of five CNN models namely AlexNet [14], VGG-Net (VGG16Net and VGG19Net) [15], GoogLeNet [16], and ResNet-50 [17] for the tested dataset. In this scenario, several performance metrics are calculated to measure the performance of the different models for COVID-19 classification. Table 3 shows the CNN experimental setup employed in the first scenario. The default parameters are employed in this experiment since the first stage is used to extract features of the CT images from the earlier layers of a CNN model to be used for the next scenario for features selection and balancing.

TABLE 3. CNN Experimental Setup.

Parameter Value
CNN Default training options
Momentum Learn 0.9000
RateDropFactor 0.1000
LearnRateDropPeriod 10
L2Regularization 1.0000c-04
GradientThresholdMethod 12norm
GradientThreshold Inf
VerboseFrequency 50
ValidationData [imds]
ValidationFrequency 50
ValidationPatience Inf
ResetInputNormalization 1
CNN Custom training options
\\ExecutionEnvironment gpu
InitiallearnRate 1.0000e-04
MaxEpochs 20
MiniBatchSize 8
Shuffle every-epoch
Verbose 0
Optimizer sgdm
LearnRateSchedule piecwise

1). First Scenario: Performance Metrics

The performance metrics calculated for the first phase are accuracy, sensitivity, specificity, precision (PPV), Negative Predictive Value (NPV), and F-score. Let Inline graphic represents the true-positive value and Inline graphic represents the true-negative value, while Inline graphic indicates the false-negative value and Inline graphic indicates the false-positive value. The metrics are defined as in the following equations.

  • Accuracy: measures the model ability to identify the whole cases correctly, regardless the cases are being positive or negative and can be formed as
    graphic file with name M314.gif
  • Sensitivity: called the true positive rate (TPR) or recall. It computes the capability of the positive case and is calculated as
    graphic file with name M315.gif
  • Specificity: called the true negative rate (TNR) or selectivity. It gets the capability of finding negative cases and is calculated as
    graphic file with name M316.gif
  • Precision: called positive predictive value (PPV). It directs the rate of true positives among all positive values. It is calculated as
    graphic file with name M317.gif
  • Negative Predictive Value (NPV): It directs rate of true negatives among all negative values. It is calculated as
    graphic file with name M318.gif
  • F-score: measures the harmonic mean of precision and sensitivity and is calculated as
    graphic file with name M319.gif

2). First Scenario: Results and Discussion

This scenario results are shown in Table 4. The results show that the precision (Pvalue) of the GoogLeNet model of 84.75% which is better than VGG19Net (83.78%), ResNet-50 (81.08%), AlexNet (75%), and VGG16Net (51.75%) models. The AlexNet model outperforms other models with an F-score of 77.88%. However, the GoogLeNet model has better specificity of 92.44% than other models. According to sensitivity, the rate of the VGG16Net model of 95.08% is better than the sensitivity rate of AlexNet (81%), ResNet-50 (62.5%), VGG19Net (62%), and GoogLeNet (50%) models, respectively. For the Pvalue, the VGG16Net model has a better percentage of 87.74%. As an overall performance metric for the models, the AlexNet model has an accuracy of 79% whereas VGG19Net, ResNet-50, GoogLeNet, and VGG16Net have the accuracy of 77.17%, 77.17%, 73.06%, and 58.21% for the tested COVID-19 dataset, respectively.

TABLE 4. Comparison of the Performance Metrics for the COVID-19 Classification Based on CNN Models.
CNN Models/Metric Accuracy Sensitivity (TPR) Specificity (TNR) Pvalue (PPV) Nvalue (NPV) F-score
AlexNet 0.7900 0.8100 0.7731 0.7500 0.8288 0.7788
VGG16Net 0.5821 0.9508 0.2844 0.5175 0.8774 0.6702
VGG19Net 0.7717 0.6200 0.8992 0.8378 0.7379 0.7126
GoogLeNet 0.7306 0.5000 0.9244 0.8475 0.6875 0.6289
ResNet-50 0.7717 0.6250 0.8862 0.8108 0.7517 0.7059

Based on this experiment, the highest accuracy that can be achieved for the CT images from the COVID-19 dataset tested in this research is 79% by the AlexNet model. Since this is not acceptable accuracy in this critical endeavor, the features are extracted from the earlier layers of the AlexNet model, according to its promising performance, to be used for the next scenario for features selection and balancing.

B. Second Scenario: Model’s Second Phase

In this scenario, the importance and performance of the proposed feature selection algorithm (SFS-Guided WOA) are investigated. The proposed algorithm in the second phase is compared to other algorithms of the original WOA [25], Grey Wolf Optimizer (GWO) [26], Genetic Algorithm (GA) [27], PSO [28], hybrid of PSO and GWO (GWO-PSO) [29], hybrid of GA and GWO (GWO-GA), Bat Algorithm (BA) [30], Biogeography-Based Optimizer (BBO) [31], Multiverse Optimization (MVO) [32], Bowerbird Optimizer (SBO) [33], and Firefly Algorithm (FA) [34] in terms of average error, average select size, average (mean) fitness, best fitness, worst fitness, and standard deviation fitness, to show its performance. Table 5 shows the configuration of the proposed (SFS-Guided WOA) algorithm in the experiments. The parameters of Inline graphic and Inline graphic in the fitness function are assigned to 0.99 and 0.01, respectively. Table 6 shows the configuration of the compared algorithms in the experiments.

TABLE 5. Proposed (SFS-Guided WOA) Algorithm Configuration.

Parameter Value
Number of search agents 10
Number of iterations 80
Number repetitions of runs 20
Problem dimension Number of features
Search domain [0,1]
K-neighbors 5
K-fold cross-validation 10
Maximum diffusion level 1
Inline graphic Parameter in Inline graphic 0.99
Inline graphic Parameter in Inline graphic 0.01

TABLE 6. Compared Algorithms Configuration for Feature Selection.

Algorithm Parameter (s) Value (s)
GWO Inline graphic 2 to 0
PSO Inertia Inline graphic, Inline graphic [0.9, 0.6]
Acceleration constants Inline graphic, Inline graphic [2, 2]
GA Mutation ratio 0.1
Crossover 0.9
Selection mechanism Roulette wheel
SBO Step size 0.94
Mutation probability 0.05
Difference between the upper and lower limit 0.02
WOA Inline graphic 2 to 0
Inline graphic [0, 1]
MVO Wormhole existence probability [0.2,1]
FA Number of fireflies 10
BA Pluse rate 0.5
Loudness 0.5
Frequency [0, 1]
BBO Probability of Immigration [0, 1]
Probability of Mutation 0.05
Probability of Habitat modification 1.0
Step size 1.0
Migration rate 1.0
Maximum immigration 1.0

1). Second Scenario: Performance Metrics

For the evaluation of the proposed SFS-Guided WOA algorithm effectiveness, the following metrics are employed. Let Inline graphic is the number repetitions of runs of an optimizer for the feature selection problem; Inline graphic is the best solution at the run number Inline graphic; Inline graphic is the number of tested points.

  • Average Error is calculated to show the accuracy of the classifier in giving the selected feature set. It is calculated as
    graphic file with name M337.gif
    where Inline graphic is the label of the classifier output for point Inline graphic, and Inline graphic is the label of the class for point Inline graphic, and Inline graphic calculates the matching between two inputs.
  • Average Fitness is the selected features average size to the total number of features in the dataset (Inline graphic). It is calculated as
    graphic file with name M344.gif
    where Inline graphic is the size of the vector Inline graphic.
  • Mean is the average of the solutions output from running an optimizer for several times Inline graphic. It is calculated as
    graphic file with name M348.gif
  • Best Fitness is the minimum fitness function of an optimizer running for several times Inline graphic. It is calculated as
    graphic file with name M350.gif
  • Worst Fitness is the worst solution found by an optimizer running for several times Inline graphic. It is calculated as
    graphic file with name M352.gif
  • Standard Deviation (SD) is the obtained best solutions variation which can be found by running an optimizer several times Inline graphic. SD can be calculated as
    graphic file with name M354.gif
    where Inline graphic is the average defined in equation 21.

2). Second Scenario: Results and Discussion

The results of the proposed SFS-Guided WOA algorithm in this experiment are shown in Table 7. The lower error indicates that the optimizer has selected the proper set of features for the next stage. The SFS-Guided WOA algorithm achieved the minimum average error of (0.1381) in selecting the proper features. The feature selection algorithms ordered from the best to the worst according to the minimum error for the current problem are SFS-Guided WOA, PSO, GWO, GWO-GA, WOA, GA, BA, GWO-PSO, FA, BBO, MVO, and lastly SBO. Note that, the proposed algorithm outperforms the original WOA algorithm. Table 7 also shows that the proposed algorithm can find the lowest fitness value (0.2013), for the selected features of the COVID-19 datasets, which is lower than the compared algorithms values. The proposed algorithm can find the best fitness value of (0.1031) compared to other optimization techniques throughout runs. On the other hand, SFS-Guided WOA can not find the worst fitness and it has the lowest standard deviation compared to other algorithms that prove the stability and robustness of the proposed algorithm.

TABLE 7. Performance of the Proposed Feature Selection Algorithm (SFS-Guided WOA) Compared to Other Algorithms.
Metric/Optimizer SFS-Guided WOA GWO GWO-PSO PSO BA WOA BBO MVO SBO GWO-GA FA GA
Average Error 0.1381 0.1553 0.1946 0.1891 0.1987 0.1889 0.1573 0.1658 0.1974 0.1754 0.1875 0.1689
Average Select size 0.0909 0.2909 0.4242 0.2909 0.4303 0.4543 0.4547 0.3874 0.4612 0.2137 0.3254 0.2333
Average Fitness 0.2013 0.2175 0.2258 0.2159 0.2388 0.2237 0.2216 0.2456 0.2556 0.2236 0.2678 0.2289
Best Fitness 0.1031 0.1378 0.1793 0.1962 0.1285 0.1878 0.2113 0.1708 0.1987 0.2014 0.1865 0.1322
Worst Fitness 0.2016 0.2047 0.2893 0.2639 0.2301 0.2639 0.2978 0.2888 0.2784 0.2776 0.2841 0.2473
Standard Deviation Fitness 0.0236 0.0283 0.0465 0.0277 0.0376 0.0299 0.0726 0.0784 0.0886 0.0289 0.0645 0.0299

Based on this experiment, the selected features are then balanced using two methods named SMOTE and LSH-SMOTE to be ready for the classification scenario. For both algorithms, the nearest neighbors parameter Inline graphic, and the oversampling percentage is 50% of features distribution (majority class = minority class). For the SMOTE algorithm, the number of instances per leaf is equal to 2. For the LSH-SMOTE algorithm, the Hashes parameter Inline graphic and the Hash tables parameter Inline graphic.

3). Second Scenario: Wilcoxon’s Rank-Sum

For getting the p-values between the proposed SFS-Guided WOA algorithm and other algorithms, Wilcoxon’s rank-sum test is employed. This statistical test can determine if the results of the proposed algorithm and other algorithms have a significant difference or not; p-value < 0.05 will demonstrate significant superiority. By contrast, a p-value >0.05 shows that the results have no significant difference. Hypothesis testing is formulated here in terms of two hypotheses; the null hypothesis (Inline graphic: Inline graphic, Inline graphic, Inline graphic, Inline graphic, Inline graphic) and the alternate hypothesis (Inline graphic: Means are not all equal). Table 8 shows the results of p-value in which p-values less than 0.05 could be achieved between the proposed algorithm and other algorithms showing the superiority of the SFS-Guided WOA algorithm and indicating that the algorithm is statistically significant. Thus, the alternate hypothesis Inline graphic is accepted.

TABLE 8. p-values of SFS-Guided WOA in Comparison to Other Algorithms Using Wilcoxon’s Rank-Sum.
GWO GWO-PSO PSO BA WAO BBO MVO SBO GWO-GA FA GA
1.13E-05 1.13E-05 1.13E-05 1.13E-05 1.13E-05 1.13E-05 1.13E-05 1.13E-05 1.13E-05 1.13E-05 1.13E-05

C. Third Scenario: Model’s Third Phase

This scenario is divided into three experiments and statistical tests. The first experiment is designed to investigate the results for the single classifiers of SVM, KNN, NN, and DT based on balanced and unbalanced features that are selected from the second scenario. The next experiment is performed to compare the proposed voting classifier (PSO-Guided WOA) with other ensemble learning techniques. In the last experiment, the proposed algorithm is compared with other voting classifier algorithms to check its effectiveness. Statistical tests of ANOVA and T-test are performed between the compared algorithms to show the effectiveness of the proposed algorithm. Table 9 shows the configuration of the proposed (PSO-Guided WOA) algorithm in the experiments. The parameters of Inline graphic and Inline graphic in the fitness function are assigned to 0.99 and 0.01, respectively. Table 10 shows the configuration of the compared algorithms in the experiments.

TABLE 9. Proposed (PSO-Guided WOA) Algorithm Configuration.

Parameter Value
Number of whales 20
Number of iterations 20
Number repetitions of runs 20
Inertia Inline graphic, Inline graphic [0.9, 0.6]
Acceleration constants Inline graphic, Inline graphic [2, 2]
K-neighbors 5
K-fold cross-validation 10
Inline graphic Parameter in Inline graphic 0.99
Inline graphic Parameter in Inline graphic 0.01

TABLE 10. Compared Algorithms Configuration for Classification.

Algorithm Parameter (s) Value (s)
GWO Inline graphic 2 to 0
No. of wolves 20
No. of iterations 20
PSO Inertia Inline graphic, Inline graphic [0.9, 0.6]
Acceleration constants Inline graphic, Inline graphic [2, 2]
No. of particles 20
Generations 20
GA Mutation ratio 0.1
Crossover 0.9
Selection mechanism Roulette wheel
Population size 20
Generations 20
WOA Inline graphic 2 to 0
Inline graphic [0, 1]
No. of whales 20
No. of iterations 20

1). Third Scenario: Performance Metrics

This scenario performance metrics are the Area Under the ROC Curve (AUC) and the Mean Square Error (MSE). AUC is a good indicator of classification performance due to being independent from the distribution of instances between classes which is also referred to as a balanced accuracy or macro-average [51]. In the current case of binary classification, the balanced accuracy is equal to the arithmetic mean of specificity and sensitivity, or AUC with binary predictions rather than scores. The AUC (balanced accuracy) value can be calculated as follows:

1).

The Mean Square Error (MSE) evaluates the classifiers performance, calculates the difference between the required and the actual output of the classifiers according to this equation:

1).

where Inline graphic indicates number of outputs, Inline graphic indicates the Inline graphicth input neuron optimal output when the Inline graphicth training instance is applied, and Inline graphic indicates optimal output actual output of the Inline graphicth input neuron when the Inline graphicth training instance appears in the input.

2). Third Scenario: Results and Discussion

The first experiment results for the SVM, KNN, NN, and DT as a single classifiers are shown in Table 11. The classifier results are shown based on three cases of no preprocessing, balancing selected features by the SMOTE algorithm, and balancing selected features by the LSH-SMOTE algorithm. Note from Table 11 that, the DT classifier achieved the highest AUC percentage of 0.911 with the minimum MSE of (0.007932). This result show the importance of balancing the selected features from the previous stage by the LSH-SMOTE algorithm.

TABLE 11. AUC and MSE of the Signal Classifiers.
Metric Preprocessing State SVM NN KNN DT
AUC without preprocessing 0.684 0.713 0.661 0.793
with SMOTE preprocessing 0.721 0.761 0.717 0.831
with LSH-SMOTE preprocessing 0.847 0.867 0.836 0.911
MSE without preprocessing 0.099845 0.082373 0.114932 0.042853
with SMOTE preprocessing 0.089411 0.067079 0.085852 0.035723
with LSH-SMOTE preprocessing 0.033685 0.026574 0.039543 0.007932

The next experiment results for comparing the proposed algorithm with other ensemble learning methods of Bagging, AdaBoost, and Majority voting are shown in Table 12. This table shows that the proposed voting classifier (PSO-Guided WOA) with LSH-SMOTE preprocessing can achieve AUC result of 0.995 which outperforms other ensemble learning techniques. The MSE of the proposed (2.49569E-05) is the minimum MSE compared with Bagging (0.028231), AdaBoost (0.014892), and Majority voting (0.005931) techniques. The last experiment results for comparing the voting classifier with other voting classifiers using WOA, GWO, GA, and PSO are shown in Table 13. The results show that the PSO-Guided WOA algorithm with AUC of 0.995 outperforms the voting WOA (AUC = 0.931), voting GWO (AUC = 0.946), voting GA (AUC = 0.939), and voting PSO (AUC = 0.954), respectively. Figure 6 shows the ROC curves of the proposed voting (PSO-Guided WOA) algorithm versus compared voting algorithms. These figures show that the proposed algorithm is able to distinguish between the COVID-19 and non-COVID-19 cases with a high AUC value near to 1.0 as shown in Table 13.

TABLE 12. Comparing the Proposed Algorithm With Other Ensemble Learning Methods.
Metric/Ensemble Learning Bagging AdaBoost Majority voting Voting (PSO-Guided WOA)
AUC with LSH-SMOTE 0.842 0.877 0.924 0.995
MSE with LSH-SMOTE 0.028231 0.014892 0.005931 2.49569E-05
TABLE 13. Comparing the Proposed Algorithm With Other Voting Classifiers.
Metric/Ensemble Learning Voting (PSO-Guided WOA) Voting WOA Voting GWO Voting GA Voting PSO
AUC with LSH-SMOTE 0.995 0.931 0.946 0.939 0.954
MSE with LSH-SMOTE 2.49569E-05 0.006084 0.003025 0.003721 0.00151
FIGURE 6.

FIGURE 6.

ROC curves of the proposed voting (PSO-Guided WOA) algorithm versus compared algorithms.

3). Third Scenario: Statistical Test

To conclude whether there is any statistical difference between the MSE of the proposed (PSO-Guided WOA) algorithm and other compared algorithms, a one-way analysis of variance (ANOVA) test was applied. The hypothesis testing can be formulated here in terms of two hypotheses; the null hypothesis (Inline graphic: Inline graphic), where A1: Voting (PSO-Guided WOA), B1: Voting WOA, C1: Voting GWO, D1: Voting GA, and E1: Voting PSO, and the alternate hypothesis (Inline graphic: Means are not all equal). The ANOVA test results are shown in Table 14. Figure 7 shows the ANOVA test for proposed and the compared algorithms versus the objective function. Based on this test results, the alternate hypothesis Inline graphic is accepted. However, we cannot tell which algorithm is better from ANOVA, so another test is conducted between every two algorithms.

TABLE 14. A One-Way Analysis of Variance (ANOVA) Test Results.
Source of Variation SS df MS F P-value F crit
Between Groups 0.000510289 4 0.000128 25.28941 2.82E-14 2.467494
Within Groups 0.000479227 95 5.04E-06
Total 0.000989516 99
FIGURE 7.

FIGURE 7.

ANOVA test for different algorithms.

A one-tailed T-Test at 0.05 significance level is performed. Hypothesis testing is formulated here in terms of two hypotheses; the null hypothesis (Inline graphic: Inline graphic) and the alternate hypothesis (Inline graphic: Means are not all equal). The results in Table 15, for 20 samples (Number repetitions of runs) as mentioned in Table 9, show that the p-values are less than 0.05 which indicates that there is a statistically significant difference between groups. Thus, the alternate hypothesis Inline graphic is accepted.

TABLE 15. A One-Tailed T-Test at 0.05 Significance Level Results. A1: Voting (PSO-Guided WOA), B1: Voting WOA, C1: Voting GWO, D1: Voting GA, and E1: Voting PSO.
A1-B1 A1-C1 A1-D1 A1-E1
Correlation 0.234821 −0.14323 0.436198 −0.3496
T.Test 2.26E-17 2.36E-07 7.65E-10 1.09E-11

VI. Discussion

The experiments in this research are designed based on three scenarios to assess the performance and accuracy of the proposed framework for COVID-19 classification. The first scenario shows that the highest classification accuracy of the compared CNN models can be achieved by the AlexNet model for the CT images from the tested COVID-19 dataset. Based on these results, the features are extracted from the earlier layers of the AlexNet model to be used for the next scenario for features selection and balancing. In the second scenario, the performance of the proposed feature selection algorithm (SFS-Guided WOA) is assessed. Results show that the proposed algorithm outperforms the compared algorithms, including the original WOA algorithm, and could find the lowest fitness value for the feature selection of the extracted features from the COVID-19 datasets. In addition, the proposed algorithm has the lowest standard deviation compared to other algorithms that prove the stability and robustness of the proposed technique. Based on the second scenario results, the selected features are then balanced using the SMOTE and LSH-SMOTE methods to be ready for the last stage which includes the final classification. The third scenario shows the performance of the proposed classification algorithm (PSO-Guided WOA). Results show that the proposed voting classifier (PSO-Guided WOA) with LSH-SMOTE preprocessing could achieve an AUC with binary predictions (balanced accuracy) result of 0.995 and a MSE of 2.49569E-05 which outperforms other state-of-the-art ensemble learning techniques. That shows the importance of balancing the selected features from the previous stage by the LSH-SMOTE algorithm. The experimental results for comparing the voting classifier with other voting classifiers using WOA, GWO, GA, and PSO show the superiority of the proposed framework to identify COVID-19 patients using CT images. Thus, the efficacy of diagnosis can be improved while avoiding the radiologists the heavy workload associated with the initial screening of COVID-19 pneumonia.

VII. Conclusion and Future Work

This article proposes a framework for COVID-19 classification with three cascaded phases. In the first phase, the hierarchical feature representation is automatically extracted from the training CT images by the CNN model of AlexNet. Afterward, the proposed feature selection algorithm, using SFS and Guided WOA techniques, is applied to select features in the second phase. The selected features are then balanced by the LSH-SMOTE algorithm to improve the classification results. In the last phase, a voting classifier, using PSO and Guided WOA techniques, is proposed to aggregate the predictions of four single classifiers, named SVM, NN, KNN, and DT, and predict the most voted class. This increases the chance that the individual classifiers will make very different types of errors to improve the ensemble’s accuracy. Two datasets are used to test the proposed model. The first is the COVID-19 dataset which has CT images containing clinical findings of COVID-19 and the second is the non-COVID-19 dataset that has extra CT images with clinical cases that have no COVID-19. For feature selection, the proposed SFS-Guided WOA algorithm is compared in experiments with the original WOA, GWO, GA, PSO, hybrid of PSO and GWO (GWO-PSO), hybrid of GA and GWO (GWO-GA), BA, BBO, MVO, SBO, and FA in terms of average error, average select size, average (mean) fitness, best fitness, worst fitness, and standard deviation fitness. Finally, the proposed voting classifier (PSO-Guided WOA) result is compared with voting WOA, voting GWO, voting GA, and Voting PSO in terms of AUC and MSE. The statistical analysis of Wilcoxon rank-sum, ANOVA, and T-Test shows the superiority of the proposed algorithms. The utilization of each successive phase is aimed to improve the overall accuracy to offer a viable and reliable paradigm in the battle against the spread of COVID-19. A future research direction will be to tune the CNN parameters to increase the overall classification accuracy in case of using other datasets that cannot achieve satisfactory performance. Moreover, the proposed algorithms can be applied to several medical image processing applications that use other imaging modalities.

Acknowledgment

The authors would like to thank Dr. Mohamed Elsayed Gawish, Radiology Registrar at Typical Medical complex in Riyadh, and Dr. Shaaban Omar, Fellow of The Royal College of Radiologists in the U.K., for their help to understand the CT image datasets. They guided the authors to deal with COVID-19 CT images of the infected cases and to differentiate them from other non-COVID-19 cases.

Biographies

graphic file with name elken-3028012.gif

El-Sayed M. El-Kenawy (Member, IEEE) is currently an Assistant Professor with the Delta Higher Institute for Engineering and Technology (DHIET), Mansoura, Egypt. He is inspiring and motivating students by providing a thorough understanding of a variety of computer concepts. He has pioneered and launched independent research programs. His research interests include computer science and machine learning field. He is an adept at explaining sometimes complex concepts in an easy-to-understand manner.

graphic file with name ibrah-3028012.gif

Abdelhameed Ibrahim (Member, IEEE) received the bachelor’s and master’s degrees in engineering from the Computer Engineering and Systems Department, in 2001 and 2005, respectively, and the Ph.D. degree in engineering from the Faculty of Engineering, Chiba University, Japan, in 2011. He was with the Faculty of Engineering, Mansoura University, Egypt, from 2001 to 2007, where he is currently an Assistant Professor of computer engineering with the Faculty of Engineering. He serves as a Reviewer for the Journal of Electronic Imaging, Optical Engineering, the IEEE Journal of Biomedical and Health Informatics, IEEE Access, Computer Standards and Interfaces, Journal of Healthcare Engineering, IET Image Processing, Multimedia Tools and Applications, and other respected journals. His research interests include computer vision, pattern recognition, cloud computing, virtualization, and live virtual machine migration.

graphic file with name mirja-3028012.gif

Seyedali Mirjalili (Senior Member, IEEE) is currently the Director of the Centre for Artificial Intelligence Research and Optimization, Torrens University Australia at Brisbane. He is internationally recognized for his advances in swarm intelligence and optimization, including the first set of algorithms from a synthetic intelligence standpoint—a radical departure from how natural systems are typically understood—and a systematic design framework to reliably benchmark, evaluate, and propose computationally cheap robust optimization algorithms. He has published over 200 publications with over 20,000 citations and an H-index of 51. As the most cited researcher in robust optimization, he is in the list of 1% highly cited researchers and named as one of the most influential researchers in the world by the Web of Science. He is working on the applications of multi-objective and robust meta-heuristic optimization techniques as well. His research interests include robust optimization, engineering optimization, multi-objective optimization, swarm intelligence, evolutionary algorithms, and artificial neural networks’. He is an Associate Editor of several journals, including Neurocomputing, Applied Soft Computing, Advances in Engineering Software, Applied Intelligence, PLOS One, and IEEE Access.

graphic file with name eid-3028012.gif

Marwa Metwally Eid received the Ph.D. degree in electronics and communications engineering from the Faculty of Engineering, Mansoura University, Egypt, in 2015. She has been an Assistant Professor with the Delta Higher Institute for Engineering and Technology, since 2011. Her current research interests include image processing, encryption, wireless communication systems, and field programmable gate array (FPGA) applications.

graphic file with name husse-3028012.gif

Sherif E. Hussein (Member, IEEE) received the B.S. and M.S. degrees (Hons.) in computer engineering from Mansoura University, Egypt, the Ph.D. degree from The University of Strathclyde, Scotland, U.K., in 2003, and the M.B.A. (Hons.) from Cardiff Metropolitan University, Cardiff, Wales, U.K. During his Ph.D. degree, he received the ORS Award, the International Scholarship Award, and the University of Strathclyde Award. He is currently an Associate Professor with Computer Engineering and Systems Department, Mansoura University, a Computer Technical Consultant of the Egyptian Ministry of Higher Education, and the author of about 40 publications. He was a Postdoctoral Researcher with VNIIM, Saint Petersburg, Russia. He is a member of the ICMI and Institute of Engineering and Technology IET. His scientific research interests include cloud computing, data science, image and signal processing, modeling, and artificial intelligence.

Contributor Information

El-Sayed M. El-Kenawy, Email: skenawy@ieee.org.

Abdelhameed Ibrahim, Email: afai79@mans.edu.eg.

References

  • [1].Lai C.-C., Shih T.-P., Ko W.-C., Tang H.-J., and Hsueh P.-R., “Severe acute respiratory syndrome coronavirus 2 (SARS-CoV-2) and coronavirus disease-2019 (COVID-19): The epidemic and the challenges,” Int. J. Antimicrobial Agents, vol. 55, no. 3, Mar. 2020, Art. no. 105924. [Online]. Available: http://www.sciencedirect.com/science/article/pii/S0924857920300674 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [2].Wang C., Horby P. W., Hayden F. G., and Gao G. F., “A novel coronavirus outbreak of global health concern,” The Lancet, vol. 395, no. 10223, pp. 470–473, 2020, doi: 10.1016/S0140-6736(20)30185-9. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [3].Wong H. Y. F., Lam H. Y. S., Fong A. H.-T., Leung S. T., Chin T. W.-Y., Lo C. S. Y., Lui M. M.-S., Lee J. C. Y., Chiu K. W.-H., Chung T., Lee E. Y. P., Wan E. Y. F., Hung F. N. I., Lam T. P. W., Kuo M., and Ng M.-Y., “Frequency and distribution of chest radiographic findings in COVID-19 positive patients,” Radiology, vol. 2019, Mar. 2019, Art. no.201160, doi: 10.1148/radiol.2020201160. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [4].Cho S., Lim S., Kim C., Wi S., Kwon T., Youn W. S., Lee S. H., Kang B. S., and Cho S., “Enhancement of soft-tissue contrast in cone-beam CT using an anti-scatter grid with a sparse sampling approach,” Phys. Medica, vol. 70, pp. 1–9, Feb. 2020, doi: 10.1016/j.ejmp.2020.01.004. [DOI] [PubMed] [Google Scholar]
  • [5].Fu J., Wang J., Guo W., and Peng P., “Multi-mounted X-Ray cone-beam computed tomography,” Nucl. Instrum. Methods Phys. Res. A, Accel. Spectrom. Detect. Assoc. Equip., vol. 888, pp. 119–125, Oct. 2018. [Online]. Available: http://www.sciencedirect.com/science/article/pii/S0168900218300615 [Google Scholar]
  • [6].Ibrahim A., Mohammed S., and Ali H. A., “Breast cancer detection and classification using thermography: A review,” in Proc. Int. Conf. Adv. Mach. Learn. Technol. Appl., Hassanien A. E., Tolba M. F., Elhoseny M., and Mostafa M., Eds. Cham, Switzerland: Springer, 2018, pp. 496–505. [Google Scholar]
  • [7].Ye K., Zhu Q., Li M., Lu Y., and Yuan H., “A feasibility study of pulmonary nodule detection by ultralow-dose CT with adaptive statistical iterative reconstruction-V technique,” Eur. J. Radiol., vol. 119, Oct. 2019, Art. no. 108652. [DOI] [PubMed] [Google Scholar]
  • [8].Fu J., Hu X., Velroyen A., Bech M., Jiang M., and Pfeiffer F., “3D algebraic iterative reconstruction for cone-beam X-Ray differential phase-contrast computed tomography,” PLoS ONE, vol. 10, no. 3, Mar. 2015, Art. no. e0117502, doi: 10.1371/journal.pone.0117502. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [9].Walker B. J., Radtke J., Chen G.-H., Eliceiri K. W., and Mackie T. R., “A beam optics study of a modular multi-source X-ray tube for novel computed tomography applications,” Nucl. Instrum. Methods Phys. Res. A, Accel. Spectrom. Detect. Assoc. Equip., vol. 868, pp. 1–9, Oct. 2017. [Online]. Available: http://www.sciencedirect.com/science/article/pii/S0168900217306885 [Google Scholar]
  • [10].Campos L. T., de Jesus F. M., de Souza Goncalves E. A., and Magalhaes L. A. G., “Computed tomography X-Ray characterization: A Monte Carlo study,” Radiat. Phys. Chem., vol. 167, Oct. 2020, Art. no. 108359. [Online]. Available: http://www.sciencedirect.com/science/article/pii/S0969806X18313252 [Google Scholar]
  • [11].Honkanen M. K. M., Matikka H., Honkanen J. T. J., Bhattarai A., Grinstaff M. W., Joukainen A., Kröger H., Jurvelin J. S., and Töyräs J., “Imaging of proteoglycan and water contents in human articular cartilage with full–body CT using dual contrast technique,” J. Orthopaedic Res., vol. 37, no. 5, pp. 1059–1070, May 2019. [Online]. Available: https://onlinelibrary.wiley.com/doi/abs/10.1002/jor.24256 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [12].Montagnon E., Cerny M., Cadrin-Chánevert A., Hamilton V., Derennes T., Ilinca A., Vandenbroucke-Menu F., Turcotte S., Kadoury S., and Tang A., “Deep learning workflow in radiology: A primer,” Insights into Imag., vol. 11, no. 1, p. 22, Dec. 2020. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [13].Ibrahim A., Mohammed S., Ali H. A., and Hussein S. E., “Breast cancer segmentation from thermal images based on chaotic Salp swarm algorithm,” IEEE Access, vol. 8, no. 1, pp. 122121–122134, 2020, doi: 10.1109/access.2020.3007336. [DOI] [Google Scholar]
  • [14].Han J., Zhang D., Cheng G., Liu N., and Xu D., “Advanced deep-learning techniques for salient and category-specific object detection: A survey,” IEEE Signal Process. Mag., vol. 35, no. 1, pp. 84–100, Jan. 2018, doi: 10.1109/msp.2017.2749125. [DOI] [Google Scholar]
  • [15].Simonyan K. and Zisserman A., “Very deep convolutional networks for large-scale image recognition,” 2014, arXiv:1409.1556. [Online]. Available: http://arxiv.org/abs/1409.1556
  • [16].Al-Dhamari A., Sudirman R., and Mahmood N. H., “Transfer deep learning along with binary support vector machine for abnormal behavior detection,” IEEE Access, vol. 8, pp. 61085–61095, 2020, doi: 10.1109/access.2020.2982906. [DOI] [Google Scholar]
  • [17].Yu S., Xie L., Liu L., and Xia D., “Learning long-term temporal features with deep neural networks for human action recognition,” IEEE Access, vol. 8, pp. 1840–1850, 2020, doi: 10.1109/access.2019.2962284. [DOI] [Google Scholar]
  • [18].Yamashita R., Nishio M., Do R. K. G., and Togashi K., “Convolutional neural networks: An overview and application in radiology,” Insights into Imag., vol. 9, no. 4, pp. 611–629, Aug. 2018. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [19].Fouad M. M., El-Desouky A. I., Al-Hajj R., and El-Kenawy E.-S. M., “Dynamic group-based cooperative optimization algorithm,” IEEE Access, vol. 8, pp. 148378–148403, 2020, doi: 10.1109/access.2020.3015892. [DOI] [Google Scholar]
  • [20].Al-qaness M. A. A., Ewees A. A., Fan H., and El Aziz M. A., “Optimization method for forecasting confirmed cases of COVID-19 in China,” J. Clin. Med., vol. 9, no. 3, p. 674, Mar. 2020, doi: 10.3390/jcm9030674. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [21].Tharwat A., “Parameter investigation of support vector machine classifier with kernel functions,” Knowl. Inf. Syst., vol. 61, no. 3, pp. 1269–1302, Dec. 2019. [Online]. Available: http://www.sciencedirect.com/science/article/pii/S0169260720309664 [Google Scholar]
  • [22].Pereira R. M., Bertolini D., Teixeira L. O., Silla C. N., and Costa Y. M. G., “COVID-19 identification in chest X-ray images on flat and hierarchical classification scenarios,” Comput. Methods Programs Biomed., vol. 194, Oct. 2020, Art. no. 105532. [Online]. Available: http://www.sciencedirect.com/science/article/pii/S0169260720309664 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [23].Jang S., Jang Y.-E., Kim Y.-J., and Yu H., “Input initialization for inversion of neural networks using k-nearest neighbor approach,” Inf. Sci., vol. 519, pp. 229–242, May 2020. [Online]. Available: http://www.sciencedirect.com/science/article/pii/S0020025520300426 [Google Scholar]
  • [24].Breiman L., “Random forests,” Mach. Learn., vol. 45, no. 1, pp. 5–32, 2001, doi: 10.1023/a:1010933404324. [DOI] [Google Scholar]
  • [25].Mirjalili S. and Lewis A., “The whale optimization algorithm,” Adv. Eng. Softw., vol. 95, pp. 51–67, May 2016. [Google Scholar]
  • [26].Al-Tashi Q., Abdul Kadir S. J., Rais H. M., Mirjalili S., and Alhussian H., “Binary optimization using hybrid grey wolf optimization for feature selection,” IEEE Access, vol. 7, pp. 39496–39508, 2019. [Google Scholar]
  • [27].Kabir M. M., Shahjahan M., and Murase K., “A new local search based hybrid genetic algorithm for feature selection,” Neurocomputing, vol. 74, no. 17, pp. 2914–2928, Oct. 2011. [Online]. Available: http://www.sciencedirect.com/science/article/pii/S0925231211002748 [Google Scholar]
  • [28].Bello R., Gomez Y., Nowe A., and Garcia M. M., “Two-step particle swarm optimization to solve the feature selection problem,” in Proc. 7th Int. Conf. Intell. Syst. Design Appl. (ISDA), Oct. 2007, pp. 691–696. [Google Scholar]
  • [29].Şenel F. A., Gokçe F., Yuksel A. S., and Yigit T., “A novel hybrid PSO GWO algorithm for optimization problems,” Eng. Comput., vol. 35, no. 4, pp. 1359–1373, Dec. 2019, doi: 10.1007/s00366-018-0668-5. [DOI] [Google Scholar]
  • [30].Karakonstantis I. and Vlachos A., “Bat algorithm applied to continuous constrained optimization problems,” J. Inf. Optim. Sci., vol. 5, pp. 1–19, Mar. 2020, doi: 10.1080/02522667.2019.1694740. [DOI] [Google Scholar]
  • [31].Simon D., “Biogeography-based optimization,” IEEE Trans. Evol. Comput., vol. 12, no. 6, pp. 702–713, Dec. 2008, doi: 10.1109/tevc.2008.919004. [DOI] [Google Scholar]
  • [32].Mirjalili S., Mirjalili S. M., and Hatamlou A., “Multi-verse optimizer: A nature-inspired algorithm for global optimization,” Neural Comput. Appl., vol. 27, no. 2, pp. 495–513, Feb. 2016, doi: 10.1007/s00521-015-1870-7. [DOI] [Google Scholar]
  • [33].Samareh Moosavi S. H. and Khatibi Bardsiri V., “Satin bowerbird optimizer: A new optimization algorithm to optimize ANFIS for software development effort estimation,” Eng. Appl. Artif. Intell., vol. 60, pp. 1–15, Apr. 2017, doi: 10.1016/j.engappai.2017.01.006. [DOI] [Google Scholar]
  • [34].Fister I. Jr, Yang X.-S., Fister I., and Brest J., “Memetic firefly algorithm for combinatorial optimization,” 2012, arXiv:1204.5165. [Online]. Available: http://arxiv.org/abs/1204.5165
  • [35].Li K., Fang Y., Li W., Pan C., Qin P., Zhong Y., Liu X., Huang M., Liao Y., and Li S., “CT image visual quantitative evaluation and clinical classification of coronavirus disease (COVID-19),” Eur. Radiol., vol. 30, no. 8, pp. 4407–4416, Mar. 2020, doi: 10.1007/s00330-020-06817-6. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [36].Chung M., Bernheim A., Mei X., Zhang N., Huang M., Zeng X., Cui J., Xu W., Yang Y., Fayad Z. A., Jacobi A., Li K., Li S., and Shan H., “CT imaging features of 2019 novel coronavirus (2019-nCoV),” Radiology, vol. 295, no. 1, pp. 202–207, Apr. 2020, doi: 10.1148/radiol.2020200230. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [37].Yang R., Li X., Liu H., Zhen Y., Zhang X., Xiong Q., Luo Y., Gao C., and Zeng W., “Chest CT severity score: An imaging tool for assessing severe COVID-19,” Radiol., Cardiothoracic Imag., vol. 2, no. 2, Apr. 2020, Art. no. e200047, doi: 10.1148/ryct.2020200047. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [38].Shi H., Han X., Jiang N., Cao Y., Alwalid O., Gu J., Fan Y., and Zheng C., “Radiological findings from 81 patients with COVID-19 pneumonia in wuhan, China: A descriptive study,” Lancet Infectious Diseases, vol. 20, no. 4, pp. 425–434, Apr. 2020. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [39].Bai H. X., Hsieh B., Xiong Z., Halsey K., Choi J. W., Tran T. M. L., Pan I., Shi L.-B., Wang D.-C., Mei J., Jiang X.-L., Zeng Q.-H., Egglin T. K., Hu P.-F., Agarwal S., Xie F., Li S., Healey T., Atalay M. K., and Liao W.-H., “Performance of radiologists in differentiating COVID-19 from viral pneumonia on chest CT,” Radiology, vol. 12, Mar. 2020, Art. no.200823, doi: 10.1148/radiol.2020200823. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [40].Chua F., “The role of CT in case ascertainment and management of COVID-19 pneumonia in the UK: Insights from high-incidence regions,” Lancet Respiratory Med., vol. 8, no. 5, pp. 438–440, May 2020, doi: 10.1016/s2213-2600(20)30132-6. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [41].Wu X., Hui H., Niu M., Li L., Wang L., He B., Yang X., Li L., Li H., Tian J., and Zha Y., “Deep learning-based multi-view fusion model for screening 2019 novel coronavirus pneumonia: A multicentre study,” Eur. J. Radiol., vol. 128, Jul. 2020, Art. no. 109041. [Online]. Available: http://www.sciencedirect.com/science/article/pii/S0720048X20302308 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [42].Ardakani A. A., Kanafi A. R., Acharya U. R., Khadem N., and Mohammadi A., “Application of deep learning technique to manage COVID-19 in routine clinical practice using CT images: Results of 10 convolutional neural networks,” Comput. Biol. Med., vol. 121, Jun. 2020, Art. no. 103795. [Online]. Available: http://www.sciencedirect.com/science/article/pii/S0010482520301645 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [43]., “Clinically applicable AI system for accurate diagnosis, quantitative measurements, and prognosis of COVID-19 pneumonia using computed tomography,” Cell, vol. 181, no. 6, pp. 1423–1433, Jun. 2020, doi: 10.1016/j.cell.2020.04.045. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [44].Panwar H., Gupta P. K., Siddiqui M. K., Morales-Menendez R., and Singh V., “Application of deep learning for fast detection of COVID-19 in X-rays using nCOVnet,” Chaos, Solitons Fractals, vol. 138, Sep. 2020, Art. no. 109944. [Online]. Available: http://www.sciencedirect.com/science/article/pii/S096007792030343X [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [45].Butt C., Gill J., Chun D., and Babu B. A., “Deep learning system to screen coronavirus disease 2019 pneumonia,” Int. J. Speech Technol., vol. 1, p. 22, Apr. 2020, doi: 10.1007/s10489-020-01714-3. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [46].Nour M., Cömert Z., and Polat K., “A novel medical diagnosis model for COVID-19 infection detection based on deep features and Bayesian optimization,” Appl. Soft Comput., vol. 46, Jul. 2020, Art. no.106580, doi: 10.1016/j.asoc.2020.106580. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [47].Hu S., Gao Y., Niu Z., Jiang Y., Li L., Xiao X., Wang M., Fang E. F., Menpes-Smith W., Xia J., Ye H., and Yang G., “Weakly supervised deep learning for COVID-19 infection detection and classification from CT images,” IEEE Access, vol. 8, pp. 118869–118883, 2020, doi: 10.1109/access.2020.3005510. [DOI] [Google Scholar]
  • [48].Elaziz M. A., Hosny K. M., Salah A., Darwish M. M., Lu S., and Sahlol A. T., “New machine learning method for image-based diagnosis of COVID-19,” PLoS ONE, vol. 15, no. 6, Jun. 2020, Art. no. e0235187, doi: 10.1371/journal.pone.0235187. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [49].Zhao J., Zhang Y., He X., and Xie P., “COVID-CT-Dataset: A CT Scan Dataset about COVID-19,” 2020, arXiv:2003.13865. [Online]. Available: https://arxiv.org/abs/2003.13865
  • [50].Naudé W., “Artificial intelligence vs COVID-19: Limitations, constraints and pitfalls,” AI Soc., vol. 35, no. 3, pp. 761–765, Sep. 2020, doi: 10.1109/access.2019.2955983. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [51].Hassib E. M., El-Desouky A. I., El-Kenawy E.-S. M., and El-Ghamrawy S. M., “An imbalanced big data mining framework for improving optimization algorithms performance,” IEEE Access, vol. 7, pp. 170774–170795, 2019, doi: 10.1109/access.2019.2955983. [DOI] [Google Scholar]
  • [52].Hassib E. M., El-Desouky A. I., Labib L. M., and El-kenawy E.-S.-M., “WOA + BRNN: An imbalanced big data classification framework using whale optimization and deep neural network,” Soft Comput., vol. 24, no. 8, pp. 5573–5592, Mar. 2019, doi: 10.1007/s00500-019-03901-y. [DOI] [Google Scholar]
  • [53].Mirjalili S., Mirjalili S. M., Saremi S., and Mirjalili S., Whale Optim. Algorithm: Theory, Literature Rev., Appl. Designing Photonic Crystal Filters. Cham: Springer International Publishing, 2020, pp. 219–238, doi: 10.1007/978-3-030-12127-3_13. [DOI] [Google Scholar]
  • [54].Cuevas E., Fausto F., and González A., Metaheuristics and Swarm Methods: A Discussion on Their Performance and Applications. Cham, Switzerland: Springer, 2020, pp. 43–67, doi: 10.1007/978-3-030-16339-6_2. [DOI] [Google Scholar]
  • [55].Fausto F., Reyna-Orta A., Cuevas E., Á.Andrade G., and Perez-Cisneros M., “From ants to whales: Metaheuristics for all tastes,” Artif. Intell. Rev., vol. 53, no. 1, pp. 753–810, Jan. 2020. [Google Scholar]
  • [56].Salimi H., “Stochastic fractal search: A powerful metaheuristic algorithm,” Knowl.-Based Syst., vol. 75, pp. 1–18, Oct. 2015. [Online]. Available: http://www.sciencedirect.com/science/article/pii/S0950705114002822 [Google Scholar]
  • [57].El-Kenawy E.-S. M., Eid M. M., Saber M., and Ibrahim A., “MbGWO-SFS: Modified binary grey wolf optimizer based on stochastic fractal search for feature selection,” IEEE Access, vol. 8, no. 1, pp. 107635–107649, 2020, doi: 10.1109/access.2020.3001151. [DOI] [Google Scholar]
  • [58].El-Kenawy E.-S. and Eid M., “Hybrid gray wolf and particle swarm optimization for feature selection,” Int. J. Innov. Comput., Inf. Control, vol. 16, no. 3, pp. 831–844, 2020. [Google Scholar]
  • [59].Ibrahim A., Noshy M., Ali H. A., and Badawy M., “PAPSO: A power-aware VM placement technique based on particle swarm optimization,” IEEE Access, vol. 8, pp. 81747–81764, 2020, doi: 10.1109/access.2020.2990828. [DOI] [Google Scholar]
  • [60].Gaber T., Tharwat A., Ibrahim A., Snael V., and Hassanien A. E., “Human thermal face recognition based on random linear oracle (RLO) ensembles,” in Proc. Int. Conf. Intell. Netw. Collaborative Syst., Sep. 2015, pp. 1–5, doi: 10.1109/incos.2015.67. [DOI] [Google Scholar]
  • [61].Ibrahim A., Tharwat A., Gaber T., and Hassanien A. E., “Optimized superpixel and AdaBoost classifier for human thermal face recognition,” Signal, Image Video Process., vol. 12, no. 4, pp. 711–719, Nov. 2017, doi: 10.1007/s11760-017-1212-6. [DOI] [Google Scholar]

Articles from Ieee Access are provided here courtesy of Institute of Electrical and Electronics Engineers

RESOURCES