Skip to main content
Elsevier - PMC COVID-19 Collection logoLink to Elsevier - PMC COVID-19 Collection
. 2021 Jun 16;176:234–248. doi: 10.1016/j.comcom.2021.06.011

Efficient deep neural networks for classification of COVID-19 based on CT images: Virtualization via software defined radio

Saman Fouladi a, MJ Ebadi b, Ali A Safaei a, Mohd Yazid Bajuri c, Ali Ahmadian d,
PMCID: PMC8205564  PMID: 34149118

Abstract

The novel 2019 coronavirus disease (COVID-19) has infected over 141 million people worldwide since April 20, 2021. More than 200 countries around the world have been affected by the coronavirus pandemic. Screening for COVID-19, we use fast and inexpensive images from computed tomography (CT) scans. In this paper, ResNet-50, VGG-16, convolutional neural network (CNN), convolutional auto-encoder neural network (CAENN), and machine learning (ML) methods are proposed for classifying Chest CT Images of COVID-19. The dataset consists of 1252 CT scans that are positive and 1230 CT scans that are negative for COVID-19 virus. The proposed models have priority over the other models that there is no need of pre-trained networks and data augmentation for them. The classification accuracies of ResNet-50, VGG-16, CNN, and CAENN were obtained 92.24%, 94.07%, 93.84%, and 93.04% respectively. Among ML classifiers, the nearest neighbor (NN) had the highest performance with an accuracy of 94%.

Keywords: Computed tomography, ResNet-50, VGG-16, Convolutional neural networks (CNN), Convolutional auto-encoder neural network (CAENN), COVID-19

1. Introduction

Coronavirus named Severe Acute Respiratory Syndrome-Corona Virus-2 (SARS-Cov-2) appeared in Wuhan at the end of 2019 [1]. Coronavirus may cause serious respiratory disease, including Acute Respiratory Distress Syndrome (ARDS), which can be fatal [2]. The most common symptoms that patients experience after being infected include fever, fatigue, cough, and respiratory distress. The degree of illness ranges from mild influenza to multi-organ failure, and even death depending on the patient’s immune system response to infection. As of April 2021, there was a serious outbreak of COVID-19 virus across the world, resulting in 141,057,106 confirmed cases and 3,015,043 confirmed deaths. This virus has been declared as a pandemic by the World Health Organization (WHO) after causing a failure in the health services of more than 200 countries due to the lack of medical staff and personnel [3].

Reverse transcription-polymerase chain reaction (RT-PCR), computed tomography (CT), Gene sequencing, and X-ray based imaging methods are all among the widely used tests for screening the patients with COVID-19 [4], [5]. Radiological screening, such as chest CT scans or chest X-ray can help us closely monitor the disease and quickly isolate infected people [6]. CT-based imaging is inexpensive, reliable, practical, and available in all hospitals as an efficient tool for diagnosis, prediction, and follow-up of the patients with COVID-19 [7]. According to the wide range of studies conducted around the globe, the sensitivity of X-ray images is less than CT scans in the detection of patients with COVID-19 [8]. However, the number of patients and the spread rate of the disease are too high for experts to deal with. It has been observed in previous studies that manual examination in hospitals is a time-consuming process [9]. Thus, intelligent technologies will have a great contribution to disease diagnosis. Hence, artificial intelligence (AI) methods for making exact and early diagnoses have attracted the attention of many researchers. These methods make significant contributions to a wide range of fields from manufacturing to the healthcare industry [10]. Subsequent to the appearance of deep learning, AI methods have entered a new age. Deep learning (DL) structures learn the features and classify them automatically by providing an adequate amount of training examples. CNN architectures of high discriminative capacity are primarily used for image processing and reshaping medical imaging. In the automated processing of radiological images, DL techniques are commonly used [11].

Numerous CNN methods of high-efficiency and high-performance have been used for medical studies in the literature. By increasing the number of COVID-19 samples in hospitals, CT and X-ray images have been shared. However, these methods were no longer so efficient for the existing COVID-19 datasets with unequal distribution of insufficient samples. Therefore, radiology experts have turned their attention to approaches that do not need such close supervision or data augmentation. In spite of the satisfactory results of new approaches, none of them is perfect in achieving the ultimate result.

Here, the recent studies using CT scans for classifying the patients with COVID-19 and healthy controls are summarized. We examined only articles that have used various types of CNNs including pre-trained networks. Several statistical tests such as accuracy, precision, sensitivity, specificity, F1 ranking, and so forth are used to evaluate the two-class classification results of CNNs. The studies were divided into three categories: articles that have used several pre-trained CNN, articles with one pre-trained CNN, and articles that have used CNN or other methods. Many researchers investigate feature selection such as [27], [28], [29], [30], [31], [32], [33]. In articles with more than one network, the maximum metrics of each network were considered. The findings of this survey are summarized in Table 1.

Table 1.

Some new related studies.

Reference Method Dataset Results
[12] The pre-trained Xception network, the pre-trained ResNet50V2 network and a proposed pre-trained CNN COVID-19: 48260
Non-COVID: 15589
Accuracy (Acc.)
proposed CNN
98.49%
Acc.
Xception
96.55%

[13] Sixteen pre-trained CNNs such as: SqueezeNet,
GoogLeNet, Inception-v3, DenseNet-201,
ResNet and etc.
COVID-19: 216
Non-COVID: 397
Acc. 95.97
MobileNet-v2
Sensitivity (Sens.) 98.99
ResNet-18
Specificity (Spec.) 96.67
DenseNet-201
F1-score 0.96
MobileNet-v2
ShuffleNet
ResNet-18
DenseNet-201

[14] AlexNet, VGG-COVID-19: ResNet-50, GoogLeNet, and 334 Net (VGG19Net and VGG16Net) are among the CNN versions. COVID-19: 334
Non-COVID: 794
Acc. 0.79
alexnet
Prec. 0.8475
GoogLeNet
Sens. 0.95
VGG16Net
Spec. 0.92
GoogLeNet

[15] Several pre-trained CNN such as: VGG-19,
VGG-16, DenseNet-169, ResNet-50, CTnet-10,
InceptionV3
COVID-19: 349
Non-COVID: 216
Acc. 0.945
VGG-19

[16] Pre-trained CNNs such as AlexNet, GoogleNet, ResNet-18, ShuffleNet COVID-19: 347
Non-COVID: 397
Acc. 0.7829
ResNet-18
Prec. 0.81
ResNet-18
Sens. 0.769
ResNet-18
Spec. 0.799
ResNet-18
F1-score 0.789
ResNet-18

[17] The ResNet-50 neural network COVID-19: 684 Non-COVID: 254 Acc. 98%

[18] The VGG19 neural network COVID-19: 349
Non-COVID: 397
Prec. 84%
Sens. 81%
F1-score 83%

[19] Pre-trained DenseNet201 based on DTL COVID-19: 1252
Non-COVID: 1230
Acc. 96%
Prec. 96%
Sens. 96%
F1-score 96%

[20] Fine-tuned inception-v3 network with multiple-way data augmentation Total data after Augmentation: 3,724 Acc. 0.995
Prec. 0.992
Sens. 0.998
Spec. 0.982
F1-score 0.995

[21] COVID-19 detection using a 3D CNN COVID-19: 540
Non-COVID: 229
Acc. 0.901
Sens. 0.907
Spec. 0.911

[22] The UNet++ neural network Total data: 35355 images Acc. 92.59%
Sens. 100%
Spec. 81.82%

[23] The GoogleNet Inception v3 Total data: 1065 images Acc. 89.5%
Sens. 88%
Spec. 87%

[24] FGCNet is the best model of the eight
suggested networks, showing the
convergence of GCN and CNN networks
with multiple-way data augmentation.
COVID-19: 113
Non-COVID: 209
Acc. 0.9714
Prec. 0.9661
Sens. 0.9771
Spec. 0.9656

[25] The network was made up of two branches, where the upper branch a lightweight design having four different layers of convolution, and another branch is known lower which was made up of blocks having denser connections to represent the learning. COVID-19: 349
Non-COVID: 397
and
Non-COVID: 1230
COVID-19: 1252
Acc. 0.9083
Prec. 0.9575
Recall (Rec.) 0.8589
F1-score 0.9087

[26] CNN COVID-19: 68
Non-COVID: 64
Acc. 0.9250
Sens. 0.9025
F1-score 0.985

Rahimzadeh et al. [12] used the ResNet50V2 network, the Xception network, and a proposed CNN for the detection of COVID-19 from lung HRCT scans. The dataset was composed of 15589 images of normal people and 48260 images of infected with COVID-19. In this study, an image processing algorithm was proposed to filter the proper CT scans of the patients, showing inside their lungs. By using this algorithm, network accuracy and speed were increased. After training the three networks, the trained networks were used for running a full-automated system of COVID-19 identification. The system was studied on two different datasets: one with more than 7796 images and the other with 41892 images of different thicknesses from almost 245 patients. The model showed an overall accuracy of 98.49% for single image classification.

Also, Pham in [13] has presented a study of sixteen pre-trained CNNs for COVID-19 classification. In this study, higher classification rates were obtained by using transfer learning instead of data augmentation. The dataset was randomly divided into 80% and 20% respectively for training and testing data with the highest accuracy of 95% achieved by MobileNet-v2, the highest sensitivity of 98% achieved by ResNet-18, the highest specificity of 96% achieved by DenseNet-201, and the highest F1-score of 96% achieved by MobileNet-v2. There were four separate networks of MobileNet-v2, ShuffleNet, ResNet-18, and DenseNet-201.

El-Kenawy et al. [14] have focused of experiments on three scenarios to evaluate accuracy and performance of the suggested framework for the classification of COVID-19. In scenario I accuracies of many CNN models were compared on CT images from the COVID-19 dataset. The highest classification accuracy was reported 79% for AlexNet model. The highest precision of 84%, the highest sensitivity of 95%, the highest specificity of 92%, and the highest F1-score of 77% were respectively reported for GoogLeNet, VGG16Net, GoogLeNet, and AlexNet models.

For the COVID-19 diagnosis, a self-developed model was built, called CTnet-10.

Shah et al. in [15], has proposed the CTnet-10 model that composed of four convolutional blocks. It passed through two convolutional blocks with 126 × 126 × 32 and 124 × 124× 32 dimensions. Then it went into a 62 × 62 × 32 max-pooling followed by two convolutional layers with dimensions of 60 × 60 × 32 and 58 × 58 × 32. It then passed by a pooling layer of 29 × 29 × 32. After passing a fully connected layer of size 256, data were classified into negative or positive classes of COVID-19. The accuracy obtained from CTnet-10 model was 82%. VGG-19, InceptionV3, ResNet-50, VGG-16, and DenseNet-169 were some of the other models tested. The highest accuracy of 94% was obtained for the VGG-19 model. Due to the small dataset of COVID-19, it was possible to use pre-trained neural networks for COVID-19 (+) or COVID-19 (-) classification.

Attallah et al. in [16] have proposed a new computer-aided diagnosis (CAD) system called MULTI-DEEP, on the base of merger of multiple NCs to distinguish COVID-19 from other cases. There were four main scenarios in the framework. In scenario I and scenario II, the dataset of 397 normal chest CT images and 347 chest CT images of COVID-19 positive cases were classified. In scenario I, four pre-trained CNNs were applied to diagnose COVID-19 and non-COVID-19 cases. The highest performance was obtained for ResNet-18 with 78% accuracy, 76% sensitivity, 79% specificity, 81% precision, and 78% F1-score. In scenario II, the authors extracted deep features from every pre-trained CNN to be used for training SVM classifiers. Compared to other CNNs, ResNet-18 network resulted in the highest performance on deep features with 92% accuracy, 93% sensitivity, 91% specificity, 91% precision, and 92% F1-score.

In [17], Serte et al. have used the ResNet-50 deep learning model to predict COVID-19 on each CT image of 3D CT scans. This model fused image-level predictions to diagnose COVID-19 on a 3D CT volume. This dataset was composed of 3D CT scans of the patients, each comprised about 40 axial slices. This dataset included 1110 3D CT scans of patients taken in hospitals of Moscow, Russia. The Mosmed-1110 dataset consisted of five categories of 3D CT volumes. These groups were named CT0, CT1 to CT4. The CT0 was comprised of 254 3D CT volumes that were normal scans. CT2 consisted of 684 3D CT scans showing COVID-19 infection on the lungs. First, middle axial lung slices were selected and then each of them passed through ResNet-50 model. The classification accuracy was obtained 98% from this method.

Horry et al. [18] first have selected the VGG19 model and then extensively tuned to appropriate parameters in order to be performed at different levels of COVID-19 detection from pneumonia or normal cases for all three modes of lung images. This led to the precision of 84%, recall of 81%, and F1-score of 83%. In this study, the dataset consisted of 349 COVID-19 and 397 Non-COVID-19 images.

DenseNet201 on the base of deep transfer learning (DTL) model was utilized in [21] by Jaiswal et al. to identify whether patients were COVID-infected or not. The model achieved an accuracy of 96%, F1-score of 96%, recall of 96%, and precision of 96%.

Several datasets are needed to build a multi-task pipeline for coronavirus detection, prediction, and classification. For this goal, El-Bana et al. [22] used three datasets including two of the most popular datasets: (1) RSNA Pneumonia Detection Challenge Dataset, (2) COVID-19 Image Data Collection Repository, and (3) COVID-19 CT Segmentation Dataset. All scans were reshaped to 512 × 512 × 3. Due to the small number of CT volumes, data augmentation techniques were used such as rotation, vertical and horizontal transformations, shearing, and zooming. The total images after data augmentation were 3,724. For fine-tuning, the Inception-v3 model was used for multi-label classifiers and multi-class classification. The multi-class classification resulted in 99% accuracy, 99% precision, 99% sensitivity, 98% specificity, and 99% F1-score.

Wang et al. [21] have built a deep learning system, using 3D volumes of CT for lesion localization and COVID-19 classification. A pre-trained UNet network was utilized for the segmentation of the lung area. The COVID-19 lesions were classified after the segmentation of 3D lung area fed into a 3D deep neural network (DNN) to predict the risk of COVID-19 infection. The classification specificity, sensitivity, and accuracy were obtained at 91%, 90%, and 90% respectively.

Chen et al. [22] have constructed a system based on UNet++ neural network for COVID-19 pneumonia detection at high CT resolution. To validate the model, 46,096 anonymous images were retrospectively collected from 106 admitted patients, of which 51 people were laboratory-confirmed cases of COVID-19 pneumonia and 55 people were control cases of other diseases in Renmin Hospital of Wuhan University. After filtering out images with good lung conditions, 35,355 images remained and were divided into retrospectively training and testing datasets. By applying findings of the radiologists as the archetype, an accuracy of 92.59%, sensitivity of 100%, and specificity of 81.82% were attained by the model per patient in the 27 prospective cases.

Wang et al. [23] have used 1065 CT images collected from 259 patients with the cohort including 180 patients of typical viral pneumonia and 79 patients of confirmed SARS-COV-2 by nucleic acid testing in three hospitals. Three main processes constituted the architecture: first, input image preprocessing; second ROI image feature-derivation and training; third classification by two fully connected layers and prediction by binary classifiers. Transfer learning was also performed involving a predefined model trained by the use of well-known GoogleNet Inception v3 CNN. The number of picture types in the training set was equal to the total number of 320 images. The remaining CT images were considered for internal validation. The internal validation resulted in a total accuracy of 89.5%, a sensitivity of 87%, and a specificity of 88%. Moreover, total accuracy of 79.3%, a sensitivity of 67%, and a specificity of 83% were reported for the external testing dataset.

Wang et al. [24] have used extracted characteristics of an auto-created CNN to learn individual representations at the image level. The CNN employed several new techniques such as average pooling by rank and data augmentation in several directions. Relational representations were learned using the Graphic Convolution Network (GCN). The Deep Merge of Functions (DFF) was developed to merge the individual features at the image level and the relational features of CCN and GNN. The best template was called FGCNet. The model achieved 97% accuracy, 96% precision, 97% sensitivity, and 96% specificity.

Wang et al. [25] have proposed a new collaborative learning framework to accurately identify COVID-19 by effective learning of heterogeneous datasets with distribution gaps.

The network was made up of two branches, where the upper branch with a lightweight design had four different layers of convolution, and another branch was composed of blocks of denser connections for learning. In this analysis, COVID-CT and SARS-CoV-2 were used to test a joint learning system utilizing two common CT datasets of COVID-19. The dataset was composed of 2482 CT images from 120 patients, of whom 1230 were non-COVID with different lung infection symptoms and 1252 were COVID-19. The clinical results of 397 CT images from 171 patients without COVID-19 and 349 CT images from 216 patients with COVID-19 were used in the COVIDCT dataset. Evaluation metrics were computed for both datasets. The achieved results of the SARS-CoV-2 dataset were better than those of the COVID-CT dataset. The overall accuracy was estimated at 90% with a recall of 85%, precision of 95%, and F1-score of 90%.

Singh et al. [26] have used a CNN for classification of positive COVID-19 and negative COVID19 cases. To extract useful features, CNN utilized several convolutional and pooling layers. For experimental purposes, different ratios of training to testing datasets were considered, such as 20:80, 30:70, 40:60, 50:50, 60:40, 70:30, 80:20, and 90:10. The accuracy, F1-score, and sensitivity of CNN models were obtained by 92%, 98%, and 90%, respectively.

The main objective of this study is to propose DL networks such as ResNet-50, VGG-16, CNN, and deep CAENN to differentiate between COVID-19 and Non-COVID using lung CT scans. The efficiency of the proposed models is compared to that of the current deep architectures on a typically accessible lung CT dataset. For neural networks of CNN and convolutional Auto-Encoder, we propose an optimal architecture and change the parameters and hyper parameters of optimization function, learning rate, and activation function. The optimal parameters of ResNet-50 and VGG-19 for training are changed but their architectures are not. The classification is done by the use of original images of dataset but no use of data augmentation techniques.

The followings are the key contributions of this article:

  • Two efficient COVID-19 classification models have been developed on the base of ResNet-50 and VGG-16 neural networks.

  • Two efficient COVID-19 classification models have been developed on the base of CNN and CAENN.

  • The proposed models have been implemented on a big dataset including COVID-19 and Non-COVID-19 CT Scans.

The other sections of the paper are: The study in the area of deep learning neural networks and ML classifier for chest CT image classification is given in Section 2. Section 3 describes the proposed models. Section 3 is dedicated to the investigational findings and discussions. The conclusion is drawn in Section 4.

2. Material and methods

2.1. Data acquisition

For this experiment, the ’COVID-19-Dataset’ CT scan dataset was used retrieved from Kaggle.1 The data set includes a total of 2482 CT scans, composed of 1230 negative CT scans and 1252 positive CT scans for SARS-CoV-2 (COVID-19) infection. Soares et al. [34] collected these images from hospitals in Sao Paulo, Brazil. This dataset is freely available in Kaggle. Fig. 1 shows a selection of CT scans from the dataset of non-COVID-19 and COVID-19 patients. The existing dataset of images must be resized before being fed to the convolutional network. All the CT images have been resized to 173 100 pixels for width and height, with a depth of 3.

Fig. 1.

Fig. 1

Sample of (a) COVID-19 and (b) Non-COVID CT scans.

2.2. Model architecture and model training

Deep learning-driven models have recently been proven to be the winner in many clinical trials. These models outperform conventional mathematical models in terms of hand-crafted features in medical image processing and machine vision issues. CNNs are a form of DL technique in which several layers are strongly trained. These networks have very efficient and popular applications in classification, image processing, and neural computer vision [32]. A CNN network is made up of three major layers: convolution, pooling, and fully connected layers. Different layers function in different ways which leads to the ultimate learning. This network can be applied either alone or alongside other networks for data classification.

Features of chest CT scans are used to accurately classify into two classes of COVID-19 and non-COVID-19. The following steps are performed to classify COVID-19-infected person using the CNN-based proposed models:

I. Feature extraction

In this step, CNN uses several convolutions and a pooling layer to monitor and evaluate possible features. Fig. 2 shows how the stride may be used by the kernel/filter to extract possible features. After that, the max pooling layer is used to reduce the spatial size of the Convolved features. It has capable of overcoming the overfitting problem. It assesses the maximum of the region from the feature map created by the convolution operator. Fig. 3 shows a max-pooling layer with a kernel size of 2 and a stride of 2. The rectified linear activation function (ReLU) is a piecewise linear function that will output the input directly if it is positive, otherwise, it will output zero. It has become the default activation function for many types of neural networks because a model that uses it is easier to train and often achieves better performance.

Fig. 2.

Fig. 2

CNN operator with kernel size 3 and stride of 1.

Fig. 3.

Fig. 3

Max pooling with a single pooled feature.

II. Classification In this step, fully connected layers classify extracted features and assess the probability of the object in the input image. Typically, an activation function and a dropout layer are used to establish non-linearity and reduce overfitting. As shown in Fig. 4, the fully connected layer classifies features into two classes.

Fig. 4.

Fig. 4

Fully Connected layer.

2.2.1. ResNet neural network

The ResNet is a type of CNN often used in the field of computer vision since winning the 2015 ILSVRC competition [34]. With a depth of 152 layers, this architecture was named the deepest at that time. Deep ResNet operate similar to deep CNNs but they implement a residual connection between each layer and the output of the following layer. This requires that each layer receives features as input from its two previous layers. The residual connections perform better upon regular neural networks in two ways. First, they reduce the vanishing gradient problem by allowing the use of a different path for gradient flow. Second, they allow the model to learn referenced functions which ensures deeper layers will execute either better or as good as shallower layers [34]. In this paper, ResNet-50 has been used and Fig. 5 describes the architecture of this neural network. “ID BLOCK” in the diagram stands for “Identity block”, and “ID BLOCK × 3” means stack 3 identity blocks together.

Fig. 5.

Fig. 5

The architecture of the ResNet-50 network.

For this network, from the total number of 2482 data, 90% (2233 numbers) were used as training data and 10% (249 numbers) as test data. The activation for all layers except the last layer was Relu function. Adam with a learning rate of 0.0001 was selected for the optimization function. This network was trained over 100 training epochs and data were transmitted to the network in batches of 8 size (batch-size). The duration of each epoch was the 20 s.

VGG-16 Neural Network

VGG-16 is a CNN model, also called the OxfordNet model, proposed by K. Simonyan and A. Zisserman from the University of Oxford in the paper “Very Deep Convolutional Networks for Large-Scale Image Recognition”. It was one of the famous models submitted to ILSVRC-2014. Number 16 refers that it has a total of 16 layers that has some weights [35]. The architecture for VGG16 network is shown in Fig. 6. From the total number of 2482 data, 90% (2233 numbers) were used as training data and 10% (249 numbers) as test data. The input to the main VGG 16 model is images of 224 × 224 × 3 pixels changed to 173 × 100 × 3 in this study. Then, there are two convolution layers of 70 × 70 × 64 size, two convolution layers of 128 filter length, three convolution layers of 256 filter length, and three convolution layers of 512 filter length. Then again there are three convolution layers of 512 filter length. The Kernel size is 3 × 3 and the pool size is 2 × 2 with step 2 × 2 for all the layers. The output of the last pooling layer is 2D which must be converted into a 1D layer to be sent to fully connected layers, done by a flatten layer. After the convolution layers, two 4096 fully connected layers and two fully connected layers were used to classify data into two classes by softmax activation function. The activation for all layers except the last layer was Relu function. Adam with a learning rate of 0.0001 was selected for the optimization function. Also, the same padding was used for all convolution layers. This network was trained over 100 training epochs and data were transmitted to the network in batches of 32 size (batch-size). The duration of each epoch was 15 s.

Fig. 6.

Fig. 6

The architecture of the VGG-16.

2.2.2. Convolutional neural network

Fig. 7 presents the proposed architecture for a two-dimensional CNN. From the total number of 2482 data, 80% (1985 numbers) and 20% (497 numbers) were used respectively as training and testing data. The network includes 2 convolution layers of 128 filter length, 2 convolution layers of 64 filter length, 2 convolution layers of 32 filter length, 2 convolution layers of 16 filter length, and 2 convolution layers of 8 filter length. The 22 kernel function was considered for all layers. In general, a convolution network is a hierarchical neural network in which the layers of convolution are interconnected with alternate pooling layers and then with several fully connected layers. However, there is no compulsion to utilize a pooling layer after each layer of convolution. It is observable in the figure that the network consists of 10 convolution layers but 5 pooling layers. In this architecture, the 22 max-pooling layer is considered with step 2. The last pooling layer output is 2D which must be converted into a 1D layer to be sent to fully connected layers, done by a flatten layer. Also, one type of padding must be used to control the output size of each convolution layer. For all the networks in this study, the same padding has been used to fill the edges of input data with similar values in adjacent cells. After the convolution layers, a 128 fully connected layer and 2 fully connected layers were used to classify data into two classes by softmax activation function. In order to prevent overfitting, batch normalization layers were considered after all convolution layers and a dropout layer with a rate of 0.1 is considered after the first fully connected layer. The activation function used for this network was Relu function excluding the last fully connected layer.

Fig. 7.

Fig. 7

The architecture of the proposed CNN.

Extracted features from the convolution layers are the input for the first fully connected layer to Ufc=128 hidden layers. The output of flatten layer equals 358=120. Thus, the number of weights is Wconv = Outflatten Ufc=128120 = 15,360 and the number of existing parameters to the second fully connected layer is 15,360 + 128 (biases) = 15,488. Table 2 illustrates the learning parameters of this network. The Sum of all the used parameters can be calculated from the sum of the values of the Param column from Table 2. The resulting sum is 150,578 of which 149,586 numbers are associated with learning and 992 numbers with non-learning parameters.

Table 2.

Parameters used in the convolution network for classification in 2 classes.

Layer (type) Output shape Param #
conv2d_1 (Conv2D) (None, 100, 173, 128) 1664
batch_normalization_1 (Batch) (None, 100, 173, 128) 512
conv2d_2 (Conv2D) (None, 100, 173, 128) 65664
batch_normalization_2 (Batch) (None, 100, 173, 128) 512
max_pooling2d_1 (MaxPooling2D) (None, 50, 86, 128) 0
conv2d_3 (Conv2D) (None, 50, 86, 64) 32832
batch_normalization_3 (Batch) (None, 50, 86, 64) 256
conv2d_4 (Conv2D) (None, 50, 86, 64) 16448
batch_normalization_4 (Batch) (None, 50, 86, 64) 256
max_pooling2d_2 (MaxPooling2D) (None, 25, 43, 64) 0
conv2d_5 (Conv2D) (None, 25, 43, 32) 8224
batch_normalization_5 (Batch) (None, 25, 43, 32) 128
conv2d_6 (Conv2D) (None, 25, 43, 32) 4128
batch_normalization_6 (Batch) (None, 25, 43, 32) 128
max_pooling2d_3 (MaxPooling2D) (None, 12, 21, 32) 0
conv2d_7 (Conv2D) (None, 12, 21, 16) 2064
batch_normalization_7 (Batch) (None, 12, 21, 16) 64
conv2d_8 (Conv2D) (None, 12, 21, 16) 1040
batch_normalization_8 (Batch) (None, 12, 21, 16) 64
max_pooling2d_4 (MaxPooling2D) (None, 6, 10, 16) 0
conv2d_9 (Conv2D) (None, 6, 10, 8) 520
batch_normalization_9 (Batch) (None, 6, 10, 8) 32
conv2d_10 (Conv2D) (None, 6, 10, 8) 264
batch_normalization_10 (Batch) (None, 6, 10, 8) 32
max_pooling2d_5 (MaxPooling2D) (None, 3, 5, 8) 0
Flatten_1 (Flatten) (None, 120) 0
Dense_1 (Dense) (None, 128) 15488
Dropout_1 (Dropout) (None, 128) 0
dense_2 (Dense) (None, 2) 258

Total params: 150,578
Trainable params: 149,586
Non-trainable params: 992

To select an optimization function, SGD, Adadelta, and Adam functions were investigated. The results of using Adam function were significantly better than those of the other two functions. The learning rate parameter was tested with values of 0.001 and 0.0001 while the best value with the least learning error was 0.0001. The network was trained over 100 training epochs and the data was transmitted to the network in batches of 16 size (batch-size). The duration of each epoch was 9 s.

2.2.3. Convolutional auto-encoder neural network

The structure of the designed CAENN for this research can be seen in Fig. 8. This architecture comprises two parts; the first includes a CAENN for training data, and the second includes a simple convolutional network for classification making use of the last encoder layer output of the first part. Out of a total number of 16449 data, 67% (11020 numbers) were used as training data and 33% (5429 numbers) were used as test data for the network. Out of the total number of 2482 data, 80% (1985 numbers) and 20% (497 numbers) were respectively used as training and testing data. The network includes two convolution layers of 128 filter length, two convolution layers of 64 filter length, and one convolution layer of 32 filter length. The decoder part consists of one convolution layer of 32 filter length, 2 convolution layers of 64 filter length, and 2 convolution layers of 128 filter length. The 22 kernel function was considered for all layers. In the encoder, after a sequence of two layers of convolution, a 22 max-pooling layer is considered. In the decoder, after a sequence of two convolution layers, an up-sampling layer with size 22 was considered. To prevent the network from overfitting, the layer of batch normalization was used after each layer of convolution. In the last layer of the encoder part, important features of the input data are extracted and this layer output can be used for the classification part of the network as can be seen in Fig. 5. The output of this layer is trained by two continuous convolution layers with 64 filter length, a kernel function of size 22, and a max-pooling layer of size 22. To avoid overfitting, layers of batch normalization and 0.1 dropout are used. After the Flatten layer, 2 fully connected layers are used to classify data into two classes by softmax activation function. The used activation function for other layers was Relu function in this network. To select an optimization function, SGD, Adadelta, and Adam functions were investigated. The results of using Adam function were significantly better than those of the other two functions. The learning rate parameter was tested with values of 0.001 and 0.0001 while the best value with the least learning error was 0.0001. The network was trained over 100 training epochs and the data was transmitted to the network in batches of 16 size (batch-size). The duration of each epoch was 9 s.

Fig. 8.

Fig. 8

The architecture of the CAENN classification.

For the CAENN, the parameters in the encoder and classifier are important for training and classification. Extracted features from the encoder last layer are trained by several layers of convolution, and the final extracted features become the input of the first fully connected layer to the hidden layer of Ufc=2. The number of weights Wconv depends on the output size of flatten layer and the number of hidden layers in the fully connected layer. The output of flatten layer equals 61164=4224. Thus, the number of weights is Wconv = Outflatten Ufc=42242=8448 and the number of existing parameters to the second fully connected layer is 8448 + 2 (biases) = 8450. Table 3 illustrates the learning parameters of this network. The sum of all the used parameters can be calculated from some of the values of the param column from Table 3. The resulting sum is 159,906, of which 158,946 numbers are associated with learning and 960 numbers with non-learning parameters.

Table 3.

Parameters used in the CAENN for classification in 2 classes.

Layer (type) Output shape Param#
input_1 (InputLayer) (None, 100, 173, 3) 0
conv2d (Conv2D) (None, 100, 173, 128) 1664
batch_normalization (Batch) (None, 100, 173, 128) 512
conv2d_1 (Conv2D) (None, 100, 173, 128) 65664
batch_normalization_1(Batch) (None, 100, 173, 128) 512
max_pooling2d (MaxPooling2D) (None, 50, 87, 128) 0
conv2d_2 (Conv2D) (None, 50, 87, 64) 32832
batch_normalization_2(Batch) (None, 50, 87, 64) 256
conv2d_3 (Conv2D) (None, 50, 87, 64) 16448
batch_normalization_3(Batch) (None, 50, 87, 64) 256
max_pooling2d_1(MaxPooling2D) (None, 25, 44, 64) 0
conv2d_4 (Conv2D) (None, 25, 44, 32) 8224
batch_normalization_4(Batch) (None, 25, 44, 32) 128
max_pooling2d_2(MaxPooling2D) (None, 13, 22, 32) 0
conv8 (Conv2D) (None, 13, 22, 64) 8256
conv2d_11 (Conv2D) (None, 13, 22, 64) 16448
batch_normalization_10(Batch) (None, 13, 22, 64) 256
max7 (MaxPooling2D) (None, 6, 11, 64) 0
dropout (Dropout) (None, 6, 11, 64) 0
flatten (Flatten) (None, 4224) 0
dense (Dense) (None, 2) 8450

Total params: 159,906
Trainable params: 158,946
Non-trainable params: 960

Our entire implementation was done in Keras with Tensorflow backend. The networks in this research were designed in Python environment, implemented using the cross library, and run in Google Colaboratory (Colab) environment. Colab provides a platform for running Python codes, especially ML, deep learning, and data analysis. Colab hardware specifications are listed in Table 4.

Table 4.

Colab hardware specification.

Hardware Description
GPU 1xTesla K80, VRAM of 12GB GDDR5, 2496 CUDA cores, compute 3.7
CPU One core, Two threads, Xeon Processors of 2.3 Ghz
RAM 12.6 GB
Disk 33 GB

3. Experiment results

The current study aims to classify CT scans from patients into COVID-19 and non-COVID-19 categories. Metrics include F1-score, recall, precision, and accuracy. Accuracy is a metric obtained from dividing the number of correctly recognized to the total number of cases. Accuracy is the proximity of a calculated value to a normal or real value. In other words, the tool can measure the exact amount whose accuracy can be measured.

Accuracy=TP+TNTP+TN+FP+FN (1)

In ML, precision for a class is the number of truly classified items (correctly labeled) divided by the sum of either true or false items labeled as belonging to that class. Recall refers to the fraction of truly classified of the total number of classified items in a class.

Precision=TPTP+FP (2)
Recall=TPTP+FN (3)

Based on the precision and recall calculations, the weighting value for F1-score can be calculated. The F1-score is a useful metric for assessing classification efficiency and defining the weighted average of precision and recall quantities. The value of this measure for a classification algorithm is equal to 1 under the ideal condition and equals zero under the worst condition. This parameter is calculated according to the following equation [36]. Table 5 summarizes the results of the ResNet-50, VGG-16, CNN, and CAENN. The training accuracy of ResNet-50 is found to be 97.9408% and its validation accuracy is 92.2244%. The training and validation accuracies of VGG-16 are 98.3224% and 94.0763% respectively. The proposed CNN has training and validation accuracies of 97.3133% and 93.8430% respectively. Also, the training accuracy of suggested CAENN is obtained as 98.5949%, and its validation accuracy is 93.0482%.

Table 5.

Results from the proposed network for classification into 2 classes.

Network Acc. Prec. Rec. F1-score
ResNet-50 0.92244979 0.9650 0.9650 0.9650
VGG-16 0.94076305 0.9550 0.9550 0.9550
CNN 0.93843057 0.99252539 0.83801088 0.84655203
CAENN 0.93048090 0.96512168 0.74901946 0.84344849

According to Table 5, the accuracy, precision, recall, and F1-score of each DL network are presented in Fig. 9.

Fig. 9.

Fig. 9

Comparison of the F1-score, recall, precision, and accuracy, for proposed networks.

The training and validation accuracies, and loss analysis of the suggested models with respect to the epochs number are seen in Fig. 10.

Fig. 10.

Fig. 10

Training and validation analysis over 100 epochs for (1) ResNet-50 network: (a) Accuracy analysis of Training and Testing, (b) Loss analysis of Training and Testing. (2) VGG-16 network: (c) Accuracy analysis of Training and Testing, (d) Loss analysis of Training and Testing. (3) CNN: (e) Accuracy analysis of Training and Testing, (f) Loss analysis of Training and Testing (4) CAENN: (g) Accuracy analysis of Training and Testing, (h) Loss analysis of Training and Testing.

It is important in medical research, particularly for critical diseases like COVID-19, to minimize false negative and false positive outcomes in the modeling process. Classification accuracy of all networks is mentioned and impacts of false negative and false positive rate are presented in Fig. 11. It is clear that networks well performed in terms of reducing the number of false negatives and false positives.

Fig. 11.

Fig. 11

The Confusion Matrix measurements of TP, TN, FP, and FN ratios of the proposed model derived from the testing dataset of (a) ResNet-50 neural network (b) VGG-16 neural network (c) CNN and (d) CAENN.

Receiver Operating Characteristics (ROC) plots are helpful for grouping, analyzing, and visualizing the output of classifiers. ROC plots are widely used in medical decision-making and have recently gained popularity in data mining and ML. The ROC curve is provided by drawing true positive rate (TPR) versus false positive rate (FPR) in a diagram of various threshold settings. Maximizing the TPR, thus lowering the FPR, is optimal. This implies that the optimal point (FPR = 0 and TPR = 1) is in the upper left corner of the plot. Fig. 12 shows the ROC curve of the networks along with class 0 and class 1 based COVID-19 classification networks. The ideal point is observable for both class 0 and class 1. The maximum value (area = 1.00) was used to calculate the area under the ROC curve (also known as the area under the curve (AUC)).

Fig. 12.

Fig. 12

Roc plot of (a) ResNet-50 neural network (b) VGG-16 neural network (c) CNN and (d) CAENN.

The outcome of classical ML classifiers such as NN, SVM, RF, SGD LR, and MLP was compared for classification into 2 classes. The obtained accuracy was 94% for nearest neighbor, 88% for random forest, 73% for logical regression, 69% for stochastic gradient descent, 59% multilayer for perceptron, and 48% for support vector machine. Fig. 13 shows the comparison of these results. The F1-score, recall, and precision for each set of Healthy Control and COVID-19 data were calculated by these methods and summarized in Table 6. For COVID-19 data, the highest precision was obtained for random forest 93%, the highest recall for SVM 100%, and the highest F1-score for KNN 94%. For Non-COVID data, the highest precision, F1-score, and recall were respectively calculated 97%, 94%, and 95% for RF.

Fig. 13.

Fig. 13

Comparison of the classification accuracy resulting from ML classifiers.

Table 6.

ML precision, F1-score, and recall COVID-19 and Non-COVID-19 classification of classifiers.

Method COVID-19
Non-COVID
Prec. Rec. F1-score Prec. Rec. F1-score
KNN 0.92 0.97 0.94 0.97 0.92 0.94
RF 0.93 0.82 0.88 0.85 0.95 0.90
LR 0.75 0.68 0.72 0.73 0.79 0.76
SGD 0.89 0.42 0.57 0.64 0.95 0.76
MLP 0.58 0.57 0.57 0.61 0.62 0.61
SVM 0.48 1.00 0.65 0.00 0.00 0.00

According to Table 6, the average precision, average recall, and average F1-score are presented for each ML classification method in Fig. 14. The highest average results are observed at the KNN, RF, LR, SGD, MLP, and SVM, respectively.

Fig. 14.

Fig. 14

Comparison of the average precision, recall, and F1-score for ML classifiers.

Similar data to ours in this study have been investigated in other studies using other neural network architectures. Also, some studies have used pre-trained networks such as VGG-16 and ResNet-50. The results of our proposed networks have been compared with those of other studies as shown in Table 7. Among our proposed models, KNN and VGG-16 outperformed other models including CNN, ResNet-50, and CAENN by at least 1% higher accuracy. The accuracy obtained from the proposed CNN is equal to that of the proposed CAENN. Similar to our study, [19] and [25] have also used the same dataset. In [13], [14] and [16], ResNet-50 and CNN neural networks have been used and in [26], only CNN has been used.

Table 7.

Comparison of the classification resulted from deep learning networks of this study and other studies with the same data.

Reference Method Acc. Prec. Sens. (Rec.) Spec. F1-score
[19] Pre-trained DenseNet201 based
DTL
0.9625 0.9629 0.9629 0.9629
[25] There are two branches of the
network:
(1) four separate convolutional
layers
(2) heavier dens connections for
Representation learning
0.9083 0.9575 0.8589 0.9087
[13] ResNet-50 0.9262 0.9114 0.9429 0.93
[14] ResNet-50 0.77 0.6250 0.8862 0.7059
[16] ResNet-18 0.925 0.916 0.933 0.918 0.925
[26] CNN 0.9250 0.9025 0.985
Proposed ResNet-50 ResNet-50 0.9224 0.9650 0.9650 0.9650
Proposed VGG-16 VGG-16 0.9407 0.9550 0.9550 0.9550
Proposed CNN CNN 0.9384 0.9925 0.8380 0.8465
Proposed CAENN 2D-CAENN 0.9304 0.9651 0.7490 0.8434
Proposed KNN 0.94 0.9450 0.9450 0.94

In [19], DenseNet201 based DTL has been suggested with a classification accuracy 2% more than that of KNN and VGG-16 classifiers in this research.

In [25], a special architecture of CNN was developed to classify COVID-19 dataset. This study obtained accuracy 4% less than that of VGG-16 and KNN classifiers, 3% less than that of CAENN and CNN, and 2% less than that of ResNet-50 network proposed in our study.

In [13], ResNet-50 was used and accuracy was obtained equal to that of ResNet-50 and less than that of other networks in our study. The accuracy of ResNet-50 used in [14] was 15% less than the accuracy of that used in our research.

Another type of ResNet is the ResNET-18 CNN. In [16], ResNET-18 and other pre-trained CNNs such as GoogleNet, AlexNet, and ShuffleNet were proposed for being used in CAD system to distinguish COVID-19 from other cases and classify them. The accuracy of proposed ResNET-18 in this study was 92%, equal to the accuracy of proposed ResNET-50 in our study, whereas it was less than that of our proposed VGG-16, CNN, KNN, and CAENN models.

In [26], a CNN with several pooling layers and convolutional layers was used to extract CT image features and classify them into two classes. This network achieved an accuracy of 92%, which is equal to the accuracy of our ResNet-50, but 1% less than that of our CNN model.

According to Table 7, the accuracy of each model is presented in Fig. 15.

Fig. 15.

Fig. 15

Comparison of the accuracy of our proposed models and other studies.

The pre-trained CNN models are widely used in various studies with promising results. A comparison was made between the results of these models in Table 8. Among them, ResNet and VGG networks are the most used with varied results. The highest accuracy and sensitivity were respectively obtained 99.5% for Inception-v3 and 100% for ResNet-50 and U-Net++.

As mentioned, we have implemented models in Google Colaboratory (Colab) environment and the duration of each epoch was 20 s, 15 s for ResNet-50 and VGG-19 respectively, and 9 s for both CNN and CAENN. As a result, 100 epoch duration for ResNet-50 was 33.33 m, for VGG-19 was 25 m and for CNN and CAENN was 15 m. In this study, we have proposed a CNN with ten convolution layers and batch normalization layer, and dropout layer to avoid overfitting. The train parameters and hyper-parameters of networks such as learning rate, batch size, and activation function were adjusted to obtain high results. Although in our study pre-processing techniques were not used, compared with [26], our proposed CNN has improved accuracy and obtained precision of 99%. Also, we have proposed a CAENN with five convolution layers in the encoder section, five convolution layers in the decoder section, and two convolution layers for classification using the output of the last layer of the encoder. In this network, batch normalization layers and dropout layers were used to avoid overfitting as well. Among the studies reviewed, none of them used CAENN but the results of this network are promising. The accuracy obtained from CAENN is almost equal to CNN (93%) and has achieved a precision of 96%.

The architecture of pre-trained neural networks such as ResNet-50 and VGG-16 are defined, so different results can be obtained by changing the parameter and hyper-parameters of these networks. We have used a VGG-16 and our obtained results are higher than studies that used VGG-16 or VGG-19 network. According to Table 8, our proposed VGG even increased the accuracy by 10%. Also, the precision of 95%, Recall of 95%, and F1-score of 95% have been achieved by this network.

In this study, another proposed pre-trained neural network was ResNet-50. For this model, we adjusted the parameters of network training to achieve optimal results of classification. Compared with [14], our ResNet-50 increased the accuracy by 15%. Although in [13], transfer learning technique was used to train the ResNet-50 network the obtained accuracy was equal to the accuracy obtained by our proposed ResNet-50. Also, our proposed ResNet-50 improved precision, Recall, and F1-score. In [16], features of CT images extracted by the ResNet-18 network and then classified by SVM. Although in our study machine learning classifiers were not used, obtained results were improved.

Table 8.

Comparison of the classification resulted from pre-trained CNN.

Classification model Acc. Prec. Spec. Sens.
Xception 96.55% 81.74% 97.45% 97.24%
Alex-Net 79% 77% 81%
MobileNet-v2 95.97% 95.14% 96.71%
VGG-19 94.5%
VGG-19 84% 81%
ResNet-18 78.29% 81% 79.9% 76.9%
ResNet-18 90.16% 90.95% 89.45%
ResNet-50 98% 80% 100%
DenseNet201 96% 96% 96%
Inception-v3 99.5% 99.2% 98.2% 99.8%
U-Net++ 92% 81.82% 100%
Google-Net 89.5% 87% 88%

4. Conclusion

In this study, the data set included 2492 CT-scan images with 1262 images of positive (COVID-19) and 1230 images of negative (non-COVID) cases. All scans were resized to 173 × 100 × 3. Using these images, four deep learning networks were implemented for differentiating between COVID and Non-COVID cases including ResNet-50, VGG-16, CNN, and CAENNs. Also, ML classifiers of SVM, NN, RF, SGD, LR, and MLP were compared for two-class classification. The ResNet-50, VGG-16, and CAENNs classified the chest CT-scans with validation accuracies of 92.24%, 94.07%, 93.84%, and 93.04%, respectively. The best performance among ML classifiers belonged to CNN with an accuracy of 94%. Also, other evaluation metrics such as recall, precision, and F1-score were computed for the proposed networks. The highest precision was obtained 99% by CNN and the highest recall and F1-score were obtained 96% by ResNet-50. In this research, the proposed networks were tested to classify COVID-19, SARS, MERS, and EBOLA images showing acceptable results. In future research, it is suggested to test other desirable networks to classify these images.

Compared to the existing studies with a similar dataset, this research achieved better results although techniques of image preprocessing and data augmentation were not used. Also, the ResNet and CNN networks used in our study showed a higher performance compared to similar networks in other studies. The use of pre-trained network or data augmentation yielded an accuracy of 99%, reported in previous research.

The use of other datasets including more images than the initially published ones is also suggested for future work. To improve the results, it is recommended to use data augmentation and data analysis techniques for further investigation. Additionally, DL-based optimal models can be used for the diagnosis of COVID-19, SARS, MERS, and EBOLA diseases from X-ray or CT images. DL models used to quantify and segment the severity of infected region of chest from CT images. So it is recommended to use these models for detecting the range of infected tissues. To aid physicians in the real-time diagnosing of COVID-19, it is suggested to develop a computer aided diagnosis (CAD) system using proposed models. The future work intends to develop deep learning models to classify recorded voice from people into COVID-19 and non-COVID-19​ categories.

CRediT authorship contribution statement

Saman Fouladi: Conceptualization, Methodology, Software. M.J. Ebadi: Data curation, Writing - original draft. Ali A. Safaei: Data curation, Writing - original draft. Mohd Yazid Bajuri: Supervision, Validation, Reviewing original draft. Ali Ahmadian: Supervision, Validation, Reviewing original draft.

Declaration of Competing Interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

Footnotes

References

  • 1.Lai C.-C., Shih T.-P., Ko W.-C., Tang H.-J., Hsueh P.-R. Severe acute respiratory syndrome coronavirus 2 (SARS-CoV-2) and coronavirus disease-2019 (COVID-19): The epidemic and the challenges, Int. J. Antimicrob. Agents. 2019;55 doi: 10.1016/j.ijantimicag.2020.105924. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 2.Chen N., Zhou M., Dong X., Qu J., Gong F., Han Y., Qiu Y., Wang J., Liu Y., Wei Y., Xia J., Yu T., Zhang X., Zhang L. Epidemiological and clinical characteristics of 99 cases of 2019 novel coronavirus pneumonia in wuhan, China: A descriptive study. Lancet. 2019;395 doi: 10.1016/S0140-6736(20)30211-7. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3.Coronavirus disease (covid-19) outbreak situation, WHO. (n.d.). https://www.who.int/emergencies/diseases/novel-coronavirus-2019.
  • 4.Yang Y., Yang M., Shen C., Wang F., Yuan J., Li J., Zhang M., Wang Z., Xing L., Wei J., Peng L., Wong G., Zheng H., Wu W., Liao M., Feng K., Li J., Yang Q., Zhao J., Zhang Z., Liu L., Liu Y. 2020. Evaluating the accuracy of different respiratory specimens in the laboratory diagnosis and monitoring the viral shedding of 2019-nCoV infections, MedRxiv. 2020.02.11.20021493, [DOI] [Google Scholar]
  • 5.Yang Y., Yang M., Shen C., Wang F., Yuan J., Li J., Zhang M., Wang Z., Xing L., Wei J., Peng L., Zheng H., Liao M., Liao M., Feng K., Li J., Yang Q., Zhao J., Liu Y. 2020. Laboratory diagnosis and monitoring the viral shedding of 2019-nCoV infections. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6.Gupta A., Singh D., Kaur M. An efficient image encryption using non-dominated sorting genetic algorithm-III based 4-D chaotic maps. J. Ambient Intell. Humaniz. Comput. 2020;11:1309–1324. doi: 10.1007/s12652-019-01493-x. [DOI] [Google Scholar]
  • 7.Salehi S., Abedi A., Balakrishnan S., Gholamrezanezhad A. Coronavirus disease 2019 (COVID-19): A systematic review of imaging findings in 919 patients. AJR. Am. J. Roentgenol. 2019;215:87–93. doi: 10.2214/AJR.20.23034. [DOI] [PubMed] [Google Scholar]
  • 8.Wong H.Y.F., Lam H., Fong A.H.-T., Leung B.S.T., Chin T., Lo C., Lui M., Lee J., Chiu W.H., Chung T., Lee E., Wan E., Hung F., Lam T., Kuo M., Ng M.-Y. Frequency and distribution of chest radiographic findings in COVID-19 positive patients. Radiology. 2020 doi: 10.1148/radiol.2020201160. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9.Tanne J.H., Hayasaki E., Zastrow M., Pulla P., Smith P., Rada A.G. Covid-19: how doctors and healthcare systems are tackling coronavirus worldwide. BMJ. 2020;368:m1090. doi: 10.1136/bmj.m1090. [DOI] [PubMed] [Google Scholar]
  • 10.Jaakkola H., Henno J., Mäkelä J., Thalheim B. 2019 42nd Int. Conv. Inf. Commun. Technol. Electron. Microelectron. 2019. Artificial intelligence yesterday, today and tomorrow; pp. 860–867. [DOI] [Google Scholar]
  • 11.Singh D., Kaur M. Fusion of medical images using deep belief networks. Cluster Comput. 2020;23 doi: 10.1007/s10586-019-02999-x. [DOI] [Google Scholar]
  • 12.Rahimzadeh M., Attar A., Sakhaei S.M. A fully automated deep learning-based network for detecting COVID-19 from a new and large lung CT scan dataset. Biomed. Signal Process. Control. 2021 doi: 10.1016/j.bspc.2021.102588. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.Pham T.D. A comprehensive study on classification of COVID-19 on computed tomography with pretrained convolutional neural networks. Sci. Rep. 2020;10:16942. doi: 10.1038/s41598-020-74164-z. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14.El-Kenawy E.-S.M., Ibrahim A., Mirjalili S., Eid M.M., Hussein S.E. Novel feature selection and voting classifier algorithms for COVID-19 classification in CT images. IEEE Access. 2020;8:179317–179335. doi: 10.1109/ACCESS.2020.3028012. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15.Shah V., Keniya R., Shridharani A., Punjabi M., Shah J., Mehendale N. 2020. Diagnosis of COVID-19 using CT scan images and deep learning techniques, MedRxiv. 2020.07.11.20151332. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16.Attallah O., Ragab D.A., Sharkas M. MULTI-DEEP: A novel CAD system for coronavirus (COVID-19) diagnosis from CT images using multiple convolution neural networks. PeerJ. 2020;8 doi: 10.7717/peerj.10086. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17.Serte S., Demirel H. Deep learning for diagnosis of COVID-19 using 3D CT scans. Comput. Biol. Med. 2021;132 doi: 10.1016/j.compbiomed.2021.104306. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18.Horry M.J., Chakraborty S., Paul M., Ulhaq A., Pradhan B., Saha M., Shukla N. COVID-19 detection through transfer learning using multimodal imaging data. IEEE Access. 2020;8:149808–149824. doi: 10.1109/ACCESS.2020.3016780. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19.Jaiswal A., Gianchandani N., Singh D., Kumar V., Kaur M. Classification of the COVID-19 infected patients using DenseNet201 based deep transfer learning. J. Biomol. Struct. Dyn. 2020:1–8. doi: 10.1080/07391102.2020.1788642. [DOI] [PubMed] [Google Scholar]
  • 20.El-Bana S., Al-Kabbany A., Sharkas M. A multi-task pipeline with specialized streams for classification and segmentation of infection manifestations in COVID-19 scans. PeerJ Comput. Sci. 2020;6 doi: 10.7717/peerj-cs.303. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21.Wang X., Deng X., Fu Q., Zhou Q., Feng J., Ma H., Liu W., Zheng C. A weakly-supervised framework for COVID-19 classification and lesion localization from chest CT. IEEE Trans. Med. Imaging. 2020;39:2615–2625. doi: 10.1109/TMI.2020.2995965. [DOI] [PubMed] [Google Scholar]
  • 22.Chen J., Wu L., Zhang J., Zhang L., Gong D., Zhao Y., Chen Q., Huang S., Yang M., Yang X., Hu S., Wang Y., Hu X., Zheng B., Zhang K., Wu H., Dong Z., Xu Y., Zhu Y., Chen X., Zhang M., Yu L., Cheng F., Yu H. Deep learning-based model for detecting 2019 novel coronavirus pneumonia on high-resolution computed tomography. Sci. Rep. 2020;10:19196. doi: 10.1038/s41598-020-76282-0. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 23.Wang S., Kang B., Ma J., Zeng X., Xiao M., Guo J., Cai M., Yang J., Li Y., Meng X., Xu B. A deep learning algorithm using CT images to screen for corona virus disease (COVID-19) Eur. Radiol. 2021 doi: 10.1007/s00330-021-07715-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 24.Wang S.-H., Govindaraj V.V., Górriz J.M., Zhang X., Zhang Y.-D. Covid-19 Classification by FGCNet with deep feature fusion from graph convolutional network and convolutional neural network. Inf. Fusion. 2021;67:208–229. doi: 10.1016/j.inffus.2020.10.004. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25.Wang Z., Liu Q., Dou Q. Contrastive cross-site learning with redesigned net for COVID-19 CT classification. IEEE J. Biomed. Heal. Inform. 2020;24:2806–2813. doi: 10.1109/JBHI.2020.3023246. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 26.Singh D., Kumar V., Vaishali Q., Kaur M. Classification of COVID-19 patients from chest CT images using multi-objective differential evolution-based convolutional neural networks. Eur. J. Clin. Microbiol. Infect. Dis. 2020;39:1379–1389. doi: 10.1007/s10096-020-03901-z. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27.Rostami M., Berahmand K., Forouzandeh S. A novel community detection based genetic algorithm for feature selection. J. Big Data. 2021;8(2) doi: 10.1186/s40537-020-00398-3. [DOI] [Google Scholar]
  • 28.Ebadi M.J., Hosseini Alireza, Hosseini M.M. A projection type steepest descent neural network for solving a class of nonsmooth optimization problems. Neurocomputing. 2017;235:164–181. doi: 10.1016/j.neucom.2017.01.010. [DOI] [Google Scholar]
  • 29.Rostami M., Berahmand K., Forouzandeh S. A novel method of constrained feature selection by the measurement of pairwise constraints uncertainty. J. Big Data. 2020;7(83) doi: 10.1186/s40537-020-00352-3. [DOI] [Google Scholar]
  • 30.Rostami M., Forouzandeh S., Berahmand K., Soltani M. Integration of multi-objective PSO based feature selection and node centrality for medical datasets. Genomics. 2020;112(6):4370–4384. doi: 10.1016/j.ygeno.2020.07.027. [DOI] [PubMed] [Google Scholar]
  • 31.Ebadi M.J., Ebrahimi A. Video data compression by progressive iterative approximation. Int. J. Interact. Multimed. Artif. Intell. 2021;6(6):189–195. doi: 10.9781/ijimai.2020.12.002. [DOI] [Google Scholar]
  • 32.Saha P., Mukherjee D., Singh P.K., et al. Graphcovidnet: A graph neural network based model for detecting COVID-19 from CT scans and X-rays of chest. Sci. Rep. 2021;11:8304. doi: 10.1038/s41598-021-87523-1. [DOI] [PMC free article] [PubMed] [Google Scholar] [Retracted]
  • 33.Gupta V., Jain N., Katariya P., Kumar A., Mohan S., Ahmadian A., Ferrara M. An emotion care model using multimodal textual analysis on COVID-19. Chaos, Solitons Fractals. 2021;144 doi: 10.1016/j.chaos.2021.110708. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 34.He K., Zhang X., Ren S., Sun J. 2016 IEEE Conf. Comput. Vis. Pattern Recognit. 2016. Deep residual learning for image recognition; pp. 770–778. [DOI] [Google Scholar]
  • 35.Zhang X., Zou J., He K., Sun J. 2015. Accelerating Very Deep Convolutional Networks for Classification and Detection. http://arxiv.org/abs/1505.06798 (Accessed 9 October 2020) [DOI] [PubMed] [Google Scholar]
  • 36.Riazi R., Asrardel M., Shafaei M., Vakilipour S., Zare H., Veisi H. A data mining study on combustion dynamics and NOx emission of a swirl stabilised combustor with secondary fuel injection. Int. J. Heavy Veh. Syst. 2016;24 doi: 10.1504/IJHVS.2017.10005324. [DOI] [Google Scholar]

Articles from Computer Communications are provided here courtesy of Elsevier

RESOURCES