Skip to main content
Springer Nature - PMC COVID-19 Collection logoLink to Springer Nature - PMC COVID-19 Collection
. 2022 Jan 3;41(6):3397–3414. doi: 10.1007/s00034-021-01939-8

Classifier Fusion for Detection of COVID-19 from CT Scans

Taranjit Kaur 1,, Tapan Kumar Gandhi 1
PMCID: PMC8722646  PMID: 35002014

Abstract

The coronavirus disease (COVID-19) is an infectious disease caused by the SARS-CoV-2 virus. COVID-19 is found to be the most infectious disease in last few decades. This disease has infected millions of people worldwide. The inadequate availability and the limited sensitivity of the testing kits have motivated the clinicians and the scientist to use Computer Tomography (CT) scans to screen COVID-19. Recent advances in technology and the availability of deep learning approaches have proved to be very promising in detecting COVID-19 with increased accuracy. However, deep learning approaches require a huge labeled training dataset, and the current availability of benchmark COVID-19 data is still small. For the limited training data scenario, the CNN usually overfits after several iterations. Hence, in this work, we have investigated different pre-trained network architectures with transfer learning for COVID-19 detection that can work even on a small medical imaging dataset. Various variants of the pre-trained ResNet model, namely ResNet18, ResNet50, and ResNet101, are investigated in the current paper for the detection of COVID-19. The experimental results reveal that transfer learned ResNet50 model outperformed other models by achieving a recall of 98.80% and an F1-score of 98.41%. To further improvise the results, the activations from different layers of best performing model are also explored for the detection using the support vector machine, logistic regression and K-nearest neighbor classifiers. Moreover, a classifier fusion strategy is also proposed that fuses the predictions from the different classifiers via majority voting. Experimental results reveal that via using learned image features and classification fusion strategy, the recall, and F1-score have improvised to 99.20% and 99.40%.

Keywords: CT images, COVID-19, Diagnosis, Transfer learning, Activations, Classifier fusion

Introduction

The emergence of COVID-19 infection has critically affected the social and economic structures of both the developing and the developed countries since December 2019 [12]. Researchers and health care workers around the globe are trying to apprehend the COVID-19 etiology and its effect on the quality of life [30].

Nowadays, Computer Tomography (CT) scans are emerging as an alternative for screening in contrast to the conventional reverse transcription-polymerase chain reaction (RT-PCR). This is owing to the limitations associated with the availability, reproducibility, and significant false-negative outcomes produced by RT-PCR kits.

Distinctive CT scans having patchy ground-glass opacities are important biomarkers that can aid in speedy detection and isolation of the subject [1, 16, 31].

As the cases are increasing at an alarming rate, an automated detection system for COVID-19 is the need of the hour that can assist in faster virus detection at different stages thereby relieving the healthcare professionals from the manual annotation task.

Several artificial intelligence (AI) based methods have evolved that automatically provide a diagnosis whether a CT image is COVID +ve or not [3, 4, 8, 1921, 29, 32, 34].

Soares et al. [28] build a publically available CT scan data for severe acute respiratory syndrome coronavirus 2 (SARS-COV-2) comprising a total of 2482 images. The authors proposed a non-iterative algorithm named eXplainable deep network (x-DNN) based on recursive calculation for the categorization of data as non-COVID and COVID +ve. The authors attained a promising F1 value of 97.31% over the validation data.

Nour et al. [21] developed a diagnosis model for SARS-COV-2 infection detection using deep discriminative features and a Bayesian algorithm. The authors used a five convolutional layered model as a deep feature extractor; thereafter, the features were fed to the standard K-nearest neighbor (KNN), support vector machine (SVM), and decision trees whose hyperparameters were fine-tuned with Bayesian optimization procedure. The authors validated their method on the publically available X-ray image dataset and concluded that the SVM algorithm provided the best prediction resulting in an accuracy of 98.97% under the data partitioning ratio of 7:3.

He et al. [11] introduced a dataset of CT scan images with hundreds of COVID +ve images freely available for research. The authors also developed the Self-Trans method integrating self-supervised contrastive learning with the transfer learning-based mechanism for the separation of infected scans from normal ones. The researchers attained an F1-value of 0.85 using a split proportion of 0.6, 0.25, and 0.15 for training, testing, and validation.

Saygılı [25] proposed an automated system for separating COVID +ve CT scans from the non-infected scan images. The authors followed the pipeline of data set acquisition, image pre-processing that includes rgb2gray transformation, image resizing & image sharpening, Feature extraction that includes Local Binary Patterns & histograms of the Oriented Gradients, Feature reduction using Principal Component Analysis, and classification using standard machine learning (ML) algorithms. They achieved an accuracy of 98.11% on the dataset provided by Soares et al. [28] using the handcrafted features with a tenfold data partitioning scheme.

Kaur and Gandhi [13] designed a method for COVID detection based upon concatenation of deep features from the learned ResNet50 and learned MobileNetv2 model. The deep features were deduced by taking activations from the ‘avg_pool’ and the ‘Conv1’ layer of the learned models. Thereafter, the features were concatenated and given as input to the SVM for classification. The feature fusion approach, although with the high dimensionality of 63,232 yielded a validation accuracy of 98.35% on the benchmark COVID CT dataset.

In another work via Kaur et al. [15], they developed a diagnosis scheme for COVID-19 signature detection using deep features and Parameter Free-BAT (PF-BAT) enhanced Fuzzy-KNN (FKNN). Firstly, the pre-trained MobileNetv2 model was fine-tuned on CT chest radiographs. Thereafter, the features were extracted by performing activations onto the fully connected layer of the fine-tuned model. The features along with the corresponding label were fed to the FKNN classifier whose hyperparameters, i.e., nearest neighbor (‘k’) and fuzzy strength measure (‘m’) were fine-tuned via PF-BAT. Experimenting on the dataset by Soares et al. [28] reveals that the system achieved an average accuracy of 99.18%.

Goel et al. [7] designed an automated method for SARS-COV-2 detection using a framework that employs Generative Adversarial Network (GAN) for augmentation, Whale optimization for hyperparameter tuning of GAN network, and classification using transfer learned Inception V3 model. The researchers achieved a prediction accuracy value of 99.22% on the benchmark CT scan dataset using train test splits as 7:3.

Sen et al. [26] developed a feature selection approach for COVID-19 signature detection from lung CT scans. The detection framework uses a Convolutional Neural Network (CNN) architecture as a deep feature extractor. Thereafter, feature selection was done in two stages, i.e., firstly filter-based method is used then the Dragonfly optimization algorithm was applied over the ranked features. The selected features were then used by SVM for classification. The prediction rate was 90% and 98.39% on the benchmark CT scan image datasets.

Surveying the recent literature divulges that the use of CT images is evolving rapidly for COVID-19 detection owing to the shortage and the limited sensitivity of the RT-PCR detection kits [17]. Furthermore, to automate and expedite the screening of distinctive CT scans, the concept of transfer learning has proven to be quite advantageous, especially if the size of the dataset is limited [13, 15]. Additionally, to generalize the performance of the different algorithms on the benchmark CT scan dataset, a public database was provided by Soares et al. [28]. The survey also highlights that the predicted accuracy over this dataset is still limited, and there is a scope to improvise this rate further via effective mathematical models.

With the inspiration to improvise the prediction results, variants of the ResNet model are explored for detection. ResNet model variants are selected because of their proven better performance for the SARS-COV-2 detection task using CT scans [19].

Additionally, the potential of the learned image features from the transfer learned model is also investigated for COVID-19 detection when combined with the established ML algorithms (SVM. KNN, Logistic Regression (LR)). A classification fusion mechanism is also proposed that combines the predictions from the different ML algorithms via majority voting. Summarizing the vital contributions of the proposed work are:

  1. A comparative study is performed with various architectures of Deep CNN’s (DCNN) like ResNet18, ResNet50, and ResNet101 for COVID-19 detection using the transfer learning concept.

  2. Results revealed that the transfer learning-based deep ResNet50 model exhibited the best performance w.r.t the other residual network variants.

  3. The potential of the image attributes extracted from the various layers of the transfer learned model is also investigated by combining them with well-established ML algorithms like SVM, KNN, and LR.

  4. A classification fusion scheme is also proposed that combines the predictions from the different classifiers via majority voting to further boost the classification performance.

In this present work, Sect. 2 provides the materials and methods, followed by Sect. 3, which describes the experimental results. Discussion and conclusions are illustrated in Sects. 4 and 5, respectively.

Materials and Methods

COVID Dataset Depiction

The proposed method has been analyzed on the publically available benchmark COVID CT scan dataset provided by Soares et al. [28]. The statistics relating to the data are provided in Table 1. Sample CT scans are shown in Fig. 1.

Table 1.

Train validation splits

Data split Non-COVID COVID Cumulative
Training 983 1003 1986
Validation 246 250 496

Fig. 1.

Fig. 1

CT scans from the dataset COVID + ve (upper row) and non-infected by SARS-COV-2 (lower row)

Performance Measures

In the present paper, Accuracy, Recall, Precision, F-score, and Area under the Curve (AUC) are selected to quantitatively validate the competence of the designed method.

Transfer Learning

One of the effective mechanisms making use of pre-trained deep models for image classification under limited data scenarios is transfer learning. In the present work, variants of a pre-trained ResNet model are used [11]. ResNet models follow a forward neural network architecture with “shortcut connections” to train CNN [10, 24]. Through these “shortcut connections” gradients can easily propagate, which makes the training faster. Primarily, ResNet18 [24], ResNet50 [10], and ResNet101 [10] with transfer learning are used in this work as they have exhibited better performance than other computational models [19]. Only the last three layers of the model variants are fine-tuned to accommodate new image categories, as shown in Fig. 2 [33].

Fig. 3.

Fig. 3

Proposed methodology

Fig. 2.

Fig. 2

Visualization for transfer learning using pre-trained models [14]

Feature Classification Using Conventional Machine Learning Algorithms

Apart from using the transfer learned model for classification, the potential of learned image features is also explored for COVID-19 detection. The network activations from different layers of the learned model are taken, and then, they are fed to the proven state-of-the-art classifiers like SVM, KNN, and the LR. The following section briefly outlines the mathematical formulation for SVM, KNN, and LR.

Support Vector Machine (SVM)

SVM is among the most popular supervised ML algorithms that function by constructing an optimal hyperplane [9]. Let M training examples (pi, zi) be represented in an N-dimensional sample space, where pi is an example pattern and zi{-1,1} is the label. Let the kernel value matrix be represented by K and αi be the Lagrange coefficients to be calculated via optimization procedure. Solving the quadratic equation given below results in an optimal separating hyperplane

maxW(α)=-12iMjMαiαjzizjK(pi;pj)+iMαi 1

The equation given above is subject to the constrain 0αiC,i and iMαizi=0.

K-Nearest Neighbor (KNN)

KNN is one of the most simplistic nonparametric pattern recognition technique [5, 18]. In the KNN algorithm, a label is assigned according to the most common labels from its k-nearest neighbors. The main advantages of the KNN classifier are its simple implementation and fewer parameters to tune, i.e., distance metric and k.

Step by Step procedure for KNN algorithm:

  • (i)

    Initialize the number of nearest neighbors (k).

  • (ii)
    Compute the distance between the test image and all the training images. Any distance criteria could be used. E.g. Euclidean distance is primarily used and it is governed by the equation
    Distance(a,b)=a-b 2
    where (a, b) are two different samples in the feature space.
  • (iii)

    Sort the distances and calculate the nearest neighbor based on the kth minimum distance.

  • (iv)

    Get the corresponding labels of the training data which falls under k for the sorted condition.

  • (v)

    Take the majority of k-nearest neighbors as the output label.

Logistic Regression (LR)

LR method uses a sigmoid or a logistic function. Considering a 2 class problem with f0(X), f1(X) as the class conditional densities, q0(X), q1(X) as posterior probabilities, and p0(X), p1(X) as prior probabilities, then according to Bayes rule [2]

q0(X)=f0(X)p0f0(X)p0+f1(X)p1=11+exp(-ξ) 3

where ξ is defined as

ξ=-lnf1(X)p1f0(X)p0=lnf0(X)p0f1(X)p1 4

Alternatively,

lnf0(X)p0f1(X)p1=WTX+w0 5

The above equation holds if f0 and f1 are Gaussian with similar covariance matrix. This is the case of logistic regression where it would result in an optimal classifier. In logistic regression, the goal is to find W and w0 that minimizes 12i=1nhWTXi+w0-yi2 where h(a)=(1+exp(-a))-1 is the sigmoid or logistic function and yi{0,1} are the targets.

Experimental Results

Experimental Setup

The experimental settings for the proposed work include using an ‘Adam” optimizer with the initial learning rate of 0.0006 to minimize the cross-entropy loss. The variant models were fitted for 20-epochs with 150 as the batch size. The image data were processed on an Intel Core i7-4500U CPU having 8 GB RAM, with a 1.8 GHz processor in a MATLAB 19a platform. Other training options include shuffling the data before each epoch using L2-regularizer with a weight decay value of 0.05 to circumvent overfitting. Moreover, the data splits, i.e., train/validation were used as per the base paper by Soares et al. [28]. The researchers in Soares et al. [28] have given the .mat files for train-validation splits that do not contain subject-related information, i.e., without any metadata on the images. The images are only designated by numbers without any indication of whether they are subject-independent or not.

Results

The experimental results via deploying the learned ResNet model variants for the COVID-19 detection task are given in Table 2. The tabulated entries show that the best results are attained by the ResNet50 model. It has achieved a precision of 98.02%, recall of 98.80%, AUC of 0.9994, F1-score of 98.41%, and a validation accuracy of 98.35%. Other models that are taken for comparison, i.e., ResNet18 and ResNet101 achieved a validation accuracy of 97.12% and 96.71%, respectively. The confusion matrix for all the three learned variants is given in Fig. 4. The smallest misclassification error is achieved by ResNet50 followed by ResNet18, and then ResNet101. The accuracy/loss versus epoch plot for the best performing model is given in Fig. 5, which indicates that validation closely follows the training. The AUC plot for ResNet50 is shown in Fig. 6, which specifies that the model attains a value of 0.9994 for True Positive Rate Vs False Positive fraction.

Table 2.

Performance comparison of the ResNet model variants over validation data

Sr. no Accuracy (%) Precision (%) Recall (%) F1-score (%) AUC
ResNet18 97.12 96.83 97.60 97.21 0.9973
ResNet50 98.35 98.02 98.80 98.41 0.9994
ResNet101 96.71 96.43 97.20 96.81 0.9944

Fig. 4.

Fig. 4

Confusion matrix a ResNet18 b ResNet50 c ResNet101 d classification fusion

Fig. 5.

Fig. 5

Training versus epoch and loss versus epoch plot for best performing transfer learned model (ResNet50)

Fig. 6.

Fig. 6

AUC curve for best performing transfer learned model (ResNet50)

Figure 7 shows the occlusion sensitivity maps for the learned ResNet50 model. It gives us an idea about which area of the scan is most decisive for classification, i.e., occluding which results in a maximum drop in the probability score. The regions that are positively contributing to the probability score are shown in red color.

Fig. 7.

Fig. 7

Activation maps obtained via transfer learned ResNet50 model

In the present work, we also tried to investigate the efficacy of the activations from the several layers of the transfer learned ResNet50 model for the COVID-19 detection task. These activations from the various layers of the trained model are used in conjunction with well-established classifiers such as SVM, KNN, and LR. Beginning from the fully connected layer, activations from specific layers are taken and the detection results are reported in Table 3. Interestingly, e.g., the activations extracted from the layer ‘res5b_branch2c’ proved to be decisive for the classification using SVM and the LR classifier. It rendered a value of 99.20% for precision, recall, F1-score, and a value of 99.18%, 0.9997 for accuracy, and AUC.

Table 3.

Performance metrics for transfer learned ResNet50 using activations from specific layers

Layer Classifier Precision (%) Recall (%) Accuracy (%) F1-score (%) AUC Dimension of training feature space Dimension of validation feature space
‘fc’ SVM 98.02 98.80 98.35 98.41 0.9932 1986 × 2 486 × 2
KNN 97.24 98.80 97.94 98.02 0.9792
LR 98.02 98.80 98.35 98.41 0.9994
Classifier fusion 98.02 98.80 98.35 98.41 0.9994
‘avg_pool’ SVM 98.41 99.20 98.77 98.80 0.9995 1986 × 2048 486 × 2048
KNN 98.80 99.20 98.97 99.00 0.9896
LR 99.20 99.20 99.18 99.20 0.9993
Classifier fusion 99.60 99.20 99.38 99.40 0.9997
‘res5c_branch2c’ SVM 98.02 98.80 98.35 98.41 0.9975 1986 × 100,352 486 × 100,352
KNN 99.59 97.60 98.56 98.59 0.9859
LR 98.41 99.20 98.77 98.80 0.9993
Classifier fusion 98.41 99.20 98.77 98.80 0.9993
‘res5c_branch2b’ SVM 98.42 99.60 98.97 99.01 0.9997 1986 × 25,088 486 × 25,088
KNN 98.81 99.60 99.18 99.20 0.9916
LR 99.20 99.20 99.18 99.20 0.9995
Classifier fusion 99.60 99.20 99.38 99.40 0.9997
‘res5c_branch2a’ SVM 98.41 98.80 98.56 98.60 0.9997 1986 × 25,088 486 × 25,088
KNN 98.02 99.20 98.56 98.61 0.9854
LR 98.41 99.20 98.77 98.80 0.9997
Classifier fusion 98.41 99.20 98.77 98.80 0.9997
‘res5b_branch2c’ SVM 99.20 99.20 99.18 99.20 0.9997 1986 × 100,352 486 × 100,352
KNN 98.02 99.20 98.56 98.61 0.9854
LR 99.60 98.80 99.18 99.20 0.9997
Classifier fusion 99.60 99.20 99.38 99.40 0.9997
‘res5b_branch2b’ SVM 98.41 99.20 98.77 98.80 0.9997 1986 × 25,088 486 × 25,088
KNN 97.66 100 98.77 98.81 0.9873
LR 97.64 99.20 98.35 98.41 0.9996
Classifier fusion 97.64 99.20 98.35 98.41 0.9996
‘res5b_branch2a’ SVM 98.02 99.20 98.56 98.61 0.9997 1986 × 25,088 486 × 25,088
KNN 99.20 99.20 99.18 99.20 0.9997
LR 98.41 99.20 98.77 98.80 0.9997
Classifier fusion 98.41 99.20 98.77 98.80 0.9997
‘res5a_branch1’ SVM 98.02 99.20 98.56 98.61 0.9994 1986 × 100,352 486 × 100,352
KNN 97.27 99.60 98.35 98.42 0.9832
LR 96.88 99.20 97.94 98.02 0.9997
Classifier fusion 98.02 99.20 98.56 98.61 0.9994
‘res5a_branch2c’ SVM 98.02 99.20 98.56 98.61 0.9996 1986 × 100,352 486 × 100,352
KNN 98.80 99.20 98.97 99.00 0.9896
LR 98.41 99.20 98.77 98.80 0.9997
Classifier fusion 98.41 99.20 98.77 98.80 0.9997
‘res5a_branch2b’ SVM 98.80 99.20 98.97 99.00 0.9994 1986 × 25,088 486 × 25,088
KNN 97.65 99.60 98.56 98.61 0.9853
LR 98.80 98.80 98.77 98.80 0.9992
Classifier fusion 98.80 99.20 98.97 99.00 0.9896
‘res5a_branch2a’ SVM 98.80 99.20 98.97 99.00 0.9994 1986 × 25,088 486 × 25,088
KNN 96.88 99.20 97.94 98.02 0.9791
LR 99.19 98.40 98.77 98.80 0.9994
Classifier fusion 99.20 99.20 99.18 99.20 0.9997
‘res4f_branch2c’ SVM 98.80 98.80 98.77 98.80 0.9991 1986 × 200,704 486 × 200,704
KNN 93.80 90.80 92.18 92.28 0.9222
LR 98.41 98.80 98.56 98.60 0.9992
Classifier fusion 98.80 98.80 98.77 98.80 0.9991
‘res4f_branch2b’ SVM 98.02 98.80 98.35 98.41 0.9989 1986 × 50,176 486 × 50,176
KNN 92.28 95.60 93.62 93.91 0.9356
LR 98.40 98.40 98.35 98.40 0.9988
Classifier fusion 98.40 98.40 98.35 98.40 0.9988
‘res4f_branch2a’ SVM 94.30 99.20 96.50 96.69 0.9977 1986 × 50,176 486 × 50,176
KNN 84.96 90.40 86.83 87.60 0.8673
LR 94.64 98.80 96.50 96.67 0.9974
Classifier fusion 88.55 87.22 86.83 87.88 0.8669

Bold signifies the best results

To further improve the classification performance, a fusion strategy is also proposed as outlined in Fig. 8, where predictions from the three different classifiers are fused according to the majority voting rule. The confusion matrix resulting from the classification fusion is shown in Fig. 4d indicating that the misclassifications are reduced to merely three samples.

Fig. 8.

Fig. 8

Classification fusion strategy

On fusing the predictions from SVM, KNN, and LR using features from the ‘res5b_branch2c’ layer, a validation accuracy of 99.38% and F1-score of 99.40% is achieved. The advantage of classification fusion is also highlighted for the activations from ‘avg_pool’ layer where a validation accuracy of 99.38% is achieved using a feature dimension of 2048 rather than using a feature space of 100,352.

Discussion

The proposed fusion approach has also been compared with other model architectures that have used the same dataset. As apparent from Table 4, the proposed prediction fusion mechanism yields a precision of 99.60%, F-score of 99.40%, Recall of 99.20%, Accuracy of 99.38% that is higher than the existing network architectures (detection accuracy of 97.38%, 98.35%, 98.39%, 98.37%, 98.99%, 94.04%, 92% reported in Kaur and Gandhi [13], Panwar et al. [22], Pathak et al. [23], Silva et al. [27], Soares et al. [28], Fouladi et al. [6], Sen et al. [26]). In contrast to the reported works in literature, the proposed prediction fusion mechanism offers some benefits in addition to achieving a promising level of detection performance: (1) It employs a single pre-trained network architecture for COVID-19 classification differing from the usage of both VGG-16 and xDNN proposed in Soares et al. [28]. (2) It does not employ any optimization procedure for fine-tuning the model hyperparameters as in Kaur et al. [15], Pathak et al. [23] where the authors have employed PF-BAT and Memetic Adaptive Differential Evolution optimization for FKNN and deep bidirectional long short-term memory network with a Mixture Density model (DBM) hyperparameter tuning.

Table 4.

Comparison of the proposed prediction fusion scheme with the recent state-of-the-art works

Method Accuracy (%) Precision (%) Recall (%) F1 Score (%) AUC
xDNN [28] 97.38 99.16 95.53 97.31 0.9736
Alexnet [28] 93.75 94.98 92.28 93.61 0.9368
VGG16 [28] 94.96 94.02 95.43 94.97 0.9496
GoogleNet [28] 91.73 90.20 93.50 91.82 0.9179
AdaBoost [28] 95.16 93.63 96.71 95.14 0.9519
Decision Tree [28] 79.44 76.81 83.13 79.84 0.7951
EfficientNet [27] 98.99 99.20 98.80
DBM [23] 97.23 98.14 97.68 97.89 0.9771
DBM + MADE [23] 98.37 98.74 98.87 98.14 0.9832
MobileNetv2 + SVM [13] 98.35 97.64 99.20 98.41 0.9912
MobileNetv2 + PF-BAT enhanced FKNN [15] 99.38 99.20 99.60 99.40 0.9958
CNN+ bi-stage feature selection + SVM [26] 98.39 98.21 97.78 98 0.9952
VGG19 [22] 94.04 95.00 94.00 94.50
U-Net++ [6] 92 100
Proposed 98.35 98.02 98.80 98.41 0.9994
Proposed (with features from ‘res5b branch2c’) 99.18 99.20 99.20 99.20 0.9997
Proposed (with classification fusion) 99.38 99.60 99.20 99.40 0.9997

Conclusion

In the present paper, we have investigated the efficacy of different pre-trained network architectures with transfer learning for COVID-19 detection using a limited CT scan dataset. Investigation reveals that the transfer learned ResNet50 model turned out to be the finest by achieving an accuracy value of 98.35% that is superior to the considered models and the existing state-of-the-art works in the literature. Moreover, the potential of the activations from different layers of the learned ResNet50 network is also explored for detection using the established ML algorithms. The exploration reveals that the activations from some of the specific layers of the learned ResNet50 model are quite decisive for classification yielding an accuracy of 99.18% using SVM and LR classifiers. A classification fusion strategy is also proposed that further improvised the accuracy to 99.38% by combining the predictions from the different classifiers via majority voting.

The proposed automated system can assist the healthcare professionals in rapid detection of the virus at different stages.

Funding

This research did not receive any specific grant from funding agencies in the public, commercial, or not-for-profit sectors.

Data Availability

The datasets generated during and/or analyzed during the current study are available in the repository available at the following link: https://www.kaggle.com/plameneduardo/sarscov2-ctscan-dataset.

Declarations

Conflict of interest

The authors declare that they have no competing interests.

Footnotes

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Contributor Information

Taranjit Kaur, Email: Taranjit.Kaur@ee.iitd.ac.in.

Tapan Kumar Gandhi, Email: tgandhi@ee.iitd.ac.in.

References

  • 1.Bernheim A, Mei X, Huang M, Yang Y, Fayad ZA, Zhang N, Diao K, Lin B, Zhu X, Li K, et al. Chest CT findings in coronavirus disease-19 (COVID-19): relationship to duration of infection. Radiology. 2020;66:200463. doi: 10.1148/radiol.2020200463. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 2.Bishop CM. Pattern Recognition and Machine Learning. Berlin: Springer; 2006. [Google Scholar]
  • 3.Chen J, Wu L, Zhang J, Zhang L, Gong D, Zhao Y, Hu S, Wang Y, Hu X, Zheng B, et al. Deep learning-based model for detecting 2019 novel coronavirus pneumonia on high-resolution computed tomography: a prospective study. Sci. Rep. 2020;10(1):1–11. doi: 10.1038/s41598-020-76282-0. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4.Choe J, Lee SM, Do K-H, Lee G, Lee J-G, Lee SM, Seo JB. Deep learning–based image conversion of CT Reconstruction Kernels Improves Radiomics Reproducibility For Pulmonary Nodules Or Masses. Radiology. 2019;292(2):365–373. doi: 10.1148/radiol.2019181960. [DOI] [PubMed] [Google Scholar]
  • 5.Cover T, Hart P. Nearest neighbor pattern classification. IEEE Trans. Inf. Theory. 1967;13(1):21–27. [Google Scholar]
  • 6.Fouladi S, Ebadi MJ, Safaei AA, Bajuri MY, Ahmadian A. Efficient deep neural networks for classification of COVID-19 based on CT images: virtualization via software defined radio. Comput. Commun. 2021;176:234–248. doi: 10.1016/j.comcom.2021.06.011. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7.Goel T, Murugan R, Mirjalili S, Chakrabartty DK. Automatic screening of covid-19 using an optimized generative adversarial network. Cognit. Comput. 2021;66:1–16. doi: 10.1007/s12559-020-09785-7. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.O. Gozes, M. Frid-Adar, H. Greenspan, P.D. Browning, H. Zhang, W. Ji, A. Bernheim, E. Siegel, Rapid ai development cycle for the coronavirus (covid-19) pandemic: initial results for automated detection & patient monitoring using deep learning ct image analysis (2020). arXiv Prepr arXiv:200305037
  • 9.H.P. Graf, E. Cosatto, L. Bottou, I. Dourdanovic, V. Vapnik, Parallel support vector machines: the cascade SVM, in Advances in Neural Information Processing Systems (2005), pp. 521–528
  • 10.K. He, X. Zhang, S. Ren, J. Sun, Deep residual learning for image recognition, in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (2016), pp. 770–778
  • 11.He X, Yang X, Zhang S, Zhao J, Zhang Y, Xing E, Xie P. Sample-efficient deep learning for COVID-19 diagnosis based on CT scans. medRxiv. 2020;66:1–10. [Google Scholar]
  • 12.Huang C, Wang Y, Li X, Ren L, Zhao J, Hu Y, Zhang L, Fan G, Xu J, Gu X, et al. Clinical features of patients infected with 2019 novel coronavirus in Wuhan, China. Lancet. 2020;395(10223):497–506. doi: 10.1016/S0140-6736(20)30183-5. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.T. Kaur, T.K. Gandhi, Automated diagnosis of COVID-19 from CT scans based on concatenation of Mobilenetv2 and ResNet50 features, in CVIP (No. 1) (2020), pp. 149–160
  • 14.Kaur T, Gandhi TK. Deep convolutional neural networks with transfer learning for automated brain image classification. Mach. Vis. Appl. 2020;31(3):1–16. [Google Scholar]
  • 15.Kaur T, Gandhi TK, Panigrahi BK. Automated diagnosis of COVID-19 using deep features and parameter free BAT optimization. IEEE J. Transl. Eng. Heal. Med. 2021;9:1–9. doi: 10.1109/JTEHM.2021.3077142. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16.Koo HJ, Lim S, Choe J, Choi S-H, Sung H, Do K-H. Radiographic and CT features of viral pneumonia. Radiographics. 2018;38(3):719–739. doi: 10.1148/rg.2018170048. [DOI] [PubMed] [Google Scholar]
  • 17.Kundu R, Basak H, Singh PK, Ahmadian A, Ferrara M, Sarkar R. Fuzzy rank-based fusion of CNN models using Gompertz function for screening COVID-19 CT-scans. Sci. Rep. 2021;11(1):1–12. doi: 10.1038/s41598-021-93658-y. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18.Liu DY, Chen HL, Yang B, Lv XE, Li LN, Liu J. Design of an enhanced fuzzy k-nearest neighbor classifier based computer aided diagnostic system for thyroid disease. J. Med. Syst. 2012;36(5):3243–3254. doi: 10.1007/s10916-011-9815-x. [DOI] [PubMed] [Google Scholar]
  • 19.Loey M, Smarandache F, Khalifa NEM. A deep transfer learning model with classical data augmentation and CGAN to detect COVID-19 from chest CT radiography digital images. Neural Comput. Appl. 2020;66:1–13. doi: 10.1007/s00521-020-05437-x. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 20.Mohammad-Rahimi H, Nadimi M, Ghalyanchi-Langeroudi A, Taheri M, Ghafouri-Fard S. Application of machine learning in diagnosis of COVID-19 through X-ray and CT images: a scoping review. Front. Cardiovasc. Med. 2021;8:185. doi: 10.3389/fcvm.2021.638011. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21.Nour M, Cömert Z, Polat K. A novel medical diagnosis model for COVID-19 infection detection based on deep features and Bayesian optimization. Appl. Soft. Comput. 2020;97:106580. doi: 10.1016/j.asoc.2020.106580. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22.Panwar H, Gupta PK, Siddiqui MK, Morales-Menendez R, Bhardwaj P, Singh V. A deep learning and grad-CAM based color visualization approach for fast detection of COVID-19 cases using chest X-ray and CT-Scan images. Chaos Solitons Fract. 2020;140:110190. doi: 10.1016/j.chaos.2020.110190. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 23.Pathak Y, Shukla PK, Arya KV. Deep bidirectional classification model for COVID-19 disease infected patients. IEEE/ACM Trans. Comput. Biol. Bioinf. 2020;6:66. doi: 10.1109/TCBB.2020.3009859. [DOI] [PubMed] [Google Scholar]
  • 24.O.A.B. Penatti, K. Nogueira, J.A. Dos Santos, Do deep features generalize from everyday objects to remote sensing and aerial scenes domains? in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops (2015) pp. 44–51
  • 25.Saygılı A. A new approach for computer-aided detection of coronavirus (COVID-19) from CT and X-ray images using machine learning methods. Appl. Soft Comput. 2021;105:107323. doi: 10.1016/j.asoc.2021.107323. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 26.Sen S, Saha S, Chatterjee S, Mirjalili S, Sarkar R. A bi-stage feature selection approach for COVID-19 prediction using chest CT images. Appl. Intell. 2021;66:1–16. doi: 10.1007/s10489-021-02292-8. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27.Silva P, Luz E, Silva G, Moreira G, Silva R, Lucio D, Menotti D. COVID-19 detection in CT images with deep learning: a voting-based scheme and cross-datasets analysis. Inform. Med. Unlocked. 2020;66:100427. doi: 10.1016/j.imu.2020.100427. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 28.Soares E, Angelov P, Biaso S, Froes MH, Abe DK. SARS-CoV-2 CT-scan dataset: a large dataset of real patients CT scans for SARS-CoV-2 identification. medRxiv. 2020;66:1–8. [Google Scholar]
  • 29.Song Y, Zheng S, Li L, Zhang X, Zhang X, Huang Z, Chen J, Zhao H, Jie Y, Wang R, et al. Deep learning enables accurate diagnosis of novel coronavirus (COVID-19) with CT images. medRxiv. 2020;6:66. doi: 10.1109/TCBB.2021.3065361. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 30.Tan Z, Li X, Gao M, Jiang L. The environmental story during the COVID-19 lockdown: how human activities affect PM2.5 concentration in China? IEEE Geosci. Remote. Sens. Lett. 2020;6:66. doi: 10.1109/LGRS.2020.3040435. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 31.Velavan TP, Meyer CG. The COVID-19 epidemic. Trop. Med. Int. Heal. 2020;25(3):278. doi: 10.1111/tmi.13383. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 32.Wang S, Kang B, Ma J, Zeng X, Xiao M, Guo J, Cai M, Yang J, Li Y, Meng X, et al. A deep learning algorithm using CT images to screen for corona virus disease (COVID-19) Eur. Radiol. 2021;66:1–9. doi: 10.1007/s00330-021-07715-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 33.J. Yosinski, J. Clune, Y. Bengio, H. Lipson, How transferable are features in deep neural networks? in Advances in Neural Information Processing Systems (2014), pp. 3320–3328
  • 34.J. Zhao, Y. Zhang, X. He, P. Xie, COVID-CT-Dataset: a CT scan dataset about COVID-19 (2020). arXiv Prepr arXiv:200313865

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Data Availability Statement

The datasets generated during and/or analyzed during the current study are available in the repository available at the following link: https://www.kaggle.com/plameneduardo/sarscov2-ctscan-dataset.


Articles from Circuits, Systems, and Signal Processing are provided here courtesy of Nature Publishing Group

RESOURCES