Abstract
The COVID-19 pandemic has collapsed the public healthcare systems, along with severely damaging the economy of the world. The SARS-CoV-2 virus also known as the coronavirus, led to community spread, causing the death of more than a million people worldwide. The primary reason for the uncontrolled spread of the virus is the lack of provision for population-wise screening. The apparatus for RT-PCR based COVID-19 detection is scarce and the testing process takes 6–9 h. The test is also not satisfactorily sensitive (71% sensitive only). Hence, Computer-Aided Detection techniques based on deep learning methods can be used in such a scenario using other modalities like chest CT-scan images for more accurate and sensitive screening. In this paper, we propose a method that uses a Sugeno fuzzy integral ensemble of four pre-trained deep learning models, namely, VGG-11, GoogLeNet, SqueezeNet v1.1 and Wide ResNet-50-2, for classification of chest CT-scan images into COVID and Non-COVID categories. The proposed framework has been tested on a publicly available dataset for evaluation and it achieves 98.93% accuracy and 98.93% sensitivity on the same. The model outperforms state-of-the-art methods on the same dataset and proves to be a reliable COVID-19 detector. The relevant source codes for the proposed approach can be found at: https://github.com/Rohit-Kundu/Fuzzy-Integral-Covid-Detection.
Keywords: COVID-19, Ensemble, Fuzzy integral, Computer-aided detection, Deep learning, Sugeno integral, Transfer learning, CT-Scan images
1. Introduction
COVID-19 is an extremely contagious disease that led to a community spread in the world. It was declared to be a public health emergency by the World Health Organization (WHO) in January 2020 and recognized as a pandemic in March 2020. Currently, more than 50 million people worldwide have been affected by the coronavirus, out of which it accounts for more than 1.25 million unfortunate deaths (2.5% mortality rate). However, many positive cases go undetected due to the lack of provision for conducting population-wise testing. The current statistics of the COVID-19 pandemic in the world are shown in Fig. 1 . All the data for the graphs have been obtained from a publicly available source by Ref. [47].
The available standard tests for COVID-19 detection is the Real-Time Polymerase Chain Reaction (RT-PCR) and the rapid antigen test. However, there are some drawbacks of the RT-PCR test [59]. First, it takes 6–9 h for the testing process. Secondly, there is the issue of unavailability of sufficient apparatus for population-wise testing, and thirdly, the RT-PCR test is not wholly reliable, being only 71% sensitive [67]. Due to these reasons, this virus has spread to an uncontrollable extent. On the other hand, the rapid antigen test [34] uses the human blood sample to detect the presence of IgG and IgM antibodies for COVID-19 detection and can produce the result within 15 min. However, the human body takes several days to produce these antibodies after the virus has entered the host, making the antibody test unreliable to detect COVID-19 in its early stages. This can lead to the spread of the virus even before the disease is diagnosed. Even though several vaccines have been developed by experts across the world, it will take a lot of time to vaccinate each individual on the planet, especially with the spread of the new variants of the COVID-19 virus that are emerging due to gene mutation in the virus. Therefore, there is a need to develop Computer-Aided Detection tools for faster and less laborious screening procedures to help control the pandemic. A large part of this research focuses on using deep learning techniques [9,42]. Deep learning is a machine learning technique that uses artificial neural networks for classifying structured or unstructured data through a complex decision-making process. As compared to the RT-PCR test, chest CT-scans are more sensitive [1,14], and more widely available. Although X-Ray images are being used for automated screening, CT-scan images give a more detailed view of the lungs and therefore, are more reliable for this study [39]. The most distinguishing feature of COVID-19 infected chest CT scans is the ”ground-glass opacities” (GGO) that appear inside the lungs [11,33]. They emerge when the alveoli (tiny air sacs) get filled with fluid and appear as a white frosted-glass like appearance in the chest CT-scan [43]. Fig. 2 shows the chest CT-image of a COVID-19 infected patient and a COVID-negative patient. The GGO in the lungs of the infected patient are marked in the image.
Deep learning has been proven successful in biomedical supervised learning applications like [2,15]. In the current image classification task, transfer learning has been employed which refers to the method of re-using a deep learning model used for a specific task, on a separate but related task. It is used when the current problem lacks sufficient data for training the model from scratch. Here, the parameters from the previous task are loaded, and the data from the current task are used for fine-tuning the model.
The present work uses the Ensemble Learning technique for classifier fusion. In doing so, four pre-trained models, namely, VGG-11 by Simonyan et al. [53], GoogLeNet by Szegedy et al. [57], SqueezeNet v1.1 by Iandola et al. [24] and Wide ResNet-50-2 by Zagoruyko et al. [69] have been used, and Sugeno Fuzzy Integral has been used for ensembling the four classifiers and generating the final prediction scores for the images. Ensemble Learning enhances the performance of the constituent models by accounting for increased diversity in predictions. Although simple fusion schemes, such as averaging probability scores, majority voting, etc., have been used in literature, such techniques cannot account for the complex co-occurrences within the data to be classified. They do not account for the different weights of classifiers based on the decision scores obtained at testing time. Therefore, there is a need to generalize the simple fusion schemes, such that the weightage of one classifier may be conditioned upon the weightage of the others, to promote the adaptive importance of different classifiers for every sample image.
In the present work, the Sugeno fuzzy integral has been used for ensembling the aforementioned classifiers to address the shortcomings of using simple fusion techniques. Fuzzy integrals [61] are effective aggregators, that use the degree of uncertainty in the decision scores as additional information for the fusion of classifiers. It can be viewed as a generalisation of aggregation operators on a set of confidence scores that use some weightage given to each information source, called fuzzy measures. The Sugeno fuzzy integral takes into consideration the confidence of predictions by the base learners to assign adaptive weights to them for each input to the model, unlike the simple fusion schemes found in the literature that use fixed pre-determined weights. This concept and its application in the present study are discussed in detail in Section 3.5.
The rest of the paper has been organized into sections as follows: Section 2 provides a brief literature survey of existing COVID-19 detection models; Section 3 explores in detail, the methodology proposed in the current study; Section 4 describes the results obtained on implementation of the proposed model on a publicly available dataset and finally, Section 5 concludes the findings from this paper and outlines some possible extension of this work.
2. Literature survey
Since the outbreak of the COVID-19 pandemic, a diverse range of classification models have been proposed for the automated detection of COVID-19 patients [28]. A large number of these methods use chest radiology images (X-Rays) like [5,63]. Mahmud et al. [38] developed the CovXNet model for differentiating COVID-19 cases from pneumonia using chest X-ray images, wherein they achieved 90.2% accuracy on the multi-class classification problem. Ozturk et al. [42] developed the DarkNet model for the binary and multi-class classification of COVID-19 and pneumonia using chest X-Ray images. They based their model on the You Only Look Once (YOLO) architecture imbibing some modifications and achieved an accuracy of 98.08% for the binary classification problem (COVID-19 vs. No Findings) and 87.02% for the multi-class problem (COVID-19 vs. Pneumonia vs. No Findings).
However, chest CT-scan images are more sensitive as they give a clearer view of the lungs, therefore enabling a more reliable diagnosis. Shi et al. [51] segmented the lungs from the CT-scan image and used machine learning techniques for location-specific feature extraction. The extracted features were fed to their proposed ”infection Size Aware Random Forest (iSAR)” classifier for the final classification. Incorporating 5-fold cross-validation on their dataset, they obtained 87.9% accuracy, 90.7% sensitivity and 83.3% specificity. Machine learning requires the extraction of handcrafted features and often results in redundant features being extracted. Deep learning, on the other hand, learns the important features on its own and provides an end-to-end classification system, and therefore has been used for developing several COVID-19 detection models like [32,50]. Gozes et al. [16] used a fusion of two deep learning models for the classification of COVID-infected patients. One model uses 2D slices of CT scans and the other uses volumetric CT-scan images. Xu et al. [66] proposed a location-attention classification model that uses features extracted by the ResNet-18 pre-trained model from the lung CT images for classification. Burdick et al. [10] conducted a multi-centre clinical trial to predict invasive mechanical ventilation of COVID-19 infected patients within 24 h of the initial encounter with the virus. However, acquiring data from clinical trials requires authorization from several bodies and is not publicly accessible.
Hu et al. [22] developed a weakly supervised COVID classification model that first performs lung segmentation and then uses a deep CNN model that extracts features at 3 different steps for multi-scale learning. Ni et al. [41] proposed a deep learning algorithm for lesion detection, segmentation and segmentation to find pneumonia lesions in COVID-19 patients.
An elaborate study on the efficacy of different transfer learning models (popular and new models) for COVID-19 detection using chest CT images have been conducted by Ardakani et al. [6]. They found ResNet-101 to be the most accurate classifier, obtaining an AUC of 0.994 on a dataset of 1082 CT images. Amyar et al. [3] developed an encoder-decoder network for the segmentation and classification of chest CT images for COVID-19 detection. For this, they use an encoder coupled with two decoders and a multi-layer perceptron for the segmentation, reconstruction and classification tasks. They achieved an average AUC of 97% on a dataset of 1349 patients’ lung CT scans.
Carvalho et al. [12] developed a framework for COVID-19 classification using CT images wherein they extracted deep features from a CNN model and then used a Genetic Algorithm for optimal feature subset selection and classification. Vinod et al. [62] proposed a ”Deep Covix-Net” model for detecting COVID-19 from both X-ray and CT images. They tested their model in both binary-class and multi-class settings achieving accuracies of 97% and 96.8% respectively.
Most of the existing methods use a single classification model. The ensemble learning paradigm is very seldom explored for addressing the COVID-19 detection problem. Thus in this paper, we develop a deep ensemble model for the classification of chest CT images for COVID-19 detection. However, instead of using popular ensemble approaches prevalent in literature, like probability averaging, or weighted probability averaging, we use a fuzzy integral-based ensemble. This leverages adaptive priority to the classifiers on the fly instead of the classical ensemble methods that set the priority given to each classifier beforehand.
2.1. Motivation and contributions
The COVID-19 pandemic has caused the front-line workers to work tirelessly for attending to COVID-infected patients, along with patients homing other diseases, putting their lives at risk. Also, there is no sign of the pandemic stopping, and since population-wise screening is impossible with the current level of RT-PCR test kits available. So, the need for the development of an automated screening model motivated us to conduct this study.
In this paper, we propose a method to improve the efficiency of the Convolutional Neural Network (CNN) models by using a fuzzy integral based ensemble. Four distinct CNN architectures are used to ensure that there are multiple sources of information, manifested by the individual characteristics of the different features. The overall workflow of the proposed model is shown in Fig. 3 . The complementary nature of the data captured by the different CNNs is verified by comparing the statistical divergences from the decision scores of each CNN model. The CNN classifiers are fused by applying the Sugeno Fuzzy integral using which has two advantages, first, it works as a generalization of classical empirical schemes, and second, conditioning the weights of each model based on the probability scores of the individual classifiers capture the complementary information more robustly.
The main contributions of this paper are as follows:
-
1.
Chest CT-scan images are used to develop the proposed model in this study which is both widely available and can be performed faster compared to the tedious RT-PCR testing process.
-
2.
To address the challenge of low availability of publicly available chest CT data of COVID-19, transfer learning models have been used in the first phase of the framework. The complementary natures of the features vectors obtained from the CNNs are verified using the Kullback-Leibler Divergence and Jensen-Shannon Divergence.
-
3.
For using the features of the various models, an ensemble technique has been used based on a fuzzy measure: the Sugeno Integral. Fuzzy integrals are a powerful fusion technique for using the characteristics of all the input information sources.
-
4.
The proposed method outperforms the existing state-of-the-art models achieving very high accuracy (98.93%) and sensitivity (98.93%) compared to the 71% sensitivity of the RT-PCR test.
3. Proposed method
The proposed framework uses four pre-trained deep learning models, which are trained on the ImageNet dataset [13]. The ImageNet dataset is a huge data source consisting of 14 million images belonging to 1000 classes. Training the models on such a large variety of images set the model parameters (weights and biases of the nodes of each layer) optimally, and only minor fine-tuning is generally required for using the trained models on a different classification task. The deep transfer learning models used in this study are:
-
1.
VGG-11 by Simonyan et al. [53].
-
2.
GoogLeNet by Szegedy et al. [57].
-
3.
SqueezeNet v1.1 by Iandola et al. [24].
-
4.
Wide ResNet-50-2 by Zagoruyko et al. [69]
VGG-11 is a very deep network whereas GoogLeNet has inception modules that contain parallel convolutions and hence, both of these networks have a high number of parameters. On the other hand, both SqueezeNet v1.1 and Wide ResNet-50-2 models boast about their computational efficiency due to a far lesser number of parameters for the number of layers they possess. The VGG-11 model can extract very deep features due to the use of linearly progressive convolution layers with small filter sizes. The GoogLeNet model can extract a diverse set of features through its inception modules. The SqueezeNet v1.1 model uses the squeeze and expand blocks to extract deep features while also controlling the number of parameters. Whereas, the WideResNet-50-2 model uses its residual connections to control the feature flow and eliminate the vanishing gradient problem, while concatenating the shallow features with the deep ones for efficient feature extraction. Thus, these four pre-trained models have been chosen for the current study to fuse the diverse properties of each model while also controlling the computational complexity. The total number of parameters in each network are shown in Table 1 .
Table 1.
CNN | Layers | Kernels | Parameters |
---|---|---|---|
VGG-11 | 11 | (3x3) | 30.15 M |
GoogLeNet | 22 | (1x1), (3x3), (7x7) | 11.98 M |
SqueezeNet v1.1 | 14 | (1x1), (3x3) | 0.72 M |
Wide ResNet-50-2 | 50 | (3x3) | 66.8 M |
These pre-trained models are used for the binary classification task at hand and the prediction probability scores for the images are stored. These probability prediction scores are then used for computing the Sugeno Fuzzy Integral to calculate the final predictions on the test set and then the predictions are used to compute the evaluation metrics. The transfer learning models and the fuzzy integral ensemble method adopted in this study are described in the following subsections.
3.1. VGG-11
The VGG models developed by the Visual Geometry Group [53], are some of the deepest CNNs in literature. The introduction of several weight layers (up to 16–19 layers) was made possible by decreasing the size of the convolution filters to 3x3 kernels. The VGG group stresses the fact that the depth of a CNN model is important for visual representations and beneficial for accurately applying to several genres of classification tasks in Computer Vision. The VGG-11 CNN architecture has been shown in Fig. 4 .
3.2. GoogLeNet
The GoogLeNet CNN model by Szegedy et al. [57], the name of which is inspired from LeNet-5 [31] is 22 layers deep network that consists of ”inception modules” instead of uniformly increasing networks. The inception block inspired by Lin et al. [35], allows a large number of units to be accommodated at each stage due to the parallel convolution and pooling, without facing an uncontrolled computational complexity because of the increased number of parameters. For controlling the computational complexity, the architecture uses dimensionality reduction in the inception blocks as shown in Fig. 5 (b), rather than the naive inception block (Fig. 5(a)). The GoogLeNet model proves that an optimal sparse architecture made from the available dense building blocks, i.e., the introduction of inception blocks, improves the performance of neural networks for Computer Vision tasks. The architecture of the GoogLeNet CNN model has been shown in Fig. 6 .
3.3. SqueezeNet v1.1
The SqueezeNet [24] is a CNN model having 50 times lesser parameters than AlexNet [26] model but achieving similar accuracies on the ImageNet data classification task. However, to employ model compression techniques [18,19] allow the SqueezeNet model to be stored under 0.5 MB space which is 510 times smaller than the space occupied by the AlexNet model.
The salient features of the SqueezeNet architecture are as follows:
-
1.
Majority of 3x3 filters have been replaced with 1x1 filters, which have 9 times fewer number parameters.
-
2.
To maintain a small total number of parameters, the number of input channels has been decreased to only 3x3 filters using ”squeeze layers”.
-
3.
The down-sampling of the images has been performed late in the architectures for obtaining larger activation maps from the convolution layers.
-
4.
A ”fire module” has been introduced, which is comprised of a ”squeeze” convolution layer that feeds into an ”expand” layer that is a mix of 1x1 and 3x3 convolutions. The fire module is shown in Fig. 7 . The squeeze layer in the fire module helps to limit the number of input channels. The following hyperparameters describe a fire module: s 1 denotes the number of 1x1 filters in the squeeze layer; e 1 denotes the number of 1x1 filters in the expand layer, and e 3 denotes the number of 3x3 filters in the expand layer.
The architecture of SqueezeNet v1.1 is shown in Fig. 8 .
3.4. Wide ResNet-50-2
Wide ResNets [69] are a genre of CNNs honing residual blocks in their architecture. They have lower depths and greater width than the traditional ResNets [20] achieving state-of-the-art results but consuming much lesser training time. They have proven that the 50 layers deep Wide ResNet (Wide ResNet-50-2) outperformed the 152 layers deep ResNet (ResNet-152) on the ImageNet [13] data classification task. Similar to the previous networks, the final classification layer of the Wide ResNet-50-2 architecture was changed for the binary classification task in the present study. The various residual blocks used in the Wide ResNet architecture has been shown in Fig. 9 , and the complete architecture has been shown in Fig. 10 .
3.5. Ensemble: Sugeno Fuzzy Integral
Ensemble learning is a strategy for fusing the salient properties of two or more base learners. Such a framework performs more robustly than its constituent models because ensembling reduces the variance in the prediction errors. However, most of the traditional ensemble frameworks [29,30] tend to assign a pre-defined classifier weight to compute the ensemble. In this study, we use a fuzzy integral-based approach, wherein instead of assigning a fixed weight to the classifiers, we leverage adaptive importance on the fly while testing. Such a method takes into account the confidence in the predictions of the base learners for every sample to assign importance to the classifier, and thus performs more robustly than the traditional approaches in the literature.
Fuzzy integrals have proven successful in several pattern recognition problems like [36,65]. They are powerful and flexible functions for aggregating information based on a fuzzy measure. This fuzzy measure represents the relevance or importance of the constituent information sources when the aggregation is computed.
A fuzzy measure [55] is the set function f that satisfies the following properties:
-
1.
f(φ) = 0, f(X) = 1
-
2.
A, B ∈ β and A ⊆ B, implies f(A) ≤ f(B)
-
3.
If C n ∈ β, C 1 ⊆ C 2 ⊆ C 3 ⊆…⊆ C n, then limn→∞ f(C n) = f(limn→∞ C n)
where β is a Borel field [37] of an arbitrary set X.
The Sugeno fuzzy-λ measures was introduced by Tahani et al. [60]. Suppose there is a set of scores S = {s 1, s 2, s 3, …, s n} where N is the number of information sources (N = 4 in our case), and e ∈ S. The Sugeno-λ measure is the function f λ: 2S → [0, 1] such that it satisfies the following conditions:
-
1.
fλ(S) = 1
-
2.
if e i ∩ e j = φ, then ∃ a λ > − 1, such that, Equation (1) holds true.
(1) |
Therefore, λ is the real root of the equation as given in Equation (2).
(2) |
The Sugeno Integral [56] is a fuzzy measure integral [46] that can be defined as follows: Suppose (A, μ) is a measurable (Borel) space, and f: X → [0, 1] is a μ-measurable function, then the Sugeno Integral of the measurable function f with respect to a fuzzy measure Ω is given as in Equation (3).
(3) |
where Ω(A i) = Ω({x i, x i+1, x i+2, …, x n}) and.
{f(x 1), f(x 2), f(x 1), …, f(x n)} are the ranges defined as f(x 1) ≤ f(x 2) ≤ f(x 3) ≤ … ≤ f(x n) The algorithm for computing the Sugeno Integral is shown in Algorithm 1.
Algorithm 1
Pseudo code for Sugeno Integral Ensemble
4. Results and discussion
This section describes the dataset used in the current study, and the results obtained by implementing the proposed approach on the dataset. Comparisons to existing standard models in literature have also been drawn to validate the efficacy of the framework.
4.1. Dataset description
The dataset used in the present work is proposed by Soares et al. [54] which they generously made publicly available on Kaggle.2 It consists of a total of 2481 chest CT-scan images divided into two categories: COVID and Non-COVID, distributed unevenly among the two classes. The distribution of images in the training and test set from each class are shown in Table 2 . The images in the dataset have been classified by Soares et al. [54] according to the results from the RT-PCR tests conducted on the patients. Patients with confirmed COVID or Non-COVID cases are only present in the dataset and thus, there is little or no risk of unnecessary bias through the non-representative selection of control patients.
Table 2.
Class | Category | Total | Train set | Test set |
---|---|---|---|---|
1 | COVID | 1252 | 876 | 376 |
2 | Non-COVID | 1229 | 860 | 369 |
4.2. Evaluation criteria
The current image classification problem is a binary classification task and the metrics that have been used for evaluating the performance of the proposed approach on this task are accuracy, specificity, precision, recall or sensitivity and f1-score. The definitions for ”True Positive (TP)”, ”False Positive (FP)”, ”True Negative (TN)” and ”False Negative (FN)” are required to understand and formulate these parameters.
The two classes in the binary classification task are the COVID-19 positive class and the COVID negative class. If the lung CT-image is of a COVID-19 positive patient and the model classifies it to be COVID positive class (i.e., correct classification), then the sample image is said to be True Positive (TP). Now, if the lung CT image is of a COVID positive patient, but the model classifies it to be COVID negative, the sample is called False Positive (FP). Similarly, a COVID negative lung CT image when classified correctly is called True Negative (FN). And, a COVID negative CT-scan image misclassified as COVID positive is called False Negative (FN). The evaluation metrics can now be formulated as Equations (4), (5), (6), (7), (8).
(4) |
(5) |
(6) |
(7) |
(8) |
In the COVID-19 classification task, the number of FP cases is more dangerous, because COVID-19 is an extremely contagious disease, and a COVID-19 positive patient whose diagnosis is determined as COVID negative will infect other people as a super-spreader. However, the FP case is not that big a problem, since a COVID negative patient diagnosed as a positive needs to just maintain precautions under prescribed guidelines since it is a disease with a very low mortality rate. The contagiousness of a virus is inversely related to its deadliness.
4.3. Implementation
The deep transfer learning models used in the current study have been trained for 100 epochs on the COVID-19 binary classification dataset of lung CT-scan images. The hyperparameters used for the model training are shown in Table 3 which have been set experimentally. The loss curves obtained on training by the four models are shown in Fig. 11 .
Table 3.
Hyperparameter | Value |
---|---|
Optimizer | Stochastic Gradient Descent |
Loss function | Cross-entropy |
Batch Size | 16 |
Initial Learning Rate | 0.0001 |
Momentum | 0.99 |
Period of Learning Rate Decay | 10 epochs |
Number of Epochs | 100 |
The probability distribution from the four classifiers has been fused using the Sugeno Fuzzy Integral as described in Section 3.5. The confusion matrix of the final predictions on the test set is shown in Fig. 12 .
The class-wise results obtained by the proposed framework on the test set of the SARS-COV-2 dataset are shown in Table 4 .
Table 4.
Class | Accuracy(%) | Specificity(%) | Precision(%) | Sensitivity(%) | F1-Score(%) |
---|---|---|---|---|---|
COVID | 98.94 | 98.92 | 98.94 | 98.94 | 98.94 |
Non-COVID | 98.92 | 98.94 | 98.92 | 98.92 | 98.92 |
Aggregate | 98.93 | 98.93 | 98.93 | 98.93 | 98.93 |
The most commonly used tests for COVID-19 detection are the RT-PCR test and the rapid antigen-antibody test. The RT-PCR test takes a long time to produce results and is not sufficiently sensitive. The rapid antigen-antibody test relies on the detection of formed antibodies in the human body and produces results in 15 min. However, the human body requires several days to form antibodies, and thus in the initial stage, an infected patient may go undetected for days. The proposed automated framework gives a high classification accuracy and sensitivity by using chest CT-scan images while also being a much faster method. The proposed method can be used as a plug-and-play model, where new test images can be passed through the model to get the predictions by the ensemble mechanism. Thus, the incorporation of the proposed method in the field under consideration is simplistic and COVID-19 cases can be screened more efficiently.
4.4. Verification of complementary features
For establishing the complementary nature (dissimilarity) in the probability vectors generated by the four pre-trained CNN models, two statistical divergence methods are incorporated: the Kullback-Leibler Divergence (KLD) [27] and the Jensen-Shannon Divergence (JSD) [40].
KLD, is closely related to relative entropy, is a measure of dissimilarity between two probability distributions, and is an non-symmetric measure. Suppose we have a probability space X. For every discrete random variable x in X, suppose there are two discrete probability distributions p(x) and q(x) on this same space obtained from two classifiers, i.e., p(x) + q(x) = 1 and p(x), q(x) > 0. Then the KLD (discrete form) from q(x) to p(x) is given as in Equation (9).
(9) |
Since , the symmetrical statistical divergence known as JSD has been proposed, which is derived from the KLD, and is a smoothed version of it. For probability distributions P = p(x) and Q = q(x) on the probability space X, suppose we have another probability distribution , then the discrete form of JSD is given as in Equation (10).
(10) |
The KLD and JSD values between every pair of classifiers from the four pre-trained CNN models are shown in Table 5 .
Table 5.
Distribution P | Distribution Q | DKL(P‖Q) | DJS(P‖Q) |
---|---|---|---|
VGG-11 | GoogLeNet | 0.108 | 0.134 |
GoogLeNet | VGG-11 | 0.456 | |
VGG-11 | SqueezeNet v1.1 | 0.651 | 0.105 |
SqueezeNet v1.1 | VGG-11 | 0.164 | |
VGG-11 | Wide ResNet-50-2 | 0.314 | 0.104 |
Wide ResNet-50-2 | VGG-11 | 0.480 | |
GoogLeNet | SqueezeNet v1.1 | 0.386 | 0.131 |
SqueezeNet v1.1 | GoogLeNet | 0.105 | |
GoogLeNet | Wide ResNet-50-2 | 0.226 | 0.128 |
Wide ResNet-50-2 | GoogLeNet | 0.194 | |
SqueezeNet v1.1 | Wide ResNet-50-2 | 0.297 | 0.106 |
Wide ResNet-50-2 | SqueezeNet v1.1 | 0.326 |
4.5. Comparison with existing methods
The proposed approach is compared with some standard deep transfer learning models, including the ones used for the ensemble in the present study, and the results thus obtained are tabulated in Table 6 . DenseNet161 performs closest to the proposed ensemble model, obtaining an accuracy of 96.91%, but still, the performance is 2.2% lower than the proposed method.
Table 6.
The results obtained by using different ensemble techniques on the SARS-COV-2 dataset [54] are shown in Table 7 . The fuzzy measures used for the Sugeno Integral has been found through extensive experimentation to be {0.35, 0.35, 0.02, 0.28} for {VGG-11, GoogLeNet, SqueezeNet v1.1, Wide ResNet-50-2} giving the best performance on several runs. The weights assigned for computing the weighted average are {0.27, 0.32, 0.23, 0.18}. Other than the weighted average approach, other ensembles that have been used for comparison are the element-wise Multiplication Rule, Maximum Probability Rule, Majority Voting Ensemble, and Average Probability Rule.
Table 7.
Ensemble technique | Accuracy |
---|---|
Multiplication Rule | 95.82% |
Maximum | 96.78% |
Majority Voting | 97.65% |
Average | 97.83% |
Weighted Average | 98.12% |
Proposed Method | 98.93% |
Among the explored ensemble techniques, the weighted average approach gives the result closest to the fuzzy integral ensemble, where the weights have also been assigned experimentally. This weighted average ensemble is a static computation process since there is no scope for dynamically refactoring the weights at the prediction time. Fuzzy integral fusion, on the other hand, can use the confidence scores of each classifier and tailors the weights on the fly, which is a different set for each sample data point. This enables further refinement of predictions even after setting the fuzzy measures.
Existing approaches for the chest CT classification for COVID-19 detection on the dataset used in this study mostly use a single CNN model for classification. Halder et al. [17] and Jaiswal et al. [25] used transfer learning with DenseNet201, for classification. Panwar et al. [44] used transfer learning with VGG-19 along with adding 5 more fully connected layers before the final classification layer. In their experiments on chest X-Ray images, they employed a Grad-CAM based colour visualization approach for interpreting the X-ray scans for further diagnosis. Angelov et al. [4] used the GoogLeNet architecture (non-pretrained model) for extracting deep features, which were used by an MLP classifier for the final classification on the chest CT-images dataset and Sen et al. [49] used a bi-stage deep feature selection approach for the COVID-19 detection problem. The results obtained by these methods on the SARS-COV-2 dataset [54], used in the current study, have been compared with the model proposed by us in Table 8 . The performance of the proposed model is seen to be appreciably better than the existing methods in the literature. As compared to the single CNN model-based approaches by Jaiswal et al. [25], Panwar et al. [44] and Angelov et al. [4], the proposed approach performs significantly better justifying the use of an ensemble framework.
Table 8.
Method | Accuracy(%) | Specificity(%) | Precision(%) | Sensitivity(%) | F1-Score(%) |
---|---|---|---|---|---|
Yazdani et al. [68] | – | 96.20 | – | 85.00 | 90.00 |
Silva et al. [52] | 87.60 | – | – | – | 86.19 |
Angelov et al. [4] | 88.60 | – | 89.70 | 88.60 | 89.15 |
Wang et al. [64] | 90.83 | – | 95.75 | 85.89 | 90.87 |
Panwar et al. [44] | 94.04 | 95.86 | 95.00 | 94.00 | 94.50 |
Sen et al. [49] | 95.32 | – | 95.30 | 95.30 | 95.30 |
Jaiswal et al. [25] | 96.25 | 96.21 | 96.29 | 96.29 | 96.29 |
Halder et al. [17] | 97.00 | 95.00 | 95.00 | 98.00 | 97.00 |
Horry et al. [21] | 97.40 | – | 99.10 | 95.50 | 97.30 |
Kundu et al. [30] | 97.81 | – | 97.77 | 97.81 | 97.77 |
Pathak et al. [45] | 98.37 | – | 98.74 | 98.87 | 98.14 |
Biswas et al. [8] | 98.79 | – | 98.79 | 98.79 | 98.79 |
Banerjee et al. [7] | 98.85 | – | – | – | – |
Proposed Method | 98.93 | 98.93 | 98.93 | 98.93 | 98.93 |
4.6. GradCAM analysis
Gradient-weighted Class Activation Maps or GradCAM [48] provides a visual representation of the regions in an image that is focused on by the CNN model. The features from the convolution layers of the trained CNN model are used to produce a gradient-based weighted activation map which is superimposed on the input image to highlight the regions that are considered discerning by the CNN model.
Fig. 13 shows the GradCAM activations by the four different models used to form the ensemble on some sample images taken from the SARS-COV-2 dataset [54]. As seen from the figure, different models focus on different parts of the same chest CT image. For example, Fig. 13(a) shows a CT image of a COVID-19 infected patient. Fig. 13(b), (c), (d) and (e) respectively show the activations produced by VGG-11, GoogLeNet, SqueezeNet v1.1 and Wide ResNet-50-2. VGG-11 focuses on the lower part of the right lung (left side of the image) and the middle part of the left lung. GoogLeNet focuses on the middle to lower part of the left lung. SqueezeNet v1.1 focuses their attention on the GGO appearing on the upper and lower regions of the right lung and middle of the left lung, and Wide ResNet-50-2 focuses on the entirety of the right lung (primarily on the lower side) where the GGO is more concentrated.
The different models focusing on the different regions of the chest CT images justify that complementary information is being extracted, making the ensemble successful. From the examples shown, it is clear that all the models focus on different regions inside the lungs of the patients. Since GGO starts appearing in the lungs of COVID-19 infected patients from the onset of the disease, the models can classify the images correctly most of the time by looking for the white opacities in the lungs. Thus, the proposed framework is fit for detecting the early stages of COVID-19, unlike the rapid antigen test which is only effective after several days of the infection. This makes the proposed framework a good COVID-19 predictor for use in the practical field.
4.7. Error analysis
The proposed framework works reliably well for even heavily noisy images examples of which are shown in Fig. 14 (a) and (f). Both the images in the figure have been taken in bad imaging conditions, the images being hazy and having improper contrast. In the CT-scan image shown in Fig. 14(f), both the lungs of the patient are not prominent, but even so, the proposed approach could classify the images correctly. For further justification, the GradCAM activations by the four base learners have been shown along with the images. In both cases, the models were able to focus on the correct regions inside the lungs to compute the predictions. And since different regions were focused on the different models, the ensemble was able to incorporate the complementary information provided by each.
However, Fig. 15 (a) and (f) shows two misclassified samples from each category- COVID and Non-COVID. Fig. 15(a) shows the lung CT-scan of a COVID-19 positive patient but has been classified as a Non-COVID case by the proposed model. The image when examined can be seen that does not prominently have GGO in either lung and possibly is a very mildly affected COVID case. The image has most likely been taken in a very early stage, where the COVID-19 virus has not even started affecting the lungs, leading to a wrong prediction. The GradCAM activations in Fig. 15(b)–(e) shows that the base learners were not able to localize on the correct regions of the lungs to extract discerning information. Fig. 15(f) shows the lung CT-scan image of a Non-COVID category patient which has been misclassified to be a COVID case by the proposed model. The bottom portion of the right (image perspective) lung has an erroneous white region, which might have been considered to be the GGO by the model (which is seen in the GradCAM activation maps in Fig. 15(g)–(j)), resulting in a wrong prediction.
Even so, only a few images have been misclassified by the proposed model, and the false positive rate is also low, which is important for a COVID classifier. Although the model fails to accurately classify all the samples provided, it still performs reliably well and is very sensitive to changes in chest CT scans.
5. Conclusion and future work
COVID-19 is an extremely contagious disease whose alarming spread on the global scale has crippled the world health care systems. It has set back the global economy to a near stand-still as the countries of the world impose lockdown throughout the country to try and halt the spread of the virus. As the human toll increases, the need for a conclusive solution becomes direr. The primary reason for the rapid spread of the virus is the lack of equipment for population-wide screening of infected patients. Coupled with this, the gold-standard RT-PCR test is not all-reliable either, being only 71% sensitive, and takes 6–9 h for a single test as well. This demands a fully automated system that can detect COVID-19 through other modalities in a fast and reliable manner.
In this paper, we have proposed a fully automated, computer-based detection method that uses lung CT-scan images for the classification of COVID-19 infected patients. Chest CT scans are more sensitive than RT-PCR tests and X-Rays, while also being more widely available than RT-PCR test kits. The proposed method uses deep learning techniques along with a fuzzy measure based integral, the Sugeno Integral for ensembling the features of four pre-trained deep transfer learning models.
The results from the proposed approach on the SARS-COV-2 dataset [54] outperforms the current state-of-the-art results on the same dataset. From the confusion matrix in Fig. 12, it is clear that the False Positive samples are very less which is important as explained in Section 4.2. Also, from Table 4, it is evident that the accuracy and sensitivity are very high (compare with the 71% sensitivity of RT-PCR). So, it can be concluded that the proposed approach is reliable and can be used for COVID-19 detection.
As discussed before, the proposed model was unable to classify a few instances of mild COVID situations. Since the COVID-19 infection only affects the inside of the lungs, and there are no pleural effusions, a patch-based approach that selects patches from the image for training, or an attention mechanism can be developed in future. Choosing the fuzzy measures is an optimization problem, but for this study, we set the measures through extensive experiments. We aim to address this problem in future as well.
Declaration of competing interest
The authors declare to have no conflicts of interest.
Acknowledgements
We would like to thank the Centre for Microprocessor Applications for Training, Education and Research (CMATER) research laboratory of the Computer Science and Engineering Department, Jadavpur University, Kolkata, India for providing us with the infrastructural support.
Footnotes
References
- 1.Ai T., Yang Z., Hou H., Zhan C., Chen C., Lv W., Tao Q., Sun Z., Xia L. Correlation of chest ct and rt-pcr testing in coronavirus disease 2019 (covid-19) in China: a report of 1014 cases. Radiology. 2020 doi: 10.1148/radiol.2020200642. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 2.Al Hasan M., Chaoji V., Salem S., Zaki M. vol. 30. 2006. Link prediction using supervised learning; pp. 798–805. (SDM06: Workshop on Link Analysis, Counter-terrorism and Security). [Google Scholar]
- 3.Amyar A., Modzelewski R., Li H., Ruan S. Multi-task deep learning based ct imaging analysis for covid-19 pneumonia: classification and segmentation. Comput. Biol. Med. 2020;126 doi: 10.1016/j.compbiomed.2020.104037. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 4.Angelov P., Almeida Soares E. medRxiv; 2020. Explainable-by-design Approach for Covid-19 Classification via Ct-Scan. [Google Scholar]
- 5.Apostolopoulos I.D., Mpesiana T.A. Covid-19: automatic detection from x-ray images utilizing transfer learning with convolutional neural networks. Physical and Engineering Sciences in Medicine. 2020:1. doi: 10.1007/s13246-020-00865-4. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 6.Ardakani A.A., Kanafi A.R., Acharya U.R., Khadem N., Mohammadi A. Application of deep learning technique to manage covid-19 in routine clinical practice using ct images: results of 10 convolutional neural networks. Comput. Biol. Med. 2020;121 doi: 10.1016/j.compbiomed.2020.103795. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 7.Bandyopadhyay R., Basu A., Cuevas E., Sarkar R. Harris hawks optimisation with simulated annealing as a deep feature selection method for screening of covid-19 ct-scans. Appl. Soft Comput. 2021;111 doi: 10.1016/j.asoc.2021.107698. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8.Biswas S., Chatterjee S., Majee A., Sen S., Schwenker F., Sarkar R. Prediction of covid-19 from chest ct images using an ensemble of deep learning models. Appl. Sci. 2021;11:7004. [Google Scholar]
- 9.Brunese L., Mercaldo F., Reginelli A., Santone A. Explainable deep learning for pulmonary disease and coronavirus covid-19 detection from x-rays. Comput. Methods Progr. Biomed. 2020;196 doi: 10.1016/j.cmpb.2020.105608. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 10.Burdick H., Lam C., Mataraso S., Siefkas A., Braden G., Dellinger R.P., McCoy A., Vincent J.-L., Green-Saxena A., Barnes G., et al. Prediction of respiratory decompensation in covid-19 patients using machine learning: the ready trial. Comput. Biol. Med. 2020;124 doi: 10.1016/j.compbiomed.2020.103949. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 11.Caruso D., Zerunian M., Polici M., Pucciarelli F., Polidori T., Rucci C., Guido G., Bracci B., de Dominicis C., Laghi A. Chest ct features of covid-19 in rome, Italy. Radiology. 2020 doi: 10.1148/radiol.2020201237. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 12.Carvalho E.D., Silva R.R., Araújo F.H., de AL Rabelo R., de Carvalho Filho A.O. An approach to the classification of covid-19 based on ct scans using convolutional features and genetic algorithms. Comput. Biol. Med. 2021 doi: 10.1016/j.compbiomed.2021.104744. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 13.Deng J., Dong W., Socher R., Li L.-J., Li K., Fei-Fei L. 2009 IEEE Conference on Computer Vision and Pattern Recognition. Ieee; 2009. Imagenet: a large-scale hierarchical image database; pp. 248–255. [Google Scholar]
- 14.Falaschi Z., Danna P.S., Arioli R., Pasché A., Zagaria D., Percivale I., Tricca S., Barini M., Aquilini F., Andreoni S., et al. Chest ct accuracy in diagnosing covid-19 during the peak of the Italian epidemic: a retrospective correlation with rt-pcr testing and analysis of discordant cases. Eur. J. Radiol. 2020;130 doi: 10.1016/j.ejrad.2020.109192. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15.Ghahramani Z., Jordan M.I. Advances in Neural Information Processing Systems. 1994. Supervised learning from incomplete data via an em approach; pp. 120–127. [Google Scholar]
- 16.Gozes O., Frid-Adar M., Greenspan H., Browning P.D., Zhang H., Ji W., Bernheim A., Siegel E. 2020. Rapid Ai Development Cycle for the Coronavirus (Covid-19) Pandemic: Initial Results for Automated Detection & Patient Monitoring Using Deep Learning Ct Image Analysis. arXiv preprint arXiv:2003.05037. [Google Scholar]
- 17.Halder A., Datta B. Covid-19 detection from lung ct-scan images using transfer learning approach. Mach. Learn.: Sci. Technol. 2021;2 doi: 10.1088/2632-2153/abf22c. [DOI] [Google Scholar]
- 18.Han S., Liu X., Mao H., Pu J., Pedram A., Horowitz M.A., Dally W.J. Eie: efficient inference engine on compressed deep neural network. Comput. Architect. News. 2016;44:243–254. [Google Scholar]
- 19.Han S., Mao H., Dally W.J. 2015. Deep Compression: Compressing Deep Neural Networks with Pruning, Trained Quantization and Huffman Coding. arXiv preprint arXiv:1510.00149. [Google Scholar]
- 20.He K., Zhang X., Ren S., Sun J. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 2016. Deep residual learning for image recognition; pp. 770–778. [Google Scholar]
- 21.Horry M.J., Chakraborty S., Pradhan B., Fallahpoor M., Hossein C., Paul M. engrXiv; 2021. Systematic Investigation into Generalization of Covid-19 Ct Deep Learning Models with Gabor Ensemble for Lung Involvement Scoring. [Google Scholar]
- 22.Hu S., Gao Y., Niu Z., Jiang Y., Li L., Xiao X., Wang M., Fang E.F., Menpes-Smith W., Xia J., et al. Weakly supervised deep learning for covid-19 infection detection and classification from ct images. IEEE Access. 2020;8:118869–118883. [Google Scholar]
- 23.Huang G., Liu Z., Pleiss G., Van Der Maaten L., Weinberger K. Convolutional networks with dense connectivity. IEEE Trans. Pattern Anal. Mach. Intell. 2019 doi: 10.1109/TPAMI.2019.2918284. [DOI] [PubMed] [Google Scholar]
- 24.Iandola F.N., Han S., Moskewicz M.W., Ashraf K., Dally W.J., Keutzer K. 2016. Squeezenet: Alexnet-Level Accuracy with 50x Fewer Parameters And< 0.5 Mb Model Size. arXiv preprint arXiv:1602.07360. [Google Scholar]
- 25.Jaiswal A., Gianchandani N., Singh D., Kumar V., Kaur M. Classification of the covid-19 infected patients using densenet201 based deep transfer learning. J. Biomol. Struct. Dyn. 2020:1–8. doi: 10.1080/07391102.2020.1788642. [DOI] [PubMed] [Google Scholar]
- 26.Krizhevsky A., Sutskever I., Hinton G.E. Imagenet classification with deep convolutional neural networks. Commun. ACM. 2017;60:84–90. [Google Scholar]
- 27.Kullback S., Leibler R.A. On information and sufficiency. Ann. Math. Stat. 1951;22:79–86. [Google Scholar]
- 28.Kundu R., Basak H., Singh P.K., Ahmadian A., Ferrara M., Sarkar R. Fuzzy rank-based fusion of cnn models using gompertz function for screening covid-19 ct-scans. Sci. Rep. 2021;11:1–12. doi: 10.1038/s41598-021-93658-y. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 29.Kundu R., Das R., Geem Z.W., Han G.-T., Sarkar R. Pneumonia detection in chest x-ray images using an ensemble of deep learning models. PLoS One. 2021;16 doi: 10.1371/journal.pone.0256630. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 30.Kundu R., Singh P.K., Ferrara M., Ahmadian A., Sarkar R. ET-NET: an ensemble of transfer learning models for prediction of covid-19 infection through chest ct-scan images. Multimed. Tool. Appl. 2021:1–20. doi: 10.1007/s11042-021-11319-8. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 31.LeCun Y., Boser B., Denker J.S., Henderson D., Howard R.E., Hubbard W., Jackel L.D. Backpropagation applied to handwritten zip code recognition. Neural Comput. 1989;1:541–551. [Google Scholar]
- 32.Li L., Qin L., Xu Z., Yin Y., Wang X., Kong B., Bai J., Lu Y., Fang Z., Song Q., et al. Artificial intelligence distinguishes covid-19 from community acquired pneumonia on chest ct. Radiology. 2020 doi: 10.1148/radiol.2020200905. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 33.Li Y., Xia L. Coronavirus disease 2019 (covid-19): role of chest ct in diagnosis and management. Am. J. Roentgenol. 2020;214:1280–1286. doi: 10.2214/AJR.20.22954. [DOI] [PubMed] [Google Scholar]
- 34.Li Z., Yi Y., Luo X., Xiong N., Liu Y., Li S., Sun R., Wang Y., Hu B., Chen W., et al. Development and clinical application of a rapid igm-igg combined antibody test for sars-cov-2 infection diagnosis. J. Med. Virol. 2020;92:1518–1524. doi: 10.1002/jmv.25727. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 35.Lin M., Chen Q., Yan S. 2013. Network in Network. arXiv preprint arXiv:1312.4400. [Google Scholar]
- 36.Liu X., Ma L., Mathew J. Machinery fault diagnosis based on fuzzy measure and fuzzy integral data fusion techniques. Mech. Syst. Signal Process. 2009;23:690–700. [Google Scholar]
- 37.Mackey G.W. Ergodic theory and virtual groups. Math. Ann. 1966;166:187–207. [Google Scholar]
- 38.Mahmud T., Rahman M.A., Fattah S.A. Covxnet: a multi-dilation convolutional neural network for automatic covid-19 and other pneumonia detection from chest x-ray images with transferable multi-receptive feature optimization. Comput. Biol. Med. 2020;122 doi: 10.1016/j.compbiomed.2020.103869. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 39.Masood A., Sheng B., Li P., Hou X., Wei X., Qin J., Feng D. Computer-assisted decision support system in pulmonary cancer detection and stage classification on ct images. J. Biomed. Inf. 2018;79:117–128. doi: 10.1016/j.jbi.2018.01.005. [DOI] [PubMed] [Google Scholar]
- 40.Menéndez M., Pardo J., Pardo L., Pardo M. The jensen-shannon divergence. J. Franklin Inst. 1997;334:307–318. [Google Scholar]
- 41.Ni Q., Sun Z.Y., Qi L., Chen W., Yang Y., Wang L., Zhang X., Yang L., Fang Y., Xing Z., et al. A deep learning approach to characterize 2019 coronavirus disease (covid-19) pneumonia in chest ct images. Eur. Radiol. 2020:1–11. doi: 10.1007/s00330-020-07044-9. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 42.Ozturk T., Talo M., Yildirim E.A., Baloglu U.B., Yildirim O., Acharya U.R. Automated detection of covid-19 cases using deep neural networks with x-ray images. Comput. Biol. Med. 2020 doi: 10.1016/j.compbiomed.2020.103792. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 43.Pan F., Ye T., Sun P., Gui S., Liang B., Li L., Zheng D., Wang J., Hesketh R.L., Yang L., et al. Time course of lung changes on chest ct during recovery from 2019 novel coronavirus (covid-19) pneumonia. Radiology. 2020;295(3) doi: 10.1148/radiol.2020200370. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 44.Panwar H., Gupta P., Siddiqui M.K., Morales-Menendez R., Bhardwaj P., Singh V. A deep learning and grad-cam based color visualization approach for fast detection of covid-19 cases using chest x-ray and ct-scan images. Chaos, Solit. Fractals. 2020 doi: 10.1016/j.chaos.2020.110190. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 45.Pathak Y., Shukla P.K., Arya K. Deep bidirectional classification model for covid-19 disease infected patients. IEEE ACM Trans. Comput. Biol. Bioinf. 2020;18(4) doi: 10.1109/TCBB.2020.3009859. [DOI] [PubMed] [Google Scholar]
- 46.Ralescu D., Adams G. The fuzzy integral. J. Math. Anal. Appl. 1980;75:562–570. [Google Scholar]
- 47.Ritchie H., Mathieu E., Rodés-Guirao L., Appel C., Giattino C., Ortiz-Ospina E., Hasell J., Macdonald B., Beltekian D., Roser M. Coronavirus pandemic (covid-19) Our World in Data. 2020 https://ourworldindata.org/coronavirus [Google Scholar]
- 48.Selvaraju R.R., Cogswell M., Das A., Vedantam R., Parikh D., Batra D. 2017 IEEE International Conference on Computer Vision (ICCV) 2017. Grad-cam: visual explanations from deep networks via gradient-based localization; pp. 618–626. [DOI] [Google Scholar]
- 49.Sen S., Saha S., Chatterjee S., Mirjalili S., Sarkar R. A bi-stage feature selection approach for covid-19 prediction using chest ct images. Appl. Intell. 2021:1–16. doi: 10.1007/s10489-021-02292-8. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 50.Shan F., Gao Y., Wang J., Shi W., Shi N., Han M., Xue Z., Shi Y. 2020. Lung Infection Quantification of Covid-19 in Ct Images with Deep Learning. arXiv preprint arXiv:2003.04655. [Google Scholar]
- 51.Shi F., Xia L., Shan F., Wu D., Wei Y., Yuan H., Jiang H., Gao Y., Sui H., Shen D. 2020. Large-scale Screening of Covid-19 from Community Acquired Pneumonia Using Infection Size-Aware Classification. arXiv preprint arXiv:2003.09860. [DOI] [PubMed] [Google Scholar]
- 52.Silva P., Luz E., Silva G., Moreira G., Silva R., Lucio D., Menotti D. Covid-19 detection in ct images with deep learning: a voting-based scheme and cross-datasets analysis. Informatics in medicine unlocked. 2020;20 doi: 10.1016/j.imu.2020.100427. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 53.Simonyan K., Zisserman A. 2014. Very Deep Convolutional Networks for Large-Scale Image Recognition. arXiv preprint arXiv:1409.1556. [Google Scholar]
- 54.Soares E., Angelov P., Biaso S., Froes M.H., Abe D.K. medRxiv; 2020. Sars-cov-2 Ct-Scan Dataset: A Large Dataset of Real Patients Ct Scans for Sars-Cov-2 Identification. [Google Scholar]
- 55.Sugeno M. Readings in Fuzzy Sets for Intelligent Systems. Elsevier; 1993. Fuzzy measures and fuzzy integrals—a survey; pp. 251–257. [Google Scholar]
- 56.Sugeno M., Murofushi T. Pseudo-additive measures and integrals. J. Math. Anal. Appl. 1987;122:197–222. [Google Scholar]
- 57.Szegedy C., Liu W., Jia Y., Sermanet P., Reed S., Anguelov D., Erhan D., Vanhoucke V., Rabinovich A. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 2015. Going deeper with convolutions; pp. 1–9. [Google Scholar]
- 58.Szegedy C., Vanhoucke V., Ioffe S., Shlens J., Wojna Z. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 2016. Rethinking the inception architecture for computer vision; pp. 2818–2826. [Google Scholar]
- 59.Tahamtan A., Ardebili A. 2020. Real-time rt-pcr in covid-19 detection: issues affecting the results. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 60.Tahani H., Keller J.M. Information fusion in computer vision using the fuzzy integral. IEEE Transactions on systems, Man, and Cybernetics. 1990;20:733–741. [Google Scholar]
- 61.Torra V., Narukawa Y. The interpretation of fuzzy integrals and their application to fuzzy systems. Int. J. Approx. Reason. 2006;41:43–58. [Google Scholar]
- 62.Vinod D.N., Jeyavadhanam B.R., Zungeru A.M., Prabaharan S. Fully automated unified prognosis of covid-19 chest x-ray/ct scan images using deep covix-net model. Comput. Biol. Med. 2021 doi: 10.1016/j.compbiomed.2021.104729. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 63.Wang L., Wong A. 2020. Covid-net: A Tailored Deep Convolutional Neural Network Design for Detection of Covid-19 Cases from Chest X-Ray Images. arXiv preprint arXiv:2003.09871. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 64.Wang Z., Liu Q., Dou Q. Contrastive cross-site learning with redesigned net for covid-19 ct classification. IEEE.J. Biomed.Health Inf. 2020;24:2806–2813. doi: 10.1109/JBHI.2020.3023246. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 65.Wu S.-L., Liu Y.-T., Hsieh T.-Y., Lin Y.-Y., Chen C.-Y., Chuang C.-H., Lin C.-T. Fuzzy integral with particle swarm optimization for a motor-imagery-based brain–computer interface. IEEE Trans. Fuzzy Syst. 2016;25:21–28. [Google Scholar]
- 66.Xu X., Jiang X., Ma C., Du P., Li X., Lv S., Yu L., Ni Q., Chen Y., Su J., et al. A deep learning system to screen novel coronavirus disease 2019 pneumonia. Engineering. 2020;6(10) doi: 10.1016/j.eng.2020.04.010. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 67.Yap J.C.-H., Ang I.Y.H., Tan S.H.X., Jacinta I., Pei C., Lewis R.F., Qian Y., Yap R.K.S., Ng B.X.Y., Tan H.Y. Covid-19 science report: diagnostics. ScholarBank@NUS. 2020 doi: 10.25540/e3y2-aqye. [DOI] [Google Scholar]
- 68.Yazdani S., Minaee S., Kafieh R., Saeedizadeh N., Sonka M. 2020. Covid Ct-Net: Predicting Covid-19 from Chest Ct Images Using Attentional Convolutional Network. arXiv preprint arXiv:2009.05096. [Google Scholar]
- 69.Zagoruyko S., Komodakis N. 2016. Wide Residual Networks. arXiv preprint arXiv:1605.07146. [Google Scholar]