Abstract
The world has experienced epidemics of coronavirus infections several times over the last two decades. Recent studies have shown that using medical imaging techniques can be useful in developing an automatic computer-aided diagnosis system to detect pandemic diseases with high accuracy at an early stage. In this study, a large margin piecewise linear classifier was developed to diagnose COVID-19 compared to a wide range of viral pneumonia, including SARS and MERS, using chest x-ray images. In the proposed method, a preprocessing pipeline was employed. Moreover, deep pre- and post-rectified linear unit (ReLU) features were extracted using the well-known VGG-Net19, which was fine-tuned to optimize transfer learning. Afterward, the canonical correlation analysis was performed for feature fusion, and fused deep features were passed into the LMPL classifier. The introduced method reached the highest performance in comparison with related state-of-the-art methods for two different schemes (normal, COVID-19, and typical viral pneumonia) and (COVID-19, SARS, and MERS pneumonia) with 99.39% and 98.86% classification accuracy, respectively.
Keywords: Computer-aided diagnosis (CAD), COVID-19, Deep feature extraction, Large margin classifier, MERS, SARS, X-ray
1. Introduction
The COVID-19 virus is very contagious and has been spreading worldwide, leading to an ongoing pandemic. The world has encountered occurrences of other severe coronavirus infections over the last few decades, including [1]:
-
•
The Severe Acute Respiratory Syndrome (SARS) epidemic originating in Guangdong, China in 2002–2003;
-
•
The Middle East Respiratory Syndrome (MERS) epidemic originating in Jeddah, Saudi Arabia in 2011;
-
•
The current COVID-19 pandemic that originated in Wuhan, China in 2019.
Various types of medical images can be used to diagnose this disease. Two commonly used types are Computed Tomography (CT) and x-rays of the chest [2].
In emerging diseases such as COVID-19, the number of available samples is initially small and, sometimes, insufficient to train deep learning models well because they tend to overfit the training data. The present study introduces a large margin piecewise linear (LMPL) classifier as a proper learning method to get the most out of the available training data. We took advantage of the efficiency of large margin classifiers to construct a piecewise linear model. The new algorithm, called LMPL, is a novel formulation with interesting benefits, such as more accurate results than the traditional approaches like support vector machine (SVM) classifiers and -Nearest Neighbors (NN). The proposed LMPL addresses the following challenges that are of great importance in machine learning:
-
•
Small sample size: Medical datasets are difficult to find due to privacy constraints that limit data sharing;
-
•
Overfitting to training data: This problem limits the generality of models to unseen data, but it is essential, particularly in transfer learning tasks, where the distribution of source and target can be so different.
-
•
Nonlinearity: Most classes or clusters in real-world problems are not linearly separable.
-
•
Multiclass classification: Some classifiers, e.g., SVMs, have a binary structure and need extensions to solve this problem, such as the One versus One (OvO) and One versus All (OvA) approaches.
-
•
Multi-modality: When a class has several modalities, its samples are located in various positions of the search space. This makes it difficult for the most common classifiers to work well.
In this paper, the VGG-net model [3] was used for feature extraction. Indeed, the deep VGG-Net19 model was retrained by transfer learning to learn deep features of x-ray images. The deep features were obtained from the seventh fully connected layer, i.e., the last layer before the output. The activation values of both pre- and post-rectified linear units (ReLU) were marked as deep features with feature vectors of dimension 4096. A preprocessing pipeline was developed for quality improvement of x-ray images. The canonical correlation analysis (CCA) was performed for feature fusion [4], and the fused features were passed to the LMPL classifier.
The introduced approach was applied for two classification tasks in this study.
-
I.
Distinguishing COVID-19 from typical viral pneumonia and normal x-ray images
-
II.
Distinguishing COVID-19 from the two severe CoV family members, i.e., SARS and MERS
1.1. Contribution
The contributions of this paper are outlined as follows:
-
•
In order to learn discriminant and robust feature representations, a novel COVID-19 diagnostic method is introduced that integrates a deep feature extraction and fusion methodology.
-
•
To gain the most informative knowledge from limited labeled data, an LMPL classifier is introduced using a cellular structure to better exploit the limited training data.
2. Related work
Since the spread of COVID-19, there has been an increasing effort to advance deep models for COVID-19 detection using medical images, such as CT scans and x-rays. In a review of the literature, some of the recent developments in deep networks for COVID-19 diagnosis are mentioned herein. Ozturk et al. [5] developed DarkCovidNet to diagnose COVID-19 using raw chest x-ray images. Their network reached the accuracy of 98.08% and 87.02% in binary and multiclass schemes. Taik et al. [6] introduced the COVIDGR-1.0 dataset and proposed the COVID-SDNet method that combines segmentation, data-augmentation, and data transformation. They achieved a high generalization accuracy of 97.72%. Mahmud et al. [7] proposed a deep CNN called CovXNet, which uses depthwise dilated convolution for efficient feature extraction from x-ray images. At first, a large number of normal and Non-COVID pneumonia (viral and bacterial) were used to train the proposed CovXNet. Then, transfer learning was used to fine-tune and retrain the net by a smaller number of COVID-19 and other pneumonia cases. CovXNet was designed and trained by various image resolutions, and a stacking algorithm was employed for optimization. CovXNet achieved an accuracy of 90.2% for multiclass classification into normal, COVID19, viral, and bacterial pneumonia. Rahman et al. [8] compiled a large x-ray dataset with its corresponding lung mask to identify COVID-19. They investigated the effect of different image enhancement techniques on COVID-19 detection. Various deep CNNs were trained on plain and segmented lung images. The method reached the best performance with 96.2% accuracy in plain images by ChexNet and 95.11% classification accuracy in segmented lung images by DenseNet201. Jain et al. [9] compared multiple deep CNN models, including Inception V3, Xception, and ResNeXt, among which Xception had the best performance with the highest accuracy of 97.97% in classifying COVID-19 using chest x-ray images. Toğaçar et al. [10] restructured the COVID-19 dataset by employing the fuzzy color technique as a preprocessing method. The images organized together with the original images were stacked. SqueezNet and MobileNetV2 were used as deep feature extractors, along with the SVM classifier, and trained on the stacked dataset. They obtained 96.28% classification accuracy using MobileNetV2. Hemdan et al. [11] introduced COVIDX-Net, which consists of seven CNN models. They employed a deep learning model to detect COVID-19 using x-ray images. Karakanis et al. [12] proposed a lightweight deep learning architecture to detect COVID-19 from x-ray images. The authors used a conditional generative adversarial network for image augmentation by generating syntactic x-ray images. They achieved 98.7% accuracy in binary classification between COVID-19 and normal pneumonia and 98.3% accuracy in a three-class model with normal, COVID-19, and bacterial pneumonia. Wang and Wong [13] developed an accurate deep network named COVID-Net for COVID-19 diagnosis with a classification accuracy of 92.4% in the two-class problem. Ioannis et al. [14] used 224 COVID-19 images to develop a deep model. They achieved the accuracy of 98.75% for a two-class problem and 93.48% for a three-class problem. Jin et al. [15] introduced a hybrid ensemble method that includes three steps to identify COVID-19 using x-ray images. At first, pre-trained Alexnet is used as a feature extractor; then, the ReliefF method is adopted to sort extracted deep features; and after reducing features' dimension, an SVM classifier is trained on selected features. They obtained 98.64% overall accuracy in multiclass classification of normal, COVID-19, and viral pneumonia. Narin et al. [16] employed the ResNet50 model for COVID-19 diagnosis and obtained an accuracy of 98% using chest x-ray images. Sethy and Behera [17] extracted deep features from x-ray images employing several CNN models and classified them with an SVM classifier. They reported the best performance for the model using the ResNet50 feature extraction and an SVM classifier. Minaee et al. [18] developed a deep learning framework to predict COVID-19 in chest x-rays. They fine-tuned four CNN models, including ResNet18, ResNet50, SqueezeNet, and DenseNet, and achieved promising results in several tasks. Various other studies have also been recently conducted on COVID-19 detection, employing several deep learning models with CT images [2,[19], [20], [21], [22], [23]] and Lung Ultrasound (LUS) [24].
A few studies have focused on the diagnosis of SARS and MERS with x-ray chest images. In a study on MERS, Hamimi [25] reported that features like the indicators of pneumonia could be found in the chest CT scan and x-ray images. Xie et al. [26] introduced a model to distinguish SARS from typical pneumonia using X-ray images. They employed three conventional classifiers: neural networks, C4.5, and Classification And Regression Tree (CART). Tahir et al. [27] employed CNN models for COVID-19 detection using a dataset containing SARS and MERS images. The authors developed a novel image preprocessing technique and deep learning algorithms and reported the outperformance of four CNN methods, including SqueezeNet, ResNet18, Inceptionv3, and DenseNet201, among which InceptionV3 achieved the highest accuracy (98.22%). Abbas et al. [28] proposed a method based on DeTraC deep CNN to classify COVID-19 using chest x-ray images. The DeTraC model consisted of three phases, i.e., Decomposition, Transfer, and Composition. DeTraC uses a class decomposition mechanism to investigate class boundaries with any irregularities in the image dataset. They achieved 93.1% accuracy in a three-class problem, including normal, COVID-19, and SARS pneumonia. Xuanyang et al. [29] introduced a SARS detection system using a lung segmentation technique and feature extraction. They applied and compared several classification methods, including neural networks, decision trees, and regression trees. The highest detection accuracy was achieved from the regression trees. However, for pneumonia classification, NN-based models produced reasonably good results.
As presented in this section, researchers worldwide have introduced several approaches to detect COVID-19, as well as SARS or MERS pneumonia. Since the features of lung infections are the same in these diseases, it is problematic for an expert to distinguish them; so, well-trained models are needed to develop an accurate and useful method in the treatment and decision-making. This study devises an effective method to distinguish the current COVID-19 from a wide range of viral pneumonia, including the same CoV family members like SARS and MERS.
3. Materials and methods
This section provides the details of the proposed COVID-19 diagnostic method. After preprocessing, in order to enhance x-ray images, the popular VGG-Net is used as a deep feature extractor from x-ray images, and the CCA feature fusion technique is applied to merge the extracted features. The LMPL classifier is introduced for coronavirus detection. Fig. 1 illustrates a schema of the proposed COVID-19 diagnostic model.
3.1. X-ray dataset
In this research, several open-access chest x-ray datasets are combined as a new large dataset on the coronavirus family. The combined dataset consists of six classes whose details are summarized in Table 1 . The total number of images of the dataset is 3387. The dataset is balanced by image augmentation to reach 6179 images, as mentioned in Sec. III. 3) Image augmentation.
Table 1.
Types | No. of X-ray Images | Source Database |
---|---|---|
Covid-19 | 423 | SIRM-ITALIAN [31] |
Novel Coronavirus 2019 Dataset [32] | ||
Radiopaedia [33] | ||
Chest Imaging (Spain) at thread reader [34] | ||
COVID-19 Radiography Database [35] | ||
SARS | 134 | SARS, MERS X-ray Images Dataset [36] |
MERS | 144 | SARS, MERS X-ray Images Dataset [36] |
Normal | 1341 | Chest X-ray Images (pneumonia) [37] |
Typical Viral Pneumonia | 1345 | Chest X-ray Images (pneumonia) [37] |
3.2. Image preprocessing pipeline
Medical images often contain a low dynamic range. This affects the overall performance of image processing approaches. Hence, it may become harder to evaluate them visually [8]. The preprocessing is performed to enhance the image information and visual quality through contrast enhancement. Another critical issue in medical imaging is imbalanced data, where the number of positive cases is lower than that of the negative ones.
-
1)
Resizing x-ray images
In order to feed the images into CNNs, they should be resized to the input dimensions of the deep network. For VGG-Net, images are resized to 224 × 224 pixels.
-
2)
Contrast limited adaptive histogram equalization
Histogram equalization (HE) is an ordinary method widely used to improve the contrast of an image and make algorithms more robust under varying conditions. HE produces a uniform distribution by computing the transformation function and it can help adjust an image by spreading out the most common intensity values of pixels in the range of 0 (black) to 255 (white). The contrast limited adaptive histogram equalization (CLAHE) [30] is an improved HE approach that enhances local contrast and edges in small image regions by calculating the equalization for each part in the image. Fig. 2 illustrates the image enhancement by the CLAHE method. It can be seen that the histogram is stretched around the entire spectrum and over all pixels of the image, which should facilitate the extraction of homogeneous features across all datasets.
-
3)
Image augmentation
In order to extend the number of training samples and avoid the risk of overfitting, the data is augmented by applying different types of transformations. This makes the dataset suitable for deep learning. Deep learning algorithms such as CNNs generally outperform a larger dataset rather than a smaller one. Besides, the imbalanced class distribution has a considerable impact on the effectiveness of the classification models. Consequently, the size of classes is balanced in the training set through data augmentation (see Fig. 3 ). Translation can be a very useful transformation to avoid positional bias in the data. In other words, translation is very helpful as most lesions can be located almost anywhere in the image.
In this work, rotation and translation operations are used to increase the number of images (upsampling) and balance the class instances with augmentation. Images are rotated 5, 10, 15, and 30°, and instances are also translated in a horizontal or vertical direction by 5%, 10%, 15% in the SARS and MERS classes that have few images, i.e., 134 and 144 ones, respectively. Finally, 1072 and 1152 images are obtained in the SARS and MERS classes, respectively. In the COVID-19 class with 423 instances, 15-degree rotation and 10% translation are performed to reach 1269 images. No augmentation is applied on normal and typical viral Pneumonia with 1341 and 1345 images. All augmented x-ray images can be found in the Zenodo repository (https://zenodo.org/record/4691987).
3.3. Deep feature extraction
Deep features are extracted by pre-trained CNN models. The fully connected (FC) layers are used to extract deep features. Since VGG-Net [3] was successful in extracting deep features for various tasks, the VGG-Net is employed as the main feature extractor to obtain robust automatic features.
The input image size in VGG networks is 224 × 224 × 2. Fig. 4 illustrates an architectural overview of the VGG-Net. When applying VGG-Net as a deep feature extraction network, the classification layer, i.e., the last fully connected layer, is removed. The features can be extracted as raw values or after being transformed by a ReLU, where an output is mapped to . The activation values from the last hidden layer are extracted as the first deep feature vector (pre-ReLU features, 4096), while the post-ReLU feature set (post-ReLU features, 4096) is extracted as the second feature vector. For datasets not large enough, transfer learning can be used. In this paper, fine-tuning is applied with transfer learning. This is done by retraining VGG-Net (the weights of the model trained using the ImageNet dataset) on the COVID-19 x-ray images and transferring the deep features that are extracted to achieve better performance. When class labels are used to fine-tune the weights of the network, the extracted features are more discriminative in classifying the problem to distinguish COVID-19 from other viral pneumonia.
3.4. Deep feature fusion
The obtained deep features are combined to acquire more informative fusion features. This can help to minimize the impact of inadequate features obtained from a single CNN. Indeed, feature fusion efficiently produces features that comprise rich information describing the image well. Appropriately combining two or more feature sets is not a trivial task. In this study, a well-known parallel feature fusion technique, i.e., CCA, is used. Feature fusion based on CCA creates two sets of transformations based on the correlation between two feature vectors resulting in transformed features with a higher correlation than the original feature sets. Suppose that and are two feature sets, where indicates the number of samples, and and denote the dimensions of and , respectively. Let and represent the covariance matrices of and , respectively, and be the covariance matrix between sets, in which . The overall covariance matrix can be presented by (1):
(1) |
The goal of CCA is to define a linear combination of these feature sets, as shown in (2).
(2) |
where are associated coefficients of and, respectively, that maximize the pair-wise correlation across the two feature sets. This objective is presented in (3):
(3) |
where, , , and . The combination of the transformed features is achieved by addition (summation) or concatenation, as follows in (4), (5).
(4) |
(5) |
where the resulting parallel fused feature matrix contains the canonical correlation discriminant features. Fig. 5 describes the CCA deep feature fusion procedure, where deep features are extracted by VGG-net from the input x-ray images.
3.5. Large margin piecewise linear (LMPL) classifier
In the last stage, fused features are used for the classification by the proposed LMPL classifier. A cellular structure is developed in this study such that space is partitioned by more than one hyper-plane, unlike in traditional models, e.g., SVMs, into some regions called cells. At first, a grid of cells is generated by predefined hyper-planes, as shown in Fig. 6 (a).
In Fig. 6(a), there are three classes: star (red), square (blue), and diamond (violet). Then, a cell table is created to store the information of each cell, such as the number of samples and its class label. The class labels are assigned to cells based on majority voting in each cell. Afterward, the parameters of the hyperplanes are tuned one by one in iterations. In other words, one of the hyper-planes is adjusted, considering that the others are fixed. Changing this hyperplane changes the region of cells bound to it from one side. Hence, the covering cell of a few instances may change. According to the actual label of each sample and the associated label of the cells, samples may prefer to be on a specific side of the hyper-plane. This preference is considered their virtual labels. For example, if an instance prefers the right side of the hyperplane, its virtual label is , and if it chooses the left side, its virtual label is . After virtual labeling, there is a binary classification problem for the hyperplane under tuning. There are three sample groups in virtual labeling, as follows:
-
1.
Normal: These samples are classified correctly on one side of the hyperplane. Their loss function is defined by the famous hinge loss, as shown in (6):
(6) |
where is the virtual label of sample , and the hinge loss function is convex.
-
2.
Negative don't care: These samples are classified incorrectly on both sides of the hyperplane. Therefore, it is not important on which side they are located, unless it is tested and proved that changing the position of samples affects the hyperplane. If the sample is located on the positive side, it tries to be located on the negative side, and vice versa. In other words, they are forced to be near the hyper-plane.
This loss function is defined by (7). By this function, the samples always get a penalty since they are misclassified.
(7) |
-
3.
Positive don't care: These samples are classified correctly on both sides of the hyperplane. Like in the second gro (negative don't care), it is not important on which side of the hyperplane they are. However, in this case, their distance to the hyper-plane is tested to be maximized. The total loss is defined by (8):
(8) |
The positive don't care group is discarded here since its samples are classified correctly and help us to preserve convexity in the final objective function. Therefore, according to the two first groups, the proposed LMPL classifier optimizes each hyperplane based on the following objective function (9):
(9) |
The scalar values and control the balance between the structural and empirical error. In this paper, both and are set to 1000.
The defined objective is optimized by quadratic programming. After optimizing each hyperplane, the cell table is updated in each iteration. Additional hyperplanes that fall out of the search space and are not useful in the classification are removed. After several iterations, the final piecewise linear decision boundary is constructed. In the test step, an unseen sample is classified.
As mentioned, the LMPL classifier utilizes the cellular structure (see Fig. 6. (a)). Each cell forms a rule with a decision space regarding its position with respect to the hyper-planes. By assigning a label to each cell, both multiclass classification and multi-modality are structurally supported (the same as a decision tree). However, contrary to decision trees, the separator hyper-planes are shared in many decision spaces to use both local and global attributes of the training sample for class separation. These cells are then adjusted by tuning the initial hyperplanes from Fig. 6 (a) to Fig. 6 (b) to optimize a large margin classifier inspired by SVMs. The use of global hyper-planes and the large margin objective function allows the method to be generalized and prevents the training data's overfitting. Fig. 6. (b) illustrates the obtained LMPL piecewise linear decision boundary where three classes are classified well by five remaining hyperplanes. As presented, the decision boundaries are in a piecewise linear form that makes the LMPL non-linearly separatable samples in the original space instead of using kernel methods. After adjusting hyperplanes, additional hyperplanes that are not useful in the classification are removed. Therefore, regarding not only the sample size but also the distribution and complexity of the decision boundaries, the complexity of the model is tuned by removing redundant hyperplanes. The large-margin approach forces the redundant hyperplanes to be removed. This structure not only makes it feasible to work with small sample-size problems but also prevents overfitting of the training data by using a large-margin approach for all the hyperplanes.
Fig. 7 represents classes by t-distributed Stochastic Neighbor Embedding (t-SNE) in 2-D and 3-D for both experiments. An extended version of this figure is shown in Fig. 8 . A multiclass problem is shown with multi-modal small samples in each class. Class samples are segmented into parts. Hence, the class pairs may not be linearly separable or they would have a very small margin if can be separated, whereas others can be piecewise linearly separated with acceptable margins.
Fig. 8 illustrates the obtained decision boundary of the introduced LMPL classifier on t-SNE embedding features with four initial hyperplanes. After a few iterations, the LMPL adjusts hyperplanes and removes additional hyperplanes. Moreover, piecewise linear decision boundaries are obtained that can separate classes. The initial number of hyperplanes is random (possibly 8 or 16), and the number of iterations can be determined manually (for example, 100). Algorithm 1 indicates the whole process of the proposed diagnostic method, and the steps of the introduced LMPL classifier are shown in Algorithm 2.
Algorithm 1
The Proposed Diagnostic Method.
Algorithm 2
Large Margin Piecewise Linear (LMPL) Classifier.
4. Experimental results
4.1. Experimental setup
The 5-fold cross-validation (CV) was utilized. Of the original labeled data, 80% were used as the training set, and the remaining samples (20%) were employed as the test set to evaluate the model (unseen instances). Results were averaged to produce the final confusion matrix. MATLAB (2020a) was utilized for training, evaluating, and testing different algorithms. The experiments were run on a computer with an Intel Core i7 @4.0 GHz processor, 24 GB RAM, 128 GB SSD, and a 64-bit Windows 10 operating system. We used mini-batch gradient descent optimization with momentum update of and learning rates of and . Besides, we selected 16 images as the size of mini-batch and five back-propagation epochs. The details of the training and test instances in each fold can be found in the Zenodo repository (https://zenodo.org/record/4691987).
4.2. Experimental results and analysis
In this study, two experiments were carried out for two different classification problems:
-
•
Experiment1: Normal, COVID-19, and typical viral pneumonia
-
•
Experiment2: COVID-19, SARS, and MERS pneumonia
-
1)
Performance analysis
Table 2, Table 3 compare several deep CNNs as an end-to-end network and as a feature extractor with the proposed LMPL (named ). Experiments show that the proposed LMPL classifier performed better than the conventional SoftMax classifier in all CNNs. VGGNet and were the best among the networks in both experiments. As illustrated in Table 4, Table 5 , the proposed LMPL outperformed the conventional classifiers in terms of performance metrics in both Experiments 1 and 2, compared to ensemble methods, including AdaboostM2, Total Boost, and Random Forests consisting of 100 trees as weak learners. Moreover, Table 6, Table 7 demonstrate the comparison of different feature fusion methods and pre-ReLU and post-ReLU deep features without fusion. Concatenation and summation techniques were used in Classical, CCA, and DCA feature fusion methods. The sum and concat operations generate 4096 and 4096 × 2 = 8192 final fused features, respectively.
Table 2.
Method | Performance Metrics (%) |
|||||
---|---|---|---|---|---|---|
Feature ExtractorLayer | Sensitivity (Recall) | Precision (PPV) | F1-score | Accuracy | Average Rank | |
96.40 (4) | 96.35 (5) | 96.37 (4) | 96.17 (4) | 4.3 | ||
97.60 (7) | 98.06 (6) | 97.82 (6) | 97.52 (6) | 6.3 | ||
Avg_pool | 96.20 (7) | 96.16 (6) | 96.17 (6) | 95.91 (6) | 6.3 | |
98.43 (3) | 98.49 (4) | 98.45 (4) | 98.10 (4) | 3.8 | ||
Avg_pool | 94.04 (10) | 92.61 (10) | 93.28 (11) | 93.18 (10) | 10.3 | |
97.48 (8) | 97.07 (10) | 97.26 (10) | 97.01 (9) | 9.3 | ||
pool5 | 96.40 (4) | 96.40 (4) | 96.35 (5) | 96.01 (5) | 4.5 | |
98.41 (4) | 98.69 (3) | 98.54 (3) | 98.36 (3) | 3.3 | ||
Avg_pool | 95.80 (8) | 94.95 (9) | 95.35 (8) | 95.11 (8) | 8.3 | |
98.26 (5) | 98.10 (5) | 98.18 (5) | 97.75 (5) | 5.0 | ||
Avg_pool | 92.00 (11) | 92.11 (11) | 92.05 (10) | 91.73 (11) | 10.8 | |
97.24 (9) | 97.34 (9) | 97.29 (9) | 96.98 (10) | 9.3 | ||
Avg_pool | 97.48 (2) | 97.66 (2) | 97.56 (2) | 97.30 (3) | 2.3 | |
98.85 (2) | 98.86 (2) | 98.85 (2) | 98.58 (2) | 2.0 | ||
Pool10 | 95.14 (9) | 95.13 (8) | 95.06 (9) | 94.76 (9) | 8.8 | |
95.39 (11) | 94.75 (11) | 95.03 (11) | 95.08 (11) | 11.0 | ||
Node_200 | 96.34 (6) | 95.68 (7) | 96.00 (7) | 95.82 (7) | 6.8 | |
97.20 (10) | 97.56 (8) | 97.36 (8) | 97.07 (8) | 8.5 | ||
fc7 | 97.24 (3) | 97.57 (3) | 97.40 (3) | 97.33 (2) | 2.8 | |
97.67 (6) | 97.84 (7) | 97.75 (7) | 97.39 (7) | 6.8 | ||
fc7 | 98.21(1) | 98.43 (1) | 98.32 (1) | 98.10 (1) | 1.0 | |
98.87 (1) | 99.08 (1) | 98.97 (1) | 98.81 (1) | 1.0 | ||
Avg. on CNN Avg on CNN+ | 95.93 | 95.73 | 95.81 | 95.58 | ||
97.76 | 97.80 | 97.77 | 97.51 |
*Bold numbers indicate the best performance.
Table 3.
Method | Performance Metrics (%) |
|||||
---|---|---|---|---|---|---|
Feature ExtractorLayer | Sensitivity (Recall) | Precision (PPV) | F1-score | Accuracy | Average Rank | |
87.47 (5) | 84.52 (7) | 85.86 (5) | 88.02 (4) | 5.3 | ||
91.97 (8) | 90.66 (7) | 91.22 (7) | 92.72 (7) | 7.3 | ||
Avg_pool | 88.11 (4) | 84.90 (6) | 86.35 (4) | 87.87 (5) | 4.8 | |
94.25 (6) | 91.26 (5) | 92.64 (5) | 93.58 (5) | 5.3 | ||
Avg_pool | 73.10 (11) | 76.76 (9) | 74.56 (11) | 78.74 (10) | 10.3 | |
91.45 (10) | 88.99 (9) | 90.12 (8) | 91.58 (9) | 9.0 | ||
pool5 | 83.76 (8) | 85.38 (5) | 84.38 (8) | 86.59 (8) | 7.3 | |
95.68 (3) | 94.35 (3) | 94.95 (3) | 95.72 (3) | 3.0 | ||
Avg_pool | 80.70 (9) | 76.38 (10) | 77.95 (9) | 80.46 (9) | 9.3 | |
92.60 (7) | 90.90 (6) | 91.70 (6) | 93.15 (6) | 6.3 | ||
Avg_pool | 75.04 (10) | 74.49 (11) | 74.75 (10) | 78.60 (11) | 10.5 | |
95.46 (4) | 89.00 (8) | 90.10 (9) | 91.73 (8) | 7.3 | ||
Avg_pool | 91.83 (3) | 91.60 (2) | 91.57 (2) | 92.87 (2) | 2.3 | |
96.59 (2) | 96.20 (2) | 96.39 (2) | 96.72 (2) | 2.0 | ||
Pool10 | 85.64 (6) | 84.51 (8) | 85.02 (7) | 87.30 (7) | 7.0 | |
85.39 (11) | 86.12 (11) | 85.45 (11) | 87.87 (11) | 11.0 | ||
Node_200 | 84.68 (7) | 85.95 (4) | 85.27 (6) | 87.73 (6) | 5.8 | |
91.80 (9) | 88.71 (10) | 90.10 (9) | 91.16 (10) | 9.5 | ||
fc7 | 92.08 (2) | 91.07 (3) | 91.39 (3) | 92.72 (3) | 2.8 | |
94.87 (5) | 91.99 (4) | 93.33 (4) | 94.15 (4) | 4.3 | ||
fc7 | 92.32 (1) | 92.58 (1) | 92.45 (1) | 93.30 (1) | 1.0 | |
96.64 (1) | 96.29 (1) | 96.46 (1) | 96.86 (1) | 1.0 | ||
Avg. on CNN Avg. on CNN+ | 85.88 | 84.38 | 84.50 | 86.75 | ||
93.34 | 91.32 | 92.04 | 93.20 |
*Bold numbers indicate the best performance.
Table 4.
Method | Performance Metrics (%) |
||||
---|---|---|---|---|---|
Sensitivity (Recall) | Precision (PPV) | F1-score | Accuracy | Average Rank | |
NaiveBayes | 97.94 (12) | 97.63 (12) | 97.78 (12) | 97.75 (12) | 12.0 |
NN ( = 3) | 98.61 (6) | 98.72 (6) | 98.67 (6) | 98.55 (6) | 6.0 |
NN (= 5) | 98.53 (7) | 98.75 (5) | 98.64 (7) | 98.52 (8) | 6.8 |
NN ( = 7) | 98.44 (8) | 98.60 (8) | 98.52 (8) | 98.39 (9) | 8.3 |
OvO SVM | 98.42 (10) | 98.52 (10) | 98.47 (10) | 98.30 (10) | 10.0 |
OvA SVM | 98.82 (3) | 98.61 (7) | 98.71 (5) | 98.68 (4) | 4.8 |
Decision Tree | 98.45 (9) | 98.56 (9) | 98.50 (9) | 98.55 (6) | 8.3 |
AdaBoostM2 | 98.63 (5) | 98.96 (4) | 98.80 (4) | 98.65 (5) | 4.5 |
TotalBoost | 98.68 (4) | 99.06 (3) | 98.87 (3) | 98.78 (3) | 3.3 |
Random Forrest | 98.84 (2) | 99.00 (2) | 98.92 (2) | 98.84 (1) | 1.8 |
SoftMax | 98.21 (11) | 98.43 (11) | 98.32 (11) | 98.10 (11) | 11.0 |
Proposed LMPL | 98.87 (1) | 99.08 (1) | 98.97 (1) | 98.81 (2) | 1.3 |
*Bold numbers indicate the best performance.
Table 5.
Method | Performance Metrics (%) |
||||
---|---|---|---|---|---|
Sensitivity (Recall) | Precision (PPV) | F1-score | Accuracy | Average Rank | |
NaiveBayes | 89.20 (12) | 94.23 (10) | 91.25 (12) | 92.58 (12) | 11.5 |
NN ( = 3) | 96.08 (2) | 95.43 (4) | 95.72 (3) | 96.29 (3) | 3.0 |
NN ( = 5) | 95.31 (5) | 94.95 (6) | 95.08 (6) | 95.72 (5) | 5.3 |
NN ( = 7) | 94.35 (8) | 94.57 (9) | 94.42 (8) | 95.15 (8) | 8.3 |
OvO SVM | 94.64 (7) | 93.10 (11) | 93.84 (9) | 95.01 (9) | 9.0 |
OvA SVM | 95.86 (4) | 94.73 (7) | 95.28 (4) | 96.01 (4) | 4.8 |
Decision Tree | 94.95 (6) | 95.20 (5) | 95.06 (7) | 95.72 (5) | 5.8 |
AdaBoostM2 | 93.07 (10) | 94.66 (8) | 93.82 (10) | 94.58 (10) | 9.5 |
TotalBoost | 93.63 (9) | 96.83 (1) | 95.13 (5) | 95.72 (5) | 5.0 |
Random Forrest | 96.03 (3) | 95.86 (3) | 95.93 (2) | 96.58 (2) | 2.5 |
SoftMax | 92.32 (11) | 92.58 (12) | 92.45 (11) | 93.30 (11) | 11.3 |
Proposed LMPL | 96.64 (1) | 96.29 (2) | 96.46 (1) | 96.86 (1) | 1.3 |
*Bold numbers indicate the best performance.
Table 6.
Method | Performance Metrics (%) |
|||||
---|---|---|---|---|---|---|
Sensitivity (Recall) | Precision (PPV) | F1-score | Accuracy | Average Rank | ||
Without Fusion | preRelu | 98.64 (8) | 98.81 (8) | 98.72 (8) | 98.52 (8) | 8.0 |
postRelu | 98.87 (6) | 99.08 (6) | 98.97 (7) | 98.81 (7) | 6.5 | |
Concat | Classical | 99.22 (4) | 99.33 (4) | 99.28 (4) | 99.13 (4) | 4.0 |
DCA | 98.86 (7) | 99.08 (6) | 98.99 (6) | 98.87 (6) | 6.3 | |
CCA | 99.37 (2) | 99.48 (1) | 99.42 (2) | 99.32 (2) | 1.8 | |
Sum | Classical | 99.27 (3) | 99.38 (3) | 99.33 (3) | 99.20 (3) | 3.0 |
DCA | 98.91 (5) | 99.18 (5) | 99.04 (5) | 99.00 (5) | 5.0 | |
CCA | 99.42 (1) | 99.47 (2) | 99.45 (1) | 99.39 (1) | 1.3 |
*Bold numbers indicate the best performance.
Table 7.
Method | Performance Metrics (%) |
|||||
---|---|---|---|---|---|---|
Sensitivity (Recall) | Precision (PPV) | F1-score | Accuracy | Average Rank | ||
Without Fusion | preRelu | 95.74 (8) | 96.03 (8) | 95.88 (8) | 96.29 (8) | 8.0 |
postRelu | 96.64 (5) | 96.29 (7) | 96.46 (7) | 96.86 (7) | 6.5 | |
Concat | Classical | 97.94 (4) | 96.97 (4) | 97.44 (4) | 97.72 (4) | 4.0 |
DCA | 96.44 (7) | 96.88 (5) | 96.66 (5) | 97.00 (5) | 5.5 | |
CCA | 98.66 (2) | 97.87 (2) | 98.26 (2) | 98.43 (2) | 2.0 | |
Sum | Classical | 98.10 (3) | 97.39 (3) | 97.74 (3) | 98.00 (3) | 3.0 |
DCA | 96.48 (6) | 96.58 (6) | 96.52 (6) | 97.00 (5) | 5.8 | |
CCA | 98.89 (1) | 98.56 (1) | 98.73 (1) | 98.86 (1) | 1.0 |
*Bold numbers indicate the best performance.
The classical method is to group two sets of feature vectors which simply concatenates the two feature sets into a single vector. Discriminant Correlation Analysis (DCA) [38] maximizes the variance between classes, as well as the correlation among features through several feature sets.
We observed that regarding the performance metrics, the CCA fusion method with summation achieved outstanding results with 99.39% and 98.86% average accuracies in comparison with the other approaches.
Table 8, Table 9 compare the final performance of the proposed method with the other preprocessing techniques for both experiments. The effectiveness of different augmentation techniques in the final performance of the proposed method is compared in Table 10, Table 11 for both experiments using the following parameters.
-
•
Brightness: Brightness is the amount of hue. As brightness varies from 0 to 1, colors change from black to white. Brightness jitter shifts the darkness and lightness of an input image. The brightness of the input image is adjusted by an offset selected randomly from the range . As expected, the image appears brighter with the rise in brightness.
-
•
Contrast: The contrast jitter randomly adjusts the difference between the darkest and brightest regions in an input image. The contrast of the input image is adjusted by a scale factor selected randomly from the range [1.2, 1.4]. As the contrast increases, shadows become darker, and highlights become brighter.
-
•
Cropping: The image is cropped to the target size 800 × 800 from the center of the image.
-
•
Flipping: A reflection transformation is created that flips the input image in the left-right direction ( dimension) and top-bottom direction ( dimension).
-
•
Gaussian noise: A zero-mean, Gaussian white noise with a variance of 0.01 is added to the image.
-
•
Salt & pepper noise: Salt and pepper noise is added with a noise density of 0.1. This affects almost 10% of the pixels.
-
2)
Statistical analysis
Table 8.
Method | Performance Metrics (%) |
||||
---|---|---|---|---|---|
Sensitivity (Recall) | Precision (PPV) | F1-score | Accuracy | Average Rank | |
No Preprocessing | 96.09 (5) | 95.89 (5) | 95.99 (5) | 96.40 (5) | 5.0 |
Image Adjustment | 97.50 (2) | 97.57 (2) | 97.52 (2) | 97.52 (2) | 2.0 |
Histogram Matching | 96.93 (4) | 96.64 (4) | 96.78 (4) | 97.01 (4) | 4.0 |
Histogram Equalization | 97.42 (3) | 97.39 (3) | 97.39 (3) | 97.43 (3) | 3.0 |
CLAHE | 99.42 (1) | 99.47 (1) | 99.45 (1) | 99.39 (1) | 1.0 |
*Bold numbers indicate the best performance.
Table 9.
Method | Performance Metrics (%) |
||||
---|---|---|---|---|---|
Sensitivity (Recall) | Precision (PPV) | F1-score | Accuracy | Average Rank | |
No Preprocessing | 89.77 (5) | 91.60 (4) | 90.62 (5) | 92.15 (5) | 4.75 |
Image Adjustment | 97.57 (3) | 96.63 (2) | 97.09 (3) | 97.57 (2) | 2.5 |
Histogram Matching | 94.78 (4) | 91.19 (5) | 92.80 (4) | 93.72 (4) | 4.25 |
Histogram Equalization | 97.94 (2) | 96.47 (3) | 97.18 (2) | 97.43 (3) | 2.5 |
CLAHE | 98.89 (1) | 98.56 (1) | 98.73 (1) | 98.86 (1) | 1.0 |
*Bold numbers indicate the best performance.
Table 10.
Method | Performance Metrics (%) |
||||
---|---|---|---|---|---|
Sensitivity (Recall) | Precision (PPV) | F1-score | Accuracy | Average Rank | |
No Augmentation | 96.12 (5) | 96.88 (5) | 96.48 (5) | 96.37 (5) | 5.0 |
Brightness & Contrast | 98.48 (3) | 98.22 (4) | 98.35 (3) | 98.10 (4) | 3.5 |
Cropping & Fliping | 98.50 (2) | 98.40 (2) | 98.45 (2) | 98.20 (2) | 2.0 |
Gaussian & Salt and pepper noise | 98.34 (4) | 98.24 (3) | 98.29 (4) | 98.20 (2) | 3.25 |
Rotation and Translation | 99.42 (1) | 99.47 (1) | 99.45 (1) | 99.39 (1) | 1.0 |
*Bold numbers indicate the best performance.
Table 11.
Method | Performance Metrics (%) |
||||
---|---|---|---|---|---|
Sensitivity (Recall) | Precision (PPV) | F1-score | Accuracy | Average Rank | |
No Augmentation | 95.12 (5) | 96.44 (5) | 95.74 (5) | 95.56 (5) | 5.0 |
Brightness & Contrast | 97.84 (3) | 97.98 (3) | 97.89 (2) | 97.62 (3) | 2.75 |
Cropping & Fliping | 97.52 (4) | 98.05 (2) | 97.76 (3) | 97.56 (4) | 3.25 |
Gaussian & Salt and pepper noise | 97.89 (2) | 97.66 (4) | 97.76 (3) | 97.68 (2) | 2.75 |
Rotation and Translation | 98.89 (1) | 98.56 (1) | 98.73 (1) | 98.86 (1) | 1.0 |
*Bold numbers indicate the best performance.
The Friedman test was used to further analyze the experimental performance of the comparable methods statistically. It is a simple, nonparametric test and safe for comparing at least three related samples. The test has no assumptions about the primary distribution of data. It ranks the algorithms for each metric independently such that the algorithm with the highest performance is ranked first, the second-best is ranked second, and this procedure continues for the next ranks. represents the average rank of the jth method depending on different metrics listed in Table 2, Table 3, Table 4, Table 5, Table 6, Table 7, Table 8, Table 9, Table 10, Table 11
In other words, is calculated as follows
(10) |
and denotes the rank of the jth method on the ith metric. In the case of ties, i.e., when two algorithms perform similarly, the same ranks are assigned.
It can be seen from Table 2, Table 3, Table 4, Table 5, Table 6, Table 7, Table 8, Table 9, Table 10, Table 11 that the proposed method improved the performance significantly and achieved the best average rank in all cases. These tables show a significant difference in the effectiveness of the algorithms. Based on the Friedman test, the performance of the proposed MLPL classifier could be significantly better. In order to determine the best fusion method, the averaged ranks of different fusion techniques were calculated, as shown in Table 6, Table 7 As presented in the tables, CCA fusion techniques with summation improved the performance considerably and achieved the best rank in both experiments.
Moreover, a comparison between preprocessing methods is performed in Table 8, Table 9 As can be seen, CLAHE obtained the first rank. The proposed augmentation techniques (rotation and translation) also showed the best effectiveness and achieved the best rank in comparison to other augmentation techniques (see Table 10, Table 11).
-
3)
Visual analysis
We provided the occlusion sensitivity [39] that provides a detailed visual analysis of the extracted deep features through the fine-tuned VGG-Net, as shown in Fig. 9 . Occlusion sensitivity is a technique to generate visual descriptions of the CNN model predictions. It provides an insight into the internal workings of a classifier based on deep CNNs.
Misclassifications or unpredicted results can be based on reasonable clarifications. Using heat maps, we investigated the prediction regions. False-negative examples were examined to find out the reasons underlying the hidden classification problem. The spatial parts that contributed most to false-negative predictions in COVID-19 classification were identified by occlusion sensitivity heat mapping, as shown in Fig. 9. The standard jet color-map is used where red and yellow denote regions with high contribution to the false-negative predictions, and blue indicates regions with a low contribution. Manual inspection showed that in these false-negative examples, the network incorrectly attended on some edges and corners of images that are not relevant to COVID-19.
The latest deep models using x-ray images (on not necessarily common datasets) are summarized in Table 12 .
Table 12.
Study | No. of cases | Method | Accuracy (%) |
---|---|---|---|
Ozturk et al. [5] | 125 COVID-19 | DarkCovidNet | |
500 No-finding | 98.08 | ||
125 COVID-19 | |||
500 Pneumonia | |||
500 No-finding | 87.02 | ||
Tabik et al. [6] | 426 COVID-19 | COVID-SD Net | 97.72 |
426 Normal | |||
Rahman et al. [8] | 3619 COVID-19 | ChexNet | 96.29 |
8851 Normal | |||
6012 Pneumonia | |||
Toğaçar et al. [10] | 295 COVID-19 | MobileNetV, SVM | 96.28 |
65 Normal | |||
98 Pneumonia | |||
Ioannis et al. [14] | 224 COVID-19 | VGG-19 | 93.48 |
700 Pneumonia | |||
504 Healthy | |||
Karakanis et al. [12] | 275 COVID-19 | ResNet, CGAN | |
275 Normal | 98.7 | ||
275 COVID-19 | |||
275 Normal | |||
275 Bacterial PNA | 98.3 | ||
Wang and Wong [13] | 53 COVID-19 | COVID-Net | 92.4 |
8066 Healthy | |||
Sethy and Behra [17] | 25 COVID-19 | ResNet 50, SVM | 95.38 |
25 No-finding | |||
Jain et al. [9] | COVID-19 | Xception | 97.97 |
Normal | |||
Pneumonia | |||
Jin et al. [15] | 543 COVID-19 | AlexNet | 98.64 |
600 Normal | |||
600 Viral PNA | |||
Hemdan et al. [11] | 25 COVID-19 | COVIDX-Net | 90.00 |
25 No-finding | |||
Narin et al. [16] | 50 COVID-19 | Res-Net 50 | 98.00 |
50 No-finding | |||
Mahmud et al. [7] | 305 COVID-19 | CovXNet | |
305 Normal | 97.4 | ||
305 COVID-19 | |||
305 Normal | |||
305 Viral PNA | |||
305 Bacterial PNA | 90.3 | ||
Minaee et al. [18] | 184 COVID-19 | SqueezeNet | 92.30 |
5000 Non-COVID | |||
6054 Pneumonia | |||
Abbaas et al. [28] | 80 Covid-19 | DeTraC | 93.10 |
105 Normal | |||
11 SARS | |||
Tahir et al. [27] | 423 COVID-19 | InceptionV3 | 97.73 |
134 SARS | |||
144 MERS | |||
Proposed method (Transferred deep features of VGG-Net, deep feature fusion, the LMPL classifier) | 423 COVID-19 | VGG-Net19 | |
1341 Normal | |||
1345 Viral PNA | 99.39 | ||
423 COVID-19 | |||
134 SARS | |||
144 MERS | 98.86 |
5. Discussion
LMPL improved the performance of all CNNs significantly and achieved the best results with a few iterations (five ones in our experiments) to avoid overfitting with small sample sizes and achieve better generalization. The LMPL outperformed conventional classifiers and provided very competitive results compared to ensemble methods with 100 trees as weak learners. However, the complexity of ensembles was high and needed space to train multiple models. The LMPL is simple and needs smaller space to produce a good model.
We can conclude that the fusion of deep features is helpful and improves the results even with classical concatenation and summation, which are fast and easy to apply. The obtained deep feature vectors were fused to acquire more informative features that minimized the impact of insufficient features obtained from one CNN model. Feature fusion efficiently produces features that comprise rich information describing the image well. Appropriately combining two or more features is not a trivial task. Due to similar visual features in various viral pneumonia, especially severe coronavirus of the SARS and MERS family, the DCA feature fusion strategy could not find useful information to maximize distances between classes.
However, x-rays are widely available and the most common approach as they are low-cost. CT scans are a more powerful method, so a chest CT scan might be needed in some cases to get a better picture and a more detailed view. Indeed, x-rays can help as a first-line diagnostic tool in most cases if they are detectable by this technology. In this study, we studied how much x-ray images could help to identify COVID-19, and we plan to extend our model for CT scans as well.
6. Conclusions and future work
Early COVID-19 detection can benefit preparing an appropriate treatment plan and facilitate medical decision-making. In this study, an LMPL classifier was presented for the diagnosis of coronavirus among a wide range of other viral pneumonia using raw chest x-ray images. The introduced method was shown to solve two classification problems: normal, COVID-19, and typical viral pneumonia; COVID-19, SARS, and MERS pneumonia. The results showed the outstanding average accuracy of the proposed method, equal to 99.39% and 98.86%, respectively, for both schemes, compared to state-of-the-art deep models.
In the future, the objective function of the introduced learning model can be extended to obtain clustering and semi-supervised models. To improve the LMPL classifier, positive don't care samples can be considered in tuning hyperplanes to get larger margins and better generalization of the model. However, such loss functions are not convex. Therefore, the solutions to deal with non-convex objective functions like gradient descent should be investigated. Other binary classifiers could be developed by the proposed cellular model. Applying kernels in the proposed classifier may help to consider complicated nonlinear decision boundaries. Extracting more effective features to make a more accurate and robust model for treating other emerging diseases with insufficient data can be considered in the future. The proposed pipeline can be adapted for chest CT scans to diagnose COVID-19 pneumonia, as well as other image processing applications of industrial and healthcare systems such as industrial cameras, process control, industrial robotics, and object recognition.
Declaration of competing interest
None Declared.
References
- 1.Mahase E. 2020. Coronavirus: Covid-19 Has Killed More People than SARS and MERS Combined, Despite Lower Case Fatality Rate. [DOI] [PubMed] [Google Scholar]
- 2.Ibrahim D.M., Elshennawy N.M., Sarhan A.M. Deep-chest: multi-classification deep learning model for diagnosing COVID-19, pneumonia, and lung cancer chest diseases. Comput. Biol. Med. 2021;132:104348. doi: 10.1016/j.compbiomed.2021.104348. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 3.Simonyan K., Zisserman A. Very deep convolutional networks for large-scale image recognition. ArXiv Prepr. ArXiv1409. 2014;1556 [Google Scholar]
- 4.Haghighat M., Abdel-Mottaleb M., Alhalabi W. Fully automatic face normalization and single sample face recognition in unconstrained environments. Expert Syst. Appl. 2016;47:23–34. [Google Scholar]
- 5.Ozturk T., Talo M., Yildirim E.A., Baloglu U.B., Yildirim O., Acharya U.R. Automated detection of COVID-19 cases using deep neural networks with X-ray images. Comput. Biol. Med. 2020 doi: 10.1016/j.compbiomed.2020.103792. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 6.Tabik S., Gómez-Ríos A., Martín-Rodríguez J.L., Sevillano-García I., Rey-Area M., Charte D., Guirado E., Suárez J.L., Luengo J., Valero-González M.A. COVIDGR dataset and COVID-SDNet methodology for predicting COVID-19 based on Chest X-Ray images. IEEE J. Biomed. Heal. Inform. 2020;24:3595–3605. doi: 10.1109/JBHI.2020.3037127. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 7.Mahmud T., Rahman M.A., Fattah S.A. CovXNet: a multi-dilation convolutional neural network for automatic COVID-19 and other pneumonia detection from chest X-ray images with transferable multi-receptive feature optimization. Comput. Biol. Med. 2020;122 doi: 10.1016/j.compbiomed.2020.103869. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8.Rahman T., Khandakar A., Qiblawey Y., Tahir A., Kiranyaz S., Kashem S.B.A., Islam M.T., Al Maadeed S., Zughaier S.M., Khan M.S. Exploring the effect of image enhancement techniques on COVID-19 detection using chest X-ray images. Comput. Biol. Med. 2021;132 doi: 10.1016/j.compbiomed.2021.104319. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 9.Jain R., Gupta M., Taneja S., Hemanth D.J. Deep learning based detection and analysis of COVID-19 on chest X-ray images. Appl. Intell. 2021;51:1690–1700. doi: 10.1007/s10489-020-01902-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 10.Toğaçar M., Ergen B., Cömert Z. COVID-19 detection using deep learning models to exploit Social Mimic Optimization and structured chest X-ray images using fuzzy color and stacking approaches. Comput. Biol. Med. 2020;121 doi: 10.1016/j.compbiomed.2020.103805. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 11.Hemdan E.E.-D., Shouman M.A., Karar M.E. vol. 11055. 2020. Covidx-net: A Framework of Deep Learning Classifiers to Diagnose Covid-19 in X-Ray Images. ArXiv Prepr. ArXiv2003. [Google Scholar]
- 12.Karakanis S., Leontidis G. Lightweight deep learning models for detecting COVID-19 from chest X-ray images. Comput. Biol. Med. 2021;130 doi: 10.1016/j.compbiomed.2020.104181. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 13.Wang L., Wong A. 2020. COVID-Net: A Tailored Deep Convolutional Neural Network Design for Detection of COVID-19 Cases from Chest X-Ray Images. ArXiv Prepr. ArXiv2003.09871. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14.Apostolopoulos I.D., Mpesiana T.A. Covid-19: automatic detection from x-ray images utilizing transfer learning with convolutional neural networks. Phys. Eng. Sci. Med. 2020:1. doi: 10.1007/s13246-020-00865-4. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15.Jin W., Dong S., Dong C., Ye X. Hybrid ensemble model for differential diagnosis between COVID-19 and common viral pneumonia by chest X-ray radiograph. Comput. Biol. Med. 2021;131 doi: 10.1016/j.compbiomed.2021.104252. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 16.Narin A., Kaya C., Pamuk Z. 2020. Automatic Detection of Coronavirus Disease (Covid-19) Using X-Ray Images and Deep Convolutional Neural Networks. ArXiv Prepr. ArXiv2003.10849. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 17.Sethy P.K., Behera S.K. 2020. Detection of coronavirus disease (covid-19) based on deep features; p. 2020. Preprints. 2020030300. [Google Scholar]
- 18.Minaee S., Kafieh R., Sonka M., Yazdani S., Soufi G.J. Deep-covid: predicting covid-19 from chest x-ray images using deep transfer learning. Med. Image Anal. 2020;65 doi: 10.1016/j.media.2020.101794. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 19.Ardakani A.A., Kanafi A.R., Acharya U.R., Khadem N., Mohammadi A. Application of deep learning technique to manage COVID-19 in routine clinical practice using CT images: results of 10 convolutional neural networks. Comput. Biol. Med. 2020;121 doi: 10.1016/j.compbiomed.2020.103795. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 20.Amyar A., Modzelewski R., Li H., Ruan S. Multi-task deep learning based CT imaging analysis for COVID-19 pneumonia: classification and segmentation. Comput. Biol. Med. 2020;126 doi: 10.1016/j.compbiomed.2020.104037. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 21.Qian X., Fu H., Shi W., Chen T., Fu Y., Shan F., Xue X. M $^ 3$ lung-sys: a deep learning system for multi-class lung pneumonia screening from CT imaging. IEEE J. Biomed. Heal. Inform. 2020;24:3539–3550. doi: 10.1109/JBHI.2020.3030853. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 22.Serte S., Demirel H. Deep learning for diagnosis of COVID-19 using 3D CT scans. Comput. Biol. Med. 2021;132 doi: 10.1016/j.compbiomed.2021.104306. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 23.Shiri I., Sorouri M., Geramifar P., Nazari M., Abdollahi M., Salimi Y., Khosravi B., Askari D., Aghaghazvini L., Hajianfar G. Machine learning-based prognostic modeling using clinical data and quantitative radiomic features from chest CT images in COVID-19 patients. Comput. Biol. Med. 2021;132:104304. doi: 10.1016/j.compbiomed.2021.104304. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 24.Dastider A.G., Sadik F., Fattah S.A. An integrated autoencoder-based hybrid CNN-LSTM model for COVID-19 severity prediction from lung ultrasound. Comput. Biol. Med. 2021;132:104296. doi: 10.1016/j.compbiomed.2021.104296. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 25.Hamimi A. MERS-CoV: Middle East respiratory syndrome corona virus: can radiology be of help? Initial single center experience, Egypt. J. Radiol. Nucl. Med. 2016;47:95–106. [Google Scholar]
- 26.Xie X., Li X., Wan S., Gong Y. Data Min. Springer; 2006. Mining x-ray images of SARS patients; pp. 282–294. [Google Scholar]
- 27.Tahir A., Qiblawey Y., Khandakar A., Rahman T., Khurshid U., Musharavati F., Kiranyaz S., Chowdhury M.E.H. vol. 11524. 2020. Coronavirus: Comparing COVID-19, SARS and MERS in the Eyes of AI. ArXiv Prepr. ArXiv2005. [Google Scholar]
- 28.Abbas A., Abdelsamea M.M., Gaber M.M. Classification of COVID-19 in chest X-ray images using DeTraC deep convolutional neural network. Appl. Intell. 2021;51:854–864. doi: 10.1007/s10489-020-01829-7. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 29.X. Xuanyang, G. Yuchang, W. Shouhong, L. Xi, Computer aided detection of SARS based on radiographs data mining, in: 2005 IEEE Eng. Med. Biol. 27th Annu. Conf., IEEE, 2006: pp. 7459–7462. [DOI] [PubMed]
- 30.Pizer S.M., Amburn E.P., Austin J.D., Cromartie R., Geselowitz A., Greer T., ter Haar Romeny B., Zimmerman J.B., Zuiderveld K. Adaptive histogram equalization and its variations, Comput. Vision, Graph. Image Process. 1987;39:355–368. [Google Scholar]
- 31.Italian Society of Medical and Interventional Radiology | SIRM, (n.d.). https://www.sirm.org/en/italian-society-of-medical-and-interventional-radiology/(accessed June 30, 2020).
- 32.Cohen J.P., Morrison P., Dao L., Roth K., Duong T.Q., Ghassemi M. 2020. COVID-19 Image Data Collection: Prospective Predictions Are the Future.https://github.com/ieee8023/covid-chestxray-dataset (accessed June 30, 2020) [Google Scholar]
- 33.COVID-19 @ Radiopaedia.Org, (n.d.). https://radiopaedia.org/playlists/25975?lang=us (accessed June 30, 2020).
- 34.Thread by @ChestImaging: This Is a Thread of COVID-19 CXR (All SARS-CoV-2 PCR+) from My Hospital (Spain)., (n.d.). https://threadreaderapp.com/thread/1243928581983670272.html (accessed June 30, 2020).
- 35.COVID-19 DATABASE | SIRM, (n.d.). http://www.sirm.org/category/senza-categoria/covid-19/(accessed June 30, 2020).
- 36.Covid-19, Sars, MERS X-Ray Images Dataset | Kaggle, (n.d.). https://www.kaggle.com/dataset/057e1b6dc41d9691e59dded4445fa8cc2f0b4b5cbcb49aef9583d95233799d5a/version/2 (accessed June 27, 2020).
- 37.Chest X-Ray Images (Pneumonia) | Kaggle, (n.d.). https://www.kaggle.com/paultimothymooney/chest-xray-pneumonia? (accessed June 27, 2020).
- 38.Bi Y., Lv M., Wei Y., Guan N., Yi W. Multi-feature fusion for thermal face recognition. Infrared Phys. Technol. 2016;77:366–374. [Google Scholar]
- 39.Zeiler M.D., Fergus R. Eur. Conf. Comput. Vis. Springer; 2014. Visualizing and understanding convolutional networks; pp. 818–833. [Google Scholar]