Skip to main content
Springer Nature - PMC COVID-19 Collection logoLink to Springer Nature - PMC COVID-19 Collection
. 2021 Jan 22;24(3):951–964. doi: 10.1007/s10044-020-00950-0

A novel framework for rapid diagnosis of COVID-19 on computed tomography scans

Tallha Akram 1, Muhammad Attique 2, Salma Gul 3, Aamir Shahzad 4, Muhammad Altaf 1, S Syed Rameez Naqvi 1, Robertas Damaševičius 5,, Rytis Maskeliūnas 6
PMCID: PMC7819829  PMID: 33500681

Abstract

Since the emergence of COVID-19, thousands of people undergo chest X-ray and computed tomography scan for its screening on everyday basis. This has increased the workload on radiologists, and a number of cases are in backlog. This is not only the case for COVID-19, but for the other abnormalities needing radiological diagnosis as well. In this work, we present an automated technique for rapid diagnosis of COVID-19 on computed tomography images. The proposed technique consists of four primary steps: (1) data collection and normalization, (2) extraction of the relevant features, (3) selection of the most optimal features and (4) feature classification. In the data collection step, we collect data for several patients from a public domain website, and perform preprocessing, which includes image resizing. In the successive step, we apply discrete wavelet transform and extended segmentation-based fractal texture analysis methods for extracting the relevant features. This is followed by application of an entropy controlled genetic algorithm for selection of the best features from each feature type, which are combined using a serial approach. In the final phase, the best features are subjected to various classifiers for the diagnosis. The proposed framework, when augmented with the Naive Bayes classifier, yields the best accuracy of 92.6%. The simulation results are supported by a detailed statistical analysis as a proof of concept.

Keywords: Covid19, Features extraction, Features selection, Features classification

Introduction

Coronavirus Disease 2019 (COVID-19) is highly contagious and has rapidly spread globally infecting almost all the countries with millions of positive cases and more than 0.4 million deaths [1], and is continuously on the rise; see Table 1 and Figs. 1 and 2. The key factor to limit this pandemic situation is the early testing and diagnosis. However, due to its pandemic nature, quick collection and testing of samples from the suspected patients is a challenging issue for clinical management. Its early detection is possible with Nucleic Acid Amplification Tests (NAAT), such as reverse transcription polymerase chain reaction (RT-PCR) [2], which is required to be interpreted by trained clinical laboratory personnel [3]. The initial symptoms of COVID-19 are fever, fatigue and dry cough, and it predominantly affects lungs. The affected lobes with ground glass changes and/or consolidations etc. can be recorded in chest radiology images [4, 5]. This is why the clinicians worldwide are using chest X-rays (CXR) and computed tomography (CT) images as an alternative and fast method for the screening and diagnosis of the COVID-19, especially supported by the fact that the RT-PCR method may take several hours to complete the process [6].

Table 1.

Covid-19 statistics: Retrieved on May 19, 2020 from www.worldometer.info

Country Total cases Total deaths Total recovered Active cases Critical cases
USA 1,612,450 95,923 374,177 1,142,350 17,964
Russia 317,554 3099 92,681 221,774 2300
Spain 280,177 27,940 196,958 55,219 1152
Italy 228,006 32,486 134,560 60,960 640
France 181,826 28,215 63,858 89,753 1745
Germany 178,918 8282 158,000 12,636 1016
China 82,967 4634 78,249 84 8
Iran 129,341 7249 100,564 21,528 2655
India 118,226 3584 48,553 66,089
Pakistan 48,091 1017 14,155 32,919 111

Fig. 1.

Fig. 1

Confirmed COVID-19 cases in 10 selected countries: Retrieved on May 19, 2020 from www.worldometer.info

Fig. 2.

Fig. 2

Confirmed deaths in 10 selected countries: Retrieved on May 19, 2020 from www.worldometer.info

Now due to the rise of this pandemic situation, thousands of people daily undergo CXR and CT scan for screening of COVID-19. This has overburdened the radiologist leading to their decreased productivity [7] in the proper detection of suspicious abnormalities [8]. In the case of contagious diseases, this backlog of radiological studies cannot be afforded; however, in the case of chronic and slow diseases such studies may be delayed. For detailed analysis of the errors and discrepancies in radiology diagnosis, the interested readers are referred to [917]. Artificial intelligence (AI) and computer vision system play an important role in classifying different complex structures found in the medical images [1821] and can be used in the computer aided diagnosis tools. Therefore, AI is being researched for radiological diagnosis since long and has proven quite successful in the cases of breast screening [22], diagnosis and quantification of emphysema severity [23], tuberculosis detection [24] and claims of detection of diagnosis idiopathic pulmonary fibrosis with similar accuracy to a human reader [25].

During the past year, several imaging-based diagnosis techniques of COVID-19 backed by AI and machine learning have been presented, along with their correlation with the RT-PCR [26, 27]. CT and CXR images are processed for the detection of pneumonia like imaging features using AI techniques. In [28], CXR images and deep convolutional neural networks (CNNs) are used to diagnose COVID-19. The models used are ResNet 50, Inception V3 and a hybrid approach based on Inception-ResNetV2, claiming accuracies of 98%,97% and 87%, respectively. In another approach [29], COVID-19 and bacterial and viral pneumonia are diagnosed and classified into negative or positive by using X-ray radiographs. The approach makes use of GoogleNet as a deep transfer model, claiming 100% accuracy. In another approach [30], the authors have claimed 86.7% accuracy in diagnosing the early signs of COVID-19, using chest CT images with deep learning . Authors in [31] propose to use two-dimensional (2D) and three-dimensional (3D) deep learning models combination and claim an accuracy of 98.2% and specificity of 92.2% with chest CT images. In [32], the authors combined the Inception CNN with Marine Predators algorithm to select the most relevant features from COVID-19 X-ray images, achieving an accuracy of 98.7%. In another study, the deep learning methods are used to extract COVID-19 graphical features and provide clinical diagnosis quite ahead of the pathogenic test helping in timely control of the spread. The authors claim a rather constrained accuracy of 85.2% [10].

On the one hand, it is widely accepted that the diagnoses based on CXR are not as efficient as those based on the CT scans [11]; on the other hand, however, the accuracies reported in the literature for the former surprisingly exceed those for the CT scans. Furthermore, these systems normally give binary decision of either negative or positive, and do not incorporate any qualitative analyses based on the recommendations of Radiological Society of North America (RSNA) [33, 34]. Hence, it may be concluded that most of the available methods are either not sufficiently reliable, or achieve a constrained diagnosis efficiency. Especially for contagious diseases, such as COVID-19, there is still space for a thorough framework that could address the aforementioned discrepancies.

In this research work, the CT scan images and CXR images are processed for the detection of radiological signs of COVID-19 using computer vision and AI techniques and classified these images as per the RSNA recommendations using a novel machine learning approach. The proposed technique is capable of minimizing inter-observer variability in image interpretation among the radiologists and hence subjectivity due to difference in experience by qualitative analysis. The system is also capable of picking up very subtle or early findings that can be missed by a radiologist. The solution is a combination of preprocessing stages, especially designed to extract the information using a set of selected feature extraction techniques. The main contributions of the proposed framework are summarized as follows:

  1. A novel entropy-based fitness optimizer function is implemented, which selects the chromosomes with maximum information. The only chromosome with maximum fitness value is selected to get the sub-optimal solution in the minimum number of iterations.

  2. To conserve maximum information and to obliterate the redundant features at the initial level, a preliminary selection process is initiated on each feature set using the entropy-controlled fitness optimizer.

  3. To exploit the complementary strength of all features, a feature fusion approach is utilized which combines all the competing features to generate a resultant feature vector.

The rest of the manuscript is organized as follows: Sect. 2 presents the two commonly used imaging analysis for COVID-19. Datasets and their collection are given in Sect. 3, followed by a detailed description of the proposed framework in Sect. 4. The results and statistical analysis are presented in Sects. 5 and 6, respectively. We conclude the manuscript in Sect. 7.

CT and CXR imaging analysis for COVID-19

COVID-19 can be detected with Nucleic Acid Amplification Tests, such as RT-PCR at very early stage [3] which is the gold standard as yet. Some studies have shown that imaging should be discouraged as primary screening tool, because these may suffer from selection bias (from inter observer variability among radiologists) with the claims that it is ten times less sensitive and less specific as compared to RT-PCR [35]. This implies that it can be negative in the early stages of the disease, and imaging features can overlap with many other infectious and noninfectious disease processes. However, in China, the chest CT has proven to have relatively higher sensitivity for COVID-19 as compared to the initial RT-PCR from swab samples [27], possibly due to the high sensitivity of CT images to lung lesion even before RT-PCR [36, 37]. As the RT-PCR takes more time, that is at least 6 h, imaging was a much faster and readily available screening tool in the surge of patients during the pandemic situation especially where RT-PCR were not available; therefore, it has played an important role in the risk stratification and screening for COVID -19. Chest radiographs are less sensitive as compared to chest CT and can give negative diagnose in case of early or mild infection, but can be used as a first line imaging modality [38]. On CXR, the findings may be airspace opacities or Ground Glass Opacities (GGO) mostly distributed in bilateral, peripheral and lower zone [38, 39]. In the early stages of disease, the CT images may be either negative or show GGO only, while at progressive stage increased GGO and crazy paving can appear [35]. The representative CXR and CT scan images of COVID-19 are shown in Fig. 3. In Fig. 3a, CXR shows Bilateral Ground Glass alveolar consolidation with peripheral distribution, which is very clear and can be seen easily, but this may not always be the case, specifically in the early stages of infection. In Fig. 3b, c, two CT images are given, showing air space consolidation and GGO. The changes are much clearly visible in Fig. 3c, while in Fig. 3b the GGO can be misinterpreted with motion blur. This is important to note as it can affect the efficient implementation of an intelligent diagnosis system. It is important to note that many respiratory viruses, such as influenza, organizing pneumonia and connective tissue disorders, can cause pneumonia like changes on both chest radiograph and CT similar to that of COVID-19, and therefore, their proper interpretation and differentiation from COVID-19 is a challenging issue [40, 41]. In order to address such ambiguities, RSNA has recommended statements on reporting CXR and CT finding related to COVID-19 [33, 34]. It is important to follow these recommendations to avoid any misinterpretation.

Fig. 3.

Fig. 3

CXR and CT images of COVID-19 patients

Dataset collection

We collected pneumonia chest CT scans of 35 subjects diagnosed positive of COVID-19 from RadioPaedia image database [2]. Links to some public domain websites are also given for verification as follows:

  1. Case 1—https://radiopaedia.org/cases/covid-19-pneumonia-2

  2. Case 3—https://radiopaedia.org/cases/covid-19-pneumonia-3

  3. Case 4—https://radiopaedia.org/cases/covid-19-pneumonia-4

  4. Case 5—https://radiopaedia.org/cases/covid-19-pneumonia-7

  5. Case 8—https://radiopaedia.org/cases/covid-19-pneumonia-8

  6. Case 15—https://radiopaedia.org/cases/covid-19-pneumonia-23

  7. Case 18—https://radiopaedia.org/cases/covid-19-pneumonia-10

  8. Case 20—https://radiopaedia.org/cases/covid-19-pneumonia-27

All the patients have their RT-PCR COVID-19 test positive and have COVID19-pneumonia. The patients’ history is also provided in the given website along with their detailed travel history. For the healthy scans, we approached [29], so that the model utilized is efficiently trained, Fig. 3a. This figure portrays healthy/normal scans in the top row, while the COVID-19 pneumonia scans are provided in the bottom row. To train the classifier, features extracted from a total of 3500 COVID19-pneumonia and 2400 normal chest CT scans are utilized from the provided link, and are resized into 512×512 resolution.

Proposed methodology

Early methods of machine learning utilize either sole or hybrid approaches for feature extraction. Though both methods have their advantages and drawbacks, generally fused feature space has more capacity to retain the dexterous features. Due to this flexibility, the hybrid approaches have gained much popularity among the researchers working in the area of computer vision. However, selection of the most appropriate feature extraction technique is quite a sensitive task, which needs to be handled carefully, otherwise, it may result in feature redundancy and, therefore, increased correlation. In this work, we utilized four different techniques—belongs to two different categories, statistical and texture. Two feature families are not considered, color and shape, because of their limited impact and significance in this application. The proposed framework, Fig. 4, is the subject of discussion in the following subsections.

Fig. 4.

Fig. 4

Proposed framework of COVID19 prediction using CT scans

Discrete wavelet transform features

The rationale behind selecting the discrete wavelet transform (DWT) for texture feature extraction is its ability to be invariant to translation, scaling and rotation. Further, in DWT, the contours can be requited from the coarsest to the finer scale, enabling the formulation to handle noise effectively. The 2D wavelet decomposition for images is similar to the 1D decomposition, in which the 2D wavelets basis Ψl,m(t) and scaling basis ϕm(t) are obtained by taking the tensor products of 1D wavelets and scaling functions. For a 2D image, the DWT performs critical subsampling along both rows and columns, and these subbands information is utilized in the next level decomposition. The followed approach utilizes filter banks, described as:

αl(k)=u6h0(u-2k)αl+1(u) 1
ςl(k)=uh1(u-2k)αl+1(u) 2

The lowpass or averaged coefficients αl(k) are created by half-band lowpass filter ho, whereas the highpass or detailed coefficients ςl(k) are created by half-band highpass filter h1. From the equations, it can be observed that filtering αl+1 with h0 and h1 produces αl(k) and ςl(k), followed by decimation by a factor of 2.

To compute DWT coefficients for two levels, two-stage filter banks are required. The initial scale, l+1, in terms of αl+1 is the original signal, which after one level of decomposition produces highpass ςl and lowpass coefficients αl. A batch of COVID-19 chest CT scans are represented by {x1,x2,x3,,xn}X(m,n)R{m,n}, where m and n are the rows and columns, respectively. Initially, both filters H0(ω) and H1(ω) are applied on x1m,n to generate a pair of images with both low and high frequencies. Afterward, the filtered images are sub-sampled by a factor of 2 and are forwarded to the next series of filters along the columns. The decimation by a factor of 2 is again carried out after filtration process in the columns. A single column decomposition generates four subband images, {LL,LH,HL,HH} of size M2,N2. The whole computation is performed to generate set of features:

LLl(u,v)=mnh0(m-2u)h0(n-2v).LLl+1(m,n), 3
HLl(u,v)=mnh0(m-2u)h1(n-2v).LLl+1(m,n), 4
LHl(u,v)=mnh1(m-2u)h0(n-2v).LLl+1(m,n), 5
HHl(u,v)=mnh1(m-2u)h1(n-2v).LLl+1(m,n), 6

The lowpass L and highpass H filters are represented by the alphabet letters on the sub images.

Extended segmentation-based fractal texture analysis (ESFTA)

As discussed earlier, texture features play much more significance role in the recognition process compared to other set up of features including shape and color. Therefore, in this work, we are employing our existing work [42] to extract the texture features of COVID-19 chest CT scans. In this technique, the fractal dimensions are computed from the stack of binary images. The technique works in two steps: (1) image partitioning into stack of binary images using pair threshold binary decomposition (PTBD), (2) fractal analysis of each binary image based on boundaries, pixel count, mean gray level. graphic file with name 10044_2020_950_Figa_HTML.jpg

Statistical features

Generally, data follow a normal distribution, which describes how the values of a variable are distributed. In case of normal distribution, it has two fundamental parameters; the mean and standard deviation. Observing a chest CT scan image under M modalities, with an assumption that all the images are spatially registered, i.e., healthy images are same and pixels correspond to the same location. From a pool of images, xi, where i={1,2,,n}X, each of the healthy image follows a Gaussian distribution Π(λ,σ). One observes a noticeable change in the distribution when provided an infected COVID-19 chest CT scan image. By following the underlying concept, two statistical parameters, entropy and skewness, are selected based on their vast applications in the field of machine learning.

Skewness

It measures asymmetry of the probability distribution about its mean. The skewness value can be negative, zero or positive. If the value is negative then the distribution curve spreads out more to the left of the mean, whereas, in case of positive value, it leans toward the right. The skewness distribution is described using the relation:

ς=E(x-μ)3σ3 7

where μ and σ are the mean and standard deviation of a random variable x, and E(t) denotes the expected value.

Entropy

Entropy offers the information regarding randomness in a signal by cogitating the system’s disorder. Due to this potential, entropy, in the current perspective, offers a useful information that can be utilized in feature representation. In this framework, Shannon entropy is utilized, which significantly improves the overall accuracy, by embedding the most relevant feature information. For both COVID-19 chest CT scans x1,x2,,xNX(m,n), contains N samples. The image space has β measure with β(X)=1, the Shannon entropy is calculated as:

E(X)=-k=1Nxklogζ(xk) 8

where ζ(xk) is observing probability for a particular pixel matrix/vector of X. This whole concept allows us to identify the most superior and dominant pixels with a better variation and with least correlation.

Feature selection framework

The genetic algorithm (GA) belongs to a class of stochastic search algorithms, which on the principle of survival of the fittest finds the sub optimal solution from a pool of solutions. In the GA framework, the population is developed by combining a set of chromosomes, where each chromosome constitutes a possible solution. In the proposing scenario, the extracted set of features are independently plugged into the GA block. The most discriminant chromosome/solution is later selected using the proposed entropy-based fitness optimizer.

Both texture and statistical features are used to generate two pairs of chromosomes, where each chromosome represents a feature type.

Φk(X)=ΨWL1(X),ΨESFA2(X),Φς3(X),ΦE4(X);kZ, 9

Z={1,,4} is the set of bounded integers, representing a feature chromosome. The entire population is generated from each set of chromosome c, continuous valued vector, having Gj genes. The continuous domain offers more convergence possibilities and also minimizes the probability for a generation to be stuck in a local minima.

Cjk=G1k,G2k,,G3k,,,Gmk, 10

where k is the chromosome index in a population, and m is the length of a chromosome. In what follows, we present a genetic operators including proposed crossover, mutation and selection operators for a thorough technical analysis.

Median-replacer crossover

Selecting a pair of chromosome, Cj1={(G11,G21,G31,,Gm1),(G12,G22,G32,,Gm2)}, where j={1,2,,m}, for an average-replacer crossover operation. Two offsprings, (Ol1 and Ol2), are generated as: Ol1=λCj1(1-λ)CJ2 and Ol2=λCj2(1-λ)CJ1. The median of both offsprings is later replaced using the min/max value extracted from the parent chromosomes. A max value is assigned 1 (max1), whereas the min value is assigned 0 (min1). A binary random sequence is generated to select min/max value from the first selected chromosome Cj1, and the same procedure applies to the second selected chromosome. Based on the generated binary rand sequence, median values of both offsprings are updated. An inversion mutation is applied on the selected number of chromosomes, whereas, for the selection, both healthy and non-healthy parents are selected for the next generation on the basis of entropy-based fitness optimizer [43].

Entropy-based fitness optimizer

To select the next-generation offsprings, the fitness of each chromosome needs to be evaluated. In this work, we developed a novel entropy-based fitness optimizer. The whole idea revolves around the fundamental property of the feature randomness calculated by the entropy function. More the entropy value is, greater the chances of a healthy chromosome. Here, the entropy calculator identifies the maximum randomness by controlling the uncertainty. For a real-valued chromosome vector Cj1={(G11,G21,G31,,Gm1)}, the Shannon entropy is calculated using the relation:

Eς=-q=0m-1Gq0/ΔGlog2βGq0/ΔG 11

where Gq1 is the gene q of the first chromosome.

Feature fusion

Features fusion is a robust strategy pursued by several researchers in the field of machine learning. The original feature space, in most of the cases, does not contain sufficient information compared to the fused feature space. Therefore, in this work, we opted feature fusion strategy to generate a resultant feature vector with enriched information. All the down-sampled features from GA block are later fused by following a cascaded design. These horizontally concatenated feature vectors is later forwarded to the classification block for final labeling using Naive Bayes classifier [44].

Results and discussion

The proposed framework for COVID-19 pneumonia is evaluated in this section with both empirical and graphical results. For the validation, 35 subjects are considered with their Coronavirus test positive, with details provided in Sect. 3. A fair training/testing ratio of (70:30) is being followed with 70% as training data and the rest is treated as testing data. To generalize the empirical results and to recognize the precise stats, a tenfold cross-validation technique is exercised. For the final classification, a Naive Bayes classifier is selected based on its improved performance. A fair comparison is also provided with the existing state-of-the-art classifiers including fine KNN (F-KNN) [45], linear support vector machine (L-SVM) [46], ensemble bagged tree (EBT) [47] and fine tree (F-Tree) [48] classifiers. To authenticate the proposed method, several performance measures deem necessary are chosen including sensitivity (SEN), precision (PR), specificity (SPE), an area under the curve (AUC) and accuracy (ACC). The mathematical form of aforementioned measures is provided in the following equations.

SEN=δTPδTP+δFN×100% 12
SPE=δTNδTN+δFP×100% 13
PR=δTPδTP+δFP×100% 14
ACC=δTP+δTNδTP+δTN+δFP+δFN×100% 15

where δTP represents true positive, δTN represents true negative, δFP denotes false positive rate, and δFN represents false negative rate. A few samples results are demonstrated in Fig. 5, where one can observe a binary labeling; corona positive and normal.

Fig. 5.

Fig. 5

Proposed prediction results. a Original images; b proposed predicted labeled image

The results are compiled by taking into consideration three different scenarios; (1) accuracy achieved using independent features, (2) accuracy achieved after employing GA for feature selection and (3) with proposed feature selection and fusion method.

Case 1 considers the features extracted independently from each technique for the classification; this is tabulated in Table 2. For a fair comparison, different classifiers are being tested and against each technique. The final results are as expected, and the Naive Bayes classifier outperforms other state-of-the-art with an average accuracy of 80.83%, while the second-best average accuracy achieved is with F-KNN (79.52%). The accuracy comparison of different feature extraction techniques with proposed is provided in Fig. 6. It is apparent from the bar plot that the independent features are of no match to the fused features for this application. One can also observe, with DWT features, most of the classifiers performed well compared to other features. This clearly shows, DWT features in this application, would be an appropriate choice compared to other features, either used solely or in the fused form.

Table 2.

Accuracy calculated using different feature extraction techniques

Classifier Entropy DWT Skewness SFTA Accuracy (%)
Naive Bayes 81.2
82.7
79.6
79.8
F-KNN 79.8
81.6
76.5
80.2
L-SVM 78.1
80.1
79.4
75.6
EBT 76.3
80.9
72.6
80.4
F-Tree 80.3
73.4
75.7
77.9

Fig. 6.

Fig. 6

Accuracy comparison of different feature extraction techniques using bar plots

In case 2, the selected features from GA are forwarded to the classifier for final labeling, Table 3. The same trend is being followed after feature selection step, and Naive Bayes classifier works exceptionally well almost for all kind features by achieving an average accuracy of 85.55%. F-KNN worked second best by achieving an average accuracy of 84%. One more time, with DWT features, classification results are exceptional with almost all the classifiers. The average classification accuracy achieved using GA-DWT features by all the selected classifiers are 84.86% compared to 83.85% using G-SFTA. Figure 7 demonstrates that the average accuracy achieved after GA increased compared to stand alone features. A vertical bar clearly indicates that the accuracy margin between the proposed and after GA selection is still comparable, which strengthens the positive significance of feature fusion.

Table 3.

Accuracy comparison of different feature extraction techniques after applying GA-based feature selection

Classifier G-Entropy G-DWT G-Skewness G-SFTA Accuracy (%)
Naive Bayes 86.7
87.4
80.4
87.7
F-KNN 84.6
85.9
80.1
85.4
L-SVM 84.2
85.4
78.6
82.9
EBT 82.3
81.8
76.3
84.6
F-Tree 81.7
83.8
79.2
78.6

Fig. 7.

Fig. 7

Accuracy comparison of different feature extraction techniques using bar plots after applying GA-based feature selection

Using the proposed framework, the achieved accuracy using the Naive Bayes classifier is 92.6%, whereas a few other classifiers (EBT, L-SVM and F-KNN) behave significantly better to achieving an average accuracy of 92.2%,92.1% and 92.0%, respectively. The authenticity of proposed framework is further validated from the selected performance parameters including sensitivity (92.5%), specificity (92.0%), precision (92.5%) and the AUC (0.99), see Table 4. From the sensitivity and specificity values, it is quite obvious that the proposed framework has successfully managed to achieve a high true positive and negative rates by correctly classifying the actual positive and actual negative samples. To further describe the performance of a classifier on a set of test data, a confusion matrix is provided, Table 5. From the stats, one can develop a clear understanding, that out of total test samples, 93% are correctly labeled as COVID-19 infected, whereas around 7% are misclassified as normal.

Table 4.

A comparison of state-of-the-art classifiers using proposed GA controlled feature selection and fusion method

Classifier Sensitivity (%) Precision (%) Specificity (%) AUC Accuracy (%)
Naive Bayes 92.5 92.5 92.0 0.99 92.6
F-KNN 92.0 92.0 91.0 0.96 92.0
L-SVM 91.7 92.0 91.0 0.96 92.1
EBT 92.2 92.3 94.0 0.98 92.2
F-Tree 91.5 91.5 93.0 0.98 91.6

Significant values are shwon in bold

Table 5.

Confusion matrix of Naive Bayes classifier after applying optimized GA for features selection

COVID-19 93% 7%
Normal 8% 92%
COVID-19 Normal

In addition, we compare the proposed method results on different training and testing samples such as 70:30, 60:40, 90:10 and so on. The results are plotted in Fig. 5. In this figure, it is shown that the 90:30 approach results are fine but if we consider the standard process of validation, then 70:30 approach results are more useful (Fig. 8).

Fig. 8.

Fig. 8

Comparison of proposed results on different training and testing sets

Statistical significance

The objective here in performing the statistical analysis is to gain a high level of confidence in the proposed method. The results are statistically significant, if they are likely not caused by chance. We employed the analysis of variance (ANOVA) to demonstrate, that either the results are statistically significant or not. In this work, we consider the proposed scenario for three different classifiers (Naive Bayes, EBT, L-SVM)—selected on the basis of their improved performance compared to the rest. A Shapiro–Wilk test is performed for assumption of normality, while Bartlett’s test—for homogeneity of variance with a significance level α=0.01. The means of our approach are x¯1, x¯2, and x¯3—calculated from the overall accuracy of both classifiers. The null hypothesis H0, given that x¯1=x¯2=x¯3, while the alternative hypothesis Ha given that x¯1x¯2x¯3. We computed the p value and tested the null hypothesis, H0, if it is rejected, p<α, then we will be applying Bonferroni post hoc test.

For the proposed entropy controlled GA method (E-GA), and with selected classifiers (Naive Bayes, EBT and L-SVM), the Shapiro–Wilk test generated p value, pu=0.8002, pv=0.9152, and pl=0.6878. By following the Bartlett’s test, the associated Chi-squared probabilities are: pu=0.371, pv=0.339, and pl=0.410. From the calculated p values of two different classifiers, which are significantly greater than α. Therefore, from the test (normality and equality of variances), we failed to reject the null hypothesis H0, and confirm that the data were distributed normally, and their variances are homogeneous. ANOVA test including five different parameters (degree of freedom (df), a sum of squared deviation (SS), mean squared error (MSE), F-statistics, and p value) is shown in Table 6. The performance range of three selected classifiers based on the proposed method is shown in Fig. 9.

Table 6.

ANOVA test on two selected classifiers based on the proposed method

Variance source SS df MSE F-Statistics p value
Between 0.6212 1 0.313 0.431 0.667
Within 4.324 4 0.716
Total 4.924 5

Fig. 9.

Fig. 9

Box-plot of accuracy values on the selected classifiers (1: SVM-C, and 2: SVM-Q)

The results are also validated by utilizing Bonferroni post hoc test, which is the most common approach to be applied whenever there exists a chance of a significant difference between the means of multiple distributions. It was certified that the proposed method performed better compared to several existing methods.

Conclusion

A computerized technique is proposed in this work for the prediction of COVID-19 from the CT scans. Textural and statistical features are extracted from raw CT images, and then, only best features are selected based on optimized genetic algorithm. The selected features are serially concatenated and later classified using the Naive Bayes classifier. The experimental process is performed on the collected COVID-19 positive and healthy samples and shows the proposed method to be effective. The main contribution of this work is an optimized genetic algorithm for best selection. Using this algorithm, the accuracy of individual feature type is improved and when all selected features are combined, then a significant change has been observed in the accuracy. Based on the performance of this algorithm, we concluded that the selection of most relevant features improves the accuracy, but on the other side, it is a high chance that we miss the important features that play a contribution in the improvement of prediction accuracy. Also, this problem may occur when we have more patients data for final testing. Therefore, in the future studies, we will focus on the reduction of these features.

Footnotes

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Change history

3/12/2021

A Correction to this paper has been published: 10.1007/s10044-021-00969-x

Contributor Information

Tallha Akram, Email: tallha@ciitwah.edu.pk.

Muhammad Attique, Email: attique@ciitwah.edu.pk.

Robertas Damaševičius, Email: robertas.damasevicius@polsl.pl.

References

  • 1.Yan L, Zhang H-T, Goncalves J, Xiao Y, Wang M, Guo Y, Sun C, Tang X, Jing L, Zhang M et al (2020) An interpretable mortality prediction model for COVID-19 patients. Nat Mach Intell 1–6
  • 2.Laboratory testing for coronavirus disease 2019 (COVID-19) in suspected human case. https://www.who.int/publications-detail/laboratory-testing-for-2019-novel-coronavirus-in-suspected-human-cases-20200117. Accessed 29 April 2020
  • 3.Accelerated emergency use authorization (EUA) summary COVID-19 RT-PCR test (Laboratory Corporation of America). https://www.fda.gov/media/136151/download. Accessed 29 April 2020
  • 4.Pan Y, Guan H, Zhou S, Wang Y, Li Q, Zhu T, Hu Q, Xia L (2020) Initial CT findings and temporal changes in patients with the novel coronavirus pneumonia (2019-nCoV): a study of 63 patients in Wuhan, China. Eur Radiol [DOI] [PMC free article] [PubMed]
  • 5.Zhu N, Zhang D, Wang W, Li X, Yang B, Song J, Zhao X, Huang B, Shi W, Lu R, Niu P, Zhan F, Ma X, Wang D, Xu W, Wu G, Gao GF, Tan W, China Novel Coronavirus Investigating and Research Team A novel coronavirus from patients with pneumonia in China, 2019. New Engl J Med. 2020;382:727–733. doi: 10.1056/NEJMoa2001017. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6.Edwards A (2020) COVID-19 tests: how they work and what’s in development. https://theconversation.com/covid-19-tests-how-they-work-and-whats-in-development-134479. Accessed 31 Mar 2020
  • 7.Valente F, Costa C, Silva A. Dicoogle, a PACS featuring profiled content based image retrieval. PLoS ONE. 2013;8:e61888. doi: 10.1371/journal.pone.0061888. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.Rohman M (2020) Has radiologist burnout. https://www.healthimaging.com/topics/practice-management/has-radiologist-burnout-finally-reached-tipping-point. Accessed 31 May 2020
  • 9.Harvey H (2019) The UK desperately needs a radiology artificial intelligence incubator. https://towardsdatascience.com/the-uk-desperately-needs-a-radiology-ai-incubator-205553dbac43. Accessed 17 Sept 2019
  • 10.Brady A, Laoide R, McCarthy P, McDermott R. Discrepancy and error in radiology: concepts, causes and consequences. Ulster Med J. 2012;8:1–3. [PMC free article] [PubMed] [Google Scholar]
  • 11.Brady APA. Error and discrepancy in radiology: inevitable or avoidable? Insights Imaging. 2017;8:171–182. doi: 10.1007/s13244-016-0534-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12.Waite S, Grigorian A, Alexander RG, Macknik SL, Carrasco M, Heeger DJ, Martinez-Conde S. Analysis of perceptual expertise in radiology—current knowledge and a new perspective. Front Hum Neurosci. 2019;13:213. doi: 10.3389/fnhum.2019.00213. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.Ekpo EU, Alakhras M, Brennan P. Errors in mammography cannot be solved through technology alone. Asian Pac J Cancer Prev APJCP. 2018;19(2):291–301. doi: 10.22034/APJCP.2018.19.2.291. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14.Ganesan K, Acharya UR, Chua CK, Min LC, Abraham KT, Ng K. Computer-aided breast cancer detection using mammograms: a review. IEEE Rev Biomed Eng. 2013;6:77–98. doi: 10.1109/RBME.2012.2232289. [DOI] [PubMed] [Google Scholar]
  • 15.Breast cancer screening in the UK. http://www.cancerresearchuk.org/aboutcancer/ type/breastcancer/ about/screening/whoisscreenedforbreastcancer. Accessed 5 May 2016
  • 16.Pinto A, Brunese L, Pinto F, Reali R, Daniele S, Romano L (2012) The concept of error and malpractice in radiology. In: Seminars in ultrasound, CT and MRI, errors and malpractice in radiology, vol 33, No. 4, pp 275–279 [DOI] [PubMed]
  • 17.Pinto A, Brunese L. Spectrum of diagnostic errors in radiology. World J Radiol. 2010;2:377–383. doi: 10.4329/wjr.v2.i10.377. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18.Wang S, Summers RM. Machine learning and radiology. Med Image Anal. 2012;16(5):933–951. doi: 10.1016/j.media.2012.02.005. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19.Ke Q, Zhang J, Wei W, Połap D, Woźniak M, Kośmider L, Damas̆evĭcius R. A neuro-heuristic approach for recognition of lung diseases from X-ray images. Expert Syst Appl. 2019;126:218–232. doi: 10.1016/j.eswa.2019.01.060. [DOI] [Google Scholar]
  • 20.Chouhan V, Singh SK, Khamparia A, Gupta D, Tiwari P, Moreira C, Damaševičius R, de Albuquerque VHC. A novel transfer learning based approach for pneumonia detection in chest X-ray images. Appl Sci (Switzerland) 2020;10(2):559. [Google Scholar]
  • 21.Khan MA, Ashraf I, Alhaisoni M, Damaševičius R, Scherer R, Rehman A, Bukhari SAC. Multimodal brain tumor classification using deep learning and robust feature selection: a machine learning application for radiologists. Diagnostics. 2020;10(8):565. doi: 10.3390/diagnostics10080565. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22.Keen CE (2019) Artificial intelligence versus 101 radiologists. https://physicsworld.com/a/artificial-intelligence-versus-101-radiologists. Accessed 17 Sept 2019
  • 23.Fischer AM, Varga-Szemes A, van Assen M, Griffith LP, Sahbaee P, Sperl JI, Nance JW, Schoepf UJ. Comparison of artificial intelligence-based fully automatic chest CT emphysema quantification to pulmonary function testing. Am J Roentgenol. 2020;214:1–7. doi: 10.2214/AJR.19.21572. [DOI] [PubMed] [Google Scholar]
  • 24.Sahlol AT, Elaziz MA, Jamal AT, Damaševičius R, Hassan OF. A novel method for detection of tuberculosis in chest radiographs using artificial ecosystem-based optimisation of deep neural network features. Symmetry. 2020;12:1146. doi: 10.3390/sym12071146. [DOI] [Google Scholar]
  • 25.Christe A, Peters AA, Drakopoulos D, Heverhagen JT, Geiser T, Stathopoulou T, Christodoulidis S, Anthimopoulos M, Mougiakakou SG, Ebner L. Computer-aided diagnosis of pulmonary fibrosis using deep learning and CT images. Investig Radiol. 2019;54:627–632. doi: 10.1097/RLI.0000000000000574. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 26.Fau Li L, Qin L, Xu Z, Yin Y, Wang X, Kong B, Bai J, Lu Y, Fang Z, Song Q, Cao K, Liu D, Wang G, Xu Q, Fang X, Zhang S, Xia J, Xia J (2020) Artificial intelligence distinguishes COVID-19 from community acquired pneumonia on chest CT. Radiology 16 [DOI] [PMC free article] [PubMed]
  • 27.Ai T, Yang Z, Hou H, Zhan C, Chen C, Lv W, Tao Q, Sun Z, Xia L (2020) Correlation of chest CT and RT-PCR testing in coronavirus disease 2019 (COVID-19) in China: a report of 1014 cases. Radiology [DOI] [PMC free article] [PubMed]
  • 28.Narin A, Kaya C, Pamuk Z (2020) Automatic detection of coronavirus disease (COVID-19) using X-ray images and deep convolutional neural networks. arXiv:2003.10849 [DOI] [PMC free article] [PubMed]
  • 29.Loey M, Smarandache F, Khalifa NEM. Within the lack of chest COVID-19 X-ray dataset: a novel detection model based on GAN and deep transfer learning. Symmetry. 2020;12(4):651. doi: 10.3390/sym12040651. [DOI] [Google Scholar]
  • 30.Xu X, Jiang X, Ma C, Du P, Li X, Lv S, Yu L, Chen Y, Su J, Lang G, Li Y, Zhao H, Xu K, Ruan L, Wu W (2020) Deep learning system to screen coronavirus disease 2019 pneumonia. arXiv:2002.09334 [physics.med-ph] [DOI] [PMC free article] [PubMed]
  • 31.Gozes O, Frid-Adar M, Greenspan H, Browning PD, Zhang H, Ji W, Bernheim A, Siegel E (2020) Rapid AI development cycle for the coronavirus (COVID-19) pandemic: initial results for automated detection and patient monitoring using deep learning CT image analysis. arXiv:2003.05037 [eess.IV]
  • 32.Sahlol AT, Yousri D, Ewees AA, Al-qaness MAA, Damaševičius R, Elaziz MA. COVID-19 image classification using deep features and fractional-order marine predators algorithm. Sci Rep. 2020;10(1):1–15. doi: 10.1038/s41598-019-56847-4. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 33.Simpson S, Kay FU, Abbara S, Bhalla S, Chung JH, Chung M, Henry TS, Kanne JP, Kligerman S, Ko JP, Litt H. Radiological Society of North America expert consensus statement on reporting chest CT findings related to COVID-19. Endorsed by the Society of Thoracic Radiology, the American College of Radiology, and RSNA. Radiol Cardiothorac Imaging. 2020;2:e200152. doi: 10.1148/ryct.2020200152. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 34.RSNA offers video series and discussion on structured reporting for assessment of COVID-19. https://www.rsna.org/en/news/2020/April/Structured-Reporting-COVID-19-Videos. Accessed 25 Apr 2020
  • 35.Murphy A, Bell DJ (2020) COVID-19. https://radiopaedia.org/articles/covid-19-3. Accessed 31 Mar 2020
  • 36.Rouger M (2020) Imaging the coronavirus disease COVID-1. https://healthcare-in-europe.com/en/news/imaging-the-coronavirus-disease-covid-19.html#. Accessed 28 Apr 2020
  • 37.Mei X, Lee H-C, Diao K-Y, Huang M, Lin B, Liu C, Xie Z, Ma Y, Robson PM, Chung M, et al. Artificial intelligence—enabled rapid diagnosis of patients with COVID-19. Nat Med. 2020;26:1–5. doi: 10.1038/s41591-020-0931-3. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 38.Wong HYF, Lam HYS, Fong AH-T, Leung ST, Chin TW-Y, Lo CSY, Lui MM-S, Lee JCY, Chiu KW-H, Chung T, Lee EYP, Wan EYF, Hung FNI, Lam TPW, Kuo M, Ng M-Y (2019) Frequency and distribution of chest radiographic findings in COVID-19 positive patients. Radiology [DOI] [PMC free article] [PubMed]
  • 39.Rodrigues JCL, Hare SS, Edey A, Devaraj A, Jacob J, Johnstone A, McStay R, Nair A, Robinson G. An update on COVID-19 for the radiologist—a British Society of Thoracic Imaging statement. Clin Radiol. 2020;75:323–325. doi: 10.1016/j.crad.2020.03.003. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 40.Zhao D, Yao F, Wang L, Zheng L, Gao Y, Ye J, Guo F, Zhao H, Gao R (2020) A comparative study on the clinical features of coronavirus 2019 (COVID-19) pneumonia with other pneumonias. Clin Infect Dis [DOI] [PMC free article] [PubMed]
  • 41.Bai HX, Hsieh B, Xiong Z, Halsey K, Choi JW, Tran TML, Pan I, Shi L-B, Wang D-C, Mei J, Jiang X-L, Zeng Q-H, Egglin TK, Hu P-F, Agarwal S, Xie F, Li S, Healey T, Atalay MK, Liao W-H (2020) Performance of radiologists in differentiating COVID-19 from viral pneumonia on chest CT. Radiology [DOI] [PMC free article] [PubMed]
  • 42.Akram T, Khan MA, Sharif M, Yasmin M (2018) Skin lesion segmentation and recognition using multichannel saliency estimation and M-SVM on selected serially fused features. J Ambient Intell Humaniz Comput 1–20
  • 43.Hussain A, Muhammad YS, Sajid MN. Performance evaluation of best–worst selection criteria for genetic algorithm. Math Comput Sci. 2017;2(6):89–97. doi: 10.11648/j.mcs.20170206.12. [DOI] [Google Scholar]
  • 44.Rish I et al (2001) An empirical study of the Naive Bayes classifier. In: IJCAI 2001 workshop on empirical methods in artificial intelligence, vol 3, pp 41–46
  • 45.Xu Y, Zhu Q, Fan Z, Qiu M, Chen Y, Liu H. Coarse to fine k nearest neighbor classifier. Pattern Recognit Lett. 2013;34(9):980–986. doi: 10.1016/j.patrec.2013.01.028. [DOI] [Google Scholar]
  • 46.Jia-Zhi D, Wei-Gang L, Xiao-He W, Jun-Yu D, Wang-Meng Z. L-SVM: a radius-margin-based svm algorithm with logdet regularization. Expert Syst Appl. 2018;102:113–125. doi: 10.1016/j.eswa.2018.02.006. [DOI] [Google Scholar]
  • 47.Dietterich TG. An experimental comparison of three methods for constructing ensembles of decision trees: bagging, boosting, and randomization. Mach Learn. 2000;40(2):139–157. doi: 10.1023/A:1007607513941. [DOI] [Google Scholar]
  • 48.Safavian SR, Landgrebe D. A survey of decision tree classifier methodology. IEEE Trans Syst Man Cybern. 1991;21(3):660–674. doi: 10.1109/21.97458. [DOI] [Google Scholar]

Articles from Pattern Analysis and Applications are provided here courtesy of Nature Publishing Group

RESOURCES