Abstract
Worldwide COVID-19 is a highly infectious and rapidly spreading disease in almost all age groups. The Computed Tomography (CT) scans of lungs are found to be accurate for the timely diagnosis of COVID-19 infection. In the proposed work, a deep learning-based P-shot N-ways Siamese network along with prototypical nearest neighbor classifiers is implemented for the classification of COVID-19 infection from lung CT scan slices. For this, a Siamese network with an identical sub-network (weight sharing) is used for image classification with a limited dataset for each class. The feature vectors are obtained from the pre-trained sub-networks having weight sharing. The performance of the proposed methodology is evaluated on the benchmark MosMed dataset having categories zero (healthy control) and numerous COVID-19 infections. The proposed methodology is evaluated on (a) chest CT scans provided by medical hospitals in Moscow, Russia for 1110 patients, and (b) case study of low-dose CT scans of 42 patients provided by Avtaran healthcare in India. The deep learning-based Siamese network (15-shot 5-ways) obtained an accuracy of 98.07%, the sensitivity of 95.66%, specificity of 98.83%, and F1-score of 95.10%. The proposed work outperforms the COVID-19 infection severity classification with limited scans availability for numerous infection categories.
Keywords: COVID-19 infection, CT scan, CNN, Siamese network
1. Introduction
In late 2019, the novel Corona Virus Infection Disease (COVID-19) is detected in humans and declared a pandemic by World Health Organization (WHO) on 11 March 2020 [1]. The COVID-19 belongs to the virus family of Severe Acute Respiratory Syndrome-Corona Virus 2 (SARS-CoV2) and Middle East Respiratory Syndrome-Corona Virus (MERS-CoV). In the end of July 2021, globally there were 1,94,080,019 active cases of coronavirus, 4,162,304 deaths reported, and 3, 696,135,440 people were vaccinated [2]. In the year 2021, researchers have claimed the presence of numerous mutants of COVID-19 such as B.1.1.7, B.1.351, and P.1, etc. [3], [4]. The mutation of coronavirus is having a huge impact on the transmission scale of infection in humans, the efficacy of testing kits, vaccination, and treatment procedures [5].
In the year 2021, the highest number of coronavirus cases are found in India with an average of 0.35 million cases per day in April [6], [7]. The number of cases rapidly increased due to community transmission and the average time duration of the RT-PCR test result took 4–5 days with a very high false-negative rate. The coronavirus majorly affected the respiratory system, i.e., breathing issues, and severe lung infection. Thus, chest CT scans are used to evaluate the severity level of coronavirus infection in suspected patients. The COVID-19 infection spreads through saliva or cough droplets of an infected person through the air. Thus, with the increment in COVID-19 cases, researchers are focusing on faster and accurate techniques for severity of COVID-19 infection measurements in lungs.
In clinical practice, COVID is detected using swab tests, namely, RT-PCR, Rapid Antigen Test (RAT), and radiographic images of lungs (X-rays, CT-scans). The COVID-19 antigen tests identify the antibodies of coronavirus. However, it suffers from various disadvantages such as the complex procedure of taking samples from the nose/throat, time consumption in the generation of the diagnosis result, and a high false-negative rate [8]. The antigen kits are customized for only COVID-19 infection detection and are unable to evaluate the impact of infection levels in the respiratory system or various organs. To overcome these issues, clinicians prefer a chest Computed Tomography (CT) scan/X-ray scan of COVID-19 infected patients.
The chest X-ray of COVID-19 suspected patients provides a preliminary diagnosis of the dense tissues and the impact of infection in the lungs in a 2-D view [9]. However, in chest CT scans, patches due to infection are visible in 3-D views (axial, coronal, sagittal). Therefore, to speed up the COVID-19 infection detection process, radiologists prefer chest CT scans over X-rays. Doctors differentiate the healthy control CT scan from the coronavirus infection patient’s CT scan measurements based on consolidation, pleural effusion, and Ground Glass Opacity’s (GGOs) in the lungs as shown in Fig. 1. Further, the measurements of these patches in the left and right lung predict the severity of coronavirus infection. However, manual annotation of the patches due to COVID-19 infection is an error-prone and tedious task.
Fig. 1.
Biomarkers are identified by the experts in the chest CT scan slice of the COVID-19 infected patient.
In recent times, researchers focused on the development of a future prediction model for COVID-19 cases variants, and improvements in the available COVID-19 measurement equipment as well [10], [11]. In [12], low dose CT scan measurements are suggested for the severity of COVID-19 infection detection in mild COVID-19 infected patients along with children and pregnant women CT scans. The COVID-19 cases are predicted using time series models such as Eigenvalue Decomposition of Hankel Matrix (EVDHM) along with the Auto-regressive Integrated Moving Average (ARIMA) model [13], Composite Monte Carlo (CMC) simulation [14]. To monitor the lung condition of critical covid patients on ventilators, a data acquisition system, i.e., Electrical Impedance Tomography (EIT) system is proposed [15].
The major challenges associated with COVID-19 infection severity measurement using CT scan-based modality are (a) limited dataset is available with expert annotation, (b) multi-class classification of the level of COVID-19 infection is not given, etc. Thus few-shot learning frame-work is introduced for classification on a small sample dataset or no labeled dataset [16]. The Few-shot learning approach include metric-learning and meta-learning-based techniques. In the computer vision field, a meta learning-based approach is introduced for face identification [17], signature verification [18], and a variety of plant leaves classification tasks, etc. [19]. In medical field, numerous disease detection task with limited dataset availability is resolved used few-shot learning based Siamese networks [20], [21], [22]. The Meta-COVID networks (2-ways, 10-shot) are implemented for binary classification of COVID-19 and healthy X-ray images with an accuracy of 96.5%, specificity of 98.4%, and precision of 98% [23].
The main contributions of the proposed work are as follows:
1. The deep learning encoder-based Siamese network is proposed for the multi-class classification of COVID-19 infection from lung CT scan slices.
2. The P-shot N-ways (N denotes the number of classes, and P denotes the images per class) based meta-learning framework evaluated using contrastive loss.
3. The Gradient-weighted Class Activation Mapping (Grad-CAM) is used to visualize the multi-class infection features learned by the proposed McS-Net.
4. The proposed methodology is implemented on the Indian dataset of suspected COVID-19 patients with numerous infection levels.
5. Performance evaluation parameters such as sensitivity, specificity, accuracy, F1-score, precision, etc. are obtained for the considered datasets with numerous variations in the encoder network and some samples per class prove the superiority of the proposed framework.
The rest of the article is organized as follows: Section 2 presents the related work; Section 3 puts forth the brief details of the COVID-19 based CT scan dataset. Section 4 illustrates the materials and methods; Section 5 presents the experimentation and discussion. Section 6 concludes the proposed work along with the future scope.
2. Related work
Various Artificial Intelligence (AI) based techniques are implemented by researchers to auto-mate the COVID-19 infection detection, classification, segmentation, and future prediction task. In [24], X-ray modality is used for multi-class classification of COVID-19 infection with an accuracy of 97.72 ± 0.95% for severe infection, 86.90 ± 3.20% for moderate level, and 61.80 ± 5.49% for mild COVID-19 infection. The Internet of Things (IoT) based pre-trained deep learning model detected the COVID-19 with a sensitivity of 96.73% [25]. The IoT-based approach using a blood sample, X-ray images, body temperature of the covid suspected patient obtained the specificity of 97.95%. Gen-ProtoPNet based deep learning model obtained the classification accuracy of 87.27% on chest X-rays for three classes (normal, COVID-19, and Pneumonia) [26]. The ResNet-SVM obtained a classification accuracy of 93% and sensitivity of 88% on X-ray images of COVID-19 patients [27]. The Genetic Algorithm (GA) with specific hyperparameter optimization technique is used for binary classification of COVID-19 infection [28]. However, the technique takes more than 25 h to get the best results due to the specific optimization parameters requirement. Table 1 put forth the brief details of state-of-the-art techniques available for COVID-19 diagnosis using radiographic images.
Table 1.
Performance evaluation of AI-based techniques available in the literature for COVID-19 diagnosis using chest CT scan.
Sr. No. | Techniques | Augmentation | Dataset | Accuracy | Sensitivity | Specificity |
---|---|---|---|---|---|---|
1 | 7-layer CNN+stochastic pooling [29] | 14-ways | Normal = 320 COVID-19 = 320 |
94.03% | 94.44% | 93.63% |
2 | Joint Classification Segmentation (JCS) system [30] | Image mixing, random horizontal flip, and crop | Normal = 350 COVID-19 = 400 |
– | 95.0% | 93.0% |
3 | Few-shot COVID-19 (2- ways, 6-shot) [31] |
Stochastic augmentation with random cropping (with resizing and color distortion) | Normal = 397 COVID-19 = 349 |
88.5% | 88.6% | 89.9% |
4 | Attention-based VGG-style network for COVID-19 (AVNC) [20] | 15-ways multiple ways on original and Mirror image | Normal = 284 COVID-19 = 306 |
– | 97.38% | 97.06% |
5 | Truncated VGG-16 [32] | Rotation, shearing, zooming, and brightness | Normal = 358 COVID = 344 |
95.7% | 96.3% | 94.8% |
6 | ResNet+DWT (Discrete Wavelet Transform) [33] | Scaling (0.85, 1.2), rotation (−20, 20), flipping, shearing (0, 50) and translation (−35, 35) in the and directions | OMNIAHCOV* Normal = 5152 COVID-19 = 6012 SARS-COV-2** Normal = 1230 COVID-19 = 1252 |
99.23%* 99.62%** |
99.27%* 99.54%** |
99.19%* 99.69%** |
In [34], combination of transfer learning and shallow learning-based approach is used to detect the severity of the COVID-19 infection. The proposed method implemented on cloud-based server detects the binary class of COVID-19 infection severity. Hybrid system of the Sugeno fuzzy integral with ensemble of four pre-trained deep learning model (Squeeze Net v1.1, wide ResNet-50-2, VGG-11, and Google Net) is used for binary classification of COVID-19 [35]. In [36], customized DenseNet201-based models are used for COVID-19 diagnosis with an accuracy of 98.18%, a precision of 97.76%, and the specificity of 98.17%. Further, joint classification and regression method is used for binary classification of severity of COVID-19 infection [37]. In [38], fully automatic approach obtained a global accuracy value of 97.06%. In [39], ResNet50 with majority voting-based model is implemented for COVID-19 classification. The fine-tuning based ResNet50 model obtained Area Under the Curve of 0.90 and 3D-ResNet50 model obtained 0.67 of AUC.
In [40], two-stage deep learning based framework is proposed to discriminate between COVID-19 infection from normal pneumonia using chest CT scans. Transfer learning-based approach is implemented on chest X-ray images to detect COVID-19 [41]. For dataset-1, MobileNet+ support vector machine (SVM) classifier obtained a F1-score of 98.5% and DenseNet201+multi-layer perceptron (MLP) obtained a F1-score of 95.6% for dataset-2. The one shot [42] technique with augmentations of the classification branch in the RoI layer obtained sensitivity of 93.55% for COVID and 99.37% of F1-score. In [43], 5656 X-ray images of each healthy and covid patient is classified using VGG network. The proposed model obtains an accuracy of 96.41%, sensitivity of 96.6%, and the specificity of 96.2%. The explainable classification module (ECM) is proposed for COVID-19 classification using chest CT scan with an accuracy of 84.83% [44].
The hybrid (SpaSA and CNN) approach is implemented on chest CT scans of COVID 19 dataset [45]. The SpaSA algorithm is used to optimize the different Convolution Neural Network (CNN) and transfer learning hyperparameters. The proposed work obtained an accuracy of 99.74% on binary class (14,486 images) and 98% accuracy on three class (17,104 images) classification framework. The 2D-Empirical wavelet transform (EWT) along-with pre-trained transfer learning-based approach (DenseNet121) is used for binary classification of chest CT scans [46]. The TL+DenseNet121 approach obtained an accuracy of 85.50%, F1-score of 85.28%, and AUC of 96.60%. In [47], a Parallel Quantum-Inspired Self-supervised Network (PQIS-Net) for automatic segmentation of lung CT images from hybrid dataset. The CNN based model is proposed for multi-class classification of chest CT scans into normal, COVID-19, and pneumonia. The proposed work obtained an average accuracy of 98% [48].
From the state-of-the-art, it can be inferred that chest CT scan-based COVID-19 diagnosis can identify the coronavirus infection with an accuracy of > 90% in comparison to X-rays or antigen-based tests [49]. The COVID-19 infection diagnosis using deep learning-based approach requires a large set of samples for each class and the model needs to be retrained for variation in the sample in a class. Also, the training on a small dataset may lead to overfitting. Further, the disadvantages of the state-of-the- art deep learning-based techniques for COVID-19 infection diagnosis using chest CT scan include (a) multiple levels of severity of the COVID-19 infection are not clear, (b) the COVID-19 infection is difficult to discriminate from certain pneumonia and lung infections, and (c) numerous techniques failed to differentiate the mutations of coronavirus.
3. Dataset
The important rationale of the proposed methodology is the dataset used for COVID-19 infection severity classification. This section discusses the details of the dataset used based on joint analysis of expert radiologist. In the proposed work, the multi-class classification framework for COVID-19 infection is inspired by the Coronavirus disease 2019 (COVID-19) Reporting and Data System (CO-RADS) [50] and CT severity scores. The CO-RADS score range is based on the patches available in the left lung and right lung [51], [52]. The CO-RADS annotation are as follows: (a) Healthy Control (HC) = CO-RADS-1, (b) less infection = CO-RADS-2, (c) Intermediate infection = CO-RADS-3, (d) High abnormalities suspicious of COVID-19 = CO-RADS-4, (e) Very high abnormality of COVID-19 = CO-RADS-5 and (f) PCR+ = CO-RADS-6. The CT severity scores for COVID-19 infection analysis is having global range of 0–5 and global score of 0–25. In the considered MosMed dataset the CT scores are based on affected lung percentage lobes such as (a) 0% = 0 points (CT score 0: CT-0), (b) > 5% = 1 point (CT-1), (c) 5%–25% = 2 points (CT-2), (d) 25%–50% = 3 points (CT-3), (e) 50%–75% = 4 points (CT-4), and (f) > 75% = 5 points (CT-5).
The proposed deep learning-based Siamese networks are evaluated on a chest CT dataset with annotations provided by joint analysis of Expert-1 (E1) and Expert-2 (E2). Fig. 2 put forth the annotation details suggested by the experts for the considered dataset. The considered chest CT scans are provided by medical hospitals in Moscow, Russia for 1110 patients within the age of 18–97 years (duration: 1st of March to the 25th of April 2020) [53]. The chest CT scans with no lung infection are named as zero or healthy control. The chest CT scans are provided in NIFTI format [54]. The SMOTE based slice selection method is used to address the class imbalance issue with a considered dataset [55]. For class zero (healthy control), the first 15 axial CT scan slices are considered. Further for the suspected COVID-19 infected patients, the 15 slices with infection patches are considered by the joint decision of experts, i.e., E1E2. However, only 2 patients’ CT scans are available for class CT-4, thus all the slices are considered.
Fig. 2.
Brief details of the chest CT scan from MosMed datasets used in the proposed work.
4. Methodology
This section discusses the detailed hierarchy of the proposed methodology, i.e., pre-processing of the dataset, brief architecture of pre-trained sub-network, contrastive loss function, and nearest neighbor classifier. The schematic architecture of the proposed methodology is shown in Fig. 3.
Fig. 3.
The schematic representation of the multi-class classification framework for COVID-19 infection severity level diagnosis using chest CT scans.
4.1. Pre-processing
In the considered dataset, chest CT scans are available in NIFTI format with a size of 512 × 512 and an average of 36 slices per patient. The input image dataset pre-processing is required to make the input dataset compatible with the proposed training network. The pre-processing pipeline includes the following steps: selection of axial slices, resizing operation, normalization, and contrast enhancement. To reduce the storage space issue and computational capacity, the input images are adjusted to the size of 256 × 256. Consider the input image dataset is = [(, ),(, ), …, (, )], here, is input images and is respective labels for . The input images with pixels ‘’ and ‘’ are normalized to the range of [0, 1] and represented as in (1).
(1) |
The contrast of the CT scan slices is enhanced using a histogram equalization approach. For histogram equalization for the normalized input image, i.e., , consider ‘ 0’ as its cumulative histogram, ‘’ is the cumulative sum of the histogram. The histogram equalization is based on the transformation to map the grayscale to their new values. The monotonic transformation ‘’ is minimized using ( ()) — for all intensities . Thus, 1, 2, …, are the pre-processed CT scan images with corresponding labels.
4.2. Pre-trained sub-networks
In the P-shot, M-ways, based learning framework, for each infection class, a maximum of 15 images are used for training. Thus, pre-trained deep learning networks trained for 1000 classes classification on ImageNet dataset are considered as the base encoder for the considered dataset [56], [57]. As the depth of layer increases, the number of trainable parameters increases and that leads to complexity for metric-based learning space. Thus, to reduce the time complexity and metric space parameters ResNet-18 based pre-trained deep learning model is used as a base encoder. Fig. 4 put the architecture details of the base encoder network, i.e, ResNet18 for the proposed work. The ResNet18 with skip connection residual network is inspired by the VGG-19 architecture. The customized weight initializer function ‘He’ for CNN layers with ReLU layer is calculated with mean value zero and standard deviation, i.e., is defined using 2. The scale of ReLU layer is represented by ‘’ and ‘’ is the filter size and ‘’ is number of channels.
(2) |
Fig. 4.
The proposed schematic architecture of the deep learning-based sub-network.
The sub-networks in Siamese networks are having two different inputs and prediction scores are obtained based on the similarity of the encoded features as shown in Fig. 5. The Siamese batch pairs are obtained using Algorithm 1. The Euclidean distance is obtained between the obtained embedded vectors and the loss function is used to update the training model.
Fig. 5.
Flowchart of the multi-class classification framework for COVID-19 infection severity level diagnosis using chest CT scans.
4.3. Contrastive loss function
The generalized loss function is used to evaluate the prediction performance of the trained model. The loss value increases if the predicted value ‘’ is different from the true label ‘’. The contrastive loss functions are based on similar and dissimilar pair images. More similar feature embeddings are obtained for the same classes. The Euclidean Distance (ED) metric for the embedded feature values, i.e, and , is obtained using (3). The 0 is an optimum value used to eliminate the undefined Euclidean distance measurement variables and set them to 1e-6.
(3) |
Further, the loss function for similar ‘’ and dissimilar pairs ‘’ is calculated using (4), (5).
(4) |
(5) |
The (m) = 1 if and are similar images, else (m) = 0. The overall contrastive loss, i.e., CL is calculated using (6).
(6) |
4.4. Prototypical nearest neighbor classification
The prototypical classification framework is based on embedded query points and finding the nearest class prototype as shown in Fig. 4. In the proposed P-shot M-ways classification framework, consider the support set with labeled sample = {}, . The prototype of each class is computed in dimensional space such that and embedding function is as defined in (7).
(7) |
The distribution over classes for a query point with learnable parameters using (8). The prototype learning is based on minimizing the probability given in (9).
(8) |
(9) |
5. Experimentation and discussion
5.1. System requirements
The deep learning-based Siamese networks used ‘Adam’ optimizer, the initial learning rate is 1e−4, gradient learning factor set to 0.9, mini-batch size is 4, the activation function is ReLU, output feature vector size is 1024D, and weight initializer is ‘He’. For the pre-trained deep learning-based encoder networks, the bias factor is set to 10. The maximum epoch is set to 50 with an early stopping criterion. The experiment is conducted on MATLAB R2021b software installed on Intel core i7 10th Generation processor (64 bit) with NVIDIA GeForce RTX2700 having 8 GB capacity and 16 GB RAM capacity.
5.2. Performance evaluation
To evaluate the performance of the proposed work, numerous performance evaluation parameters are considered such as Accuracy, sensitivity, specificity, AUC, and F1-score. The generalized confusion matrix A for the multi-class classification framework is given 10.
(10) |
Here, the true-positive is ‘’, the true-negative is ‘’, the false-negative is ‘’, and the false-positive is defined as ‘’. Consider the High-Infection (HI) class label, to evaluate its performance evaluation parameters are obtained using 10–14. Here, , , , and .
(11) |
(12) |
(13) |
(14) |
The mathematical expressions for performance parameters are given in 15–20. The Accuracy parameter for the different classes such as Extremely High Infection (EHI), High Infection (HI),Intermediate-Infection (II), Low-Infection (LI), and Zero (Z) is represented as , respectively. Here, sensitivity is denoted as , specificity is represented as , Precision measurement is denoted as (Pr, Pr, Pr Pr Pr), F1-score is represented as (), and Mathews Correlation Coefficient is expressed as (MCC, MCC, MCC, MCC, MCC).
(15) |
(16) |
(17) |
(18) |
(19) |
(20) |
The global performance evaluation parameters, i.e., macro-accuracy (), macro-sensitivity (), macro-specificity (),macro-Precision (), macro-F1-score (), and macro-MCC () are calculated using 21–26. Table 2 put forth the details of global performance evaluation parameters conducted on 5 random trials conducted on a 15-shot 5-ways multi-class classification framework. The global performance evaluation parameters ensure the stable performance of the deep learning encoder-inspired encoder-based multi-class classification framework. The proposed work on MosMed dataset obtained an average accuracy of 97.48%, specificity of 98.45%, the sensitivity of 94.35%, F1 score of 93.68%, the precision of 93.44%, and MCC of 92.27%.
Table 2.
Global performance evaluation parameters for five trials conducted with the base encoder ResNet18 on the MosMed dataset.
Trial | Performance parameters (%) |
|||||
---|---|---|---|---|---|---|
1 | 97.24 | 93.86 | 98.31 | 92.78 | 93.05 | 91.53 |
2 | 98.07 | 95.66 | 98.83 | 94.90 | 95.10 | 94.03 |
3 | 97.99 | 95.78 | 98.78 | 94.86 | 95.01 | 93.99 |
4 | 97.57 | 94.18 | 98.49 | 93.70 | 93.88 | 92.40 |
5 | 96.57 | 92.28 | 97.88 | 90.96 | 91.37 | 89.40 |
Average | 97.48 | 94.35 | 98.45 | 93.44 | 93.68 | 92.27 |
(21) |
(22) |
(23) |
(24) |
(25) |
(26) |
To evaluate the robustness of the proposed work, the performance parameters are evaluated for the individual classes as well and given in Table 3. It can be inferred that less accuracy and sensitivity is obtained for ‘LI’ class, less specificity and precision is obtained for ‘II’ class. Also, F1-score is 90% and MCC is < 90% for both ‘LI’ and ‘II’ classes.
Table 3.
Performance evaluation parameters as per severity levels of COVID-19 infection in chest CT scans with fine-tuned base encoder ResNet18 encoder on MosMed dataset.
Classes | A | Se | Sp | Pr | F1 | MCC |
---|---|---|---|---|---|---|
Z | 100 | 100 | 100 | 100 | 100 | 100 |
LI | 99.16 | 96.20 | 100 | 100 | 98.04 | 97.54 |
II | 96.23 | 98.81 | 95.69 | 83.00 | 90.22 | 88.42 |
HI | 95.82 | 83.90 | 99.72 | 99.00 | 90.83 | 88.65 |
EHI | 98.74 | 100 | 98.52 | 92.31 | 96.00 | 95.36 |
5.3. Grad-CAM based infection visualization
The Grad-CAM technique is used to visualize the strongest activation features map learned by the proposed model. The Grad-CAM is based on the linear combination of weights of channel activation maps with an applied ReLU layer. It evaluates the network prediction using differential outputs (class scores) of the gradients. The activation function ReLU provides the feature map of the region of interest, i.e., (severity of infection) in the specified class. In a proposed multi-class infection classification task, consider the score of class is represented by ‘’ for output ‘ r ’, and feature maps are denoted by ‘’. The obtained score Map (:, :, ) corresponding to categorical label ‘pc’ is the gradient of the final classification score for the particular class, w.r.t. each feature in the feature layer. The obtained Grad-Map is up-sampled to the size of 224 × 224 size, i.e., input image size. Fig. 6 shows the obtained Grad-MAP overlapped on respective input images to visualize the severity of infection in lungs CT scans of suspected COVID-19 patients.
Fig. 6.
Grad-CAM based severity of infection visualization using features learned by the proposed methodology on MosMed dataset.
5.4. Case study on Indian patients
The chest CT scans of 42 COVID suspected patients are collected in April 2021 for the patient (age 21–85 years). The low dose CT scan measurements are obtained on the multi-detector 16 slice CT Aquilion model from Toshiba. The high voltage circuit generates 60 kW power for a 7.5 MHU X-ray tube. The exposure time is 0.35 s for partial scans and 0.50 s for each full scan measurement. The length of the detector along the -axis is set to 38 mm with 896 channels or columns across 40 rows. The slice thickness range varies from 16 × 0.5 mm to 16 × 1 mm. The average total time taken by the technician for chest CT slices measurements is 10 min for setup creation.
The CT scans are provided in DICOM 3.0 format having size dimensions 512 × 512 and an average of 150 frames are given for each patient. The input images are pre-processed to make them compatible with the proposed methodology. In the Indian dataset, for each patient 15 slices are considered. Further, Fig. 7 presents the performance evaluation details of the proposed work on the Indian dataset. Based on demographic region, an average recovery time period for COVID-19 patients was 14 days which is reduced to 7 days depending on COVID-19 infection severity [58]. In the proposed work, the chest CT scan of the dataset is collected for 30 days. However, the COVID-19 infection severity is detected for chest CT scan of an average 10 day period of COVID-19 detection to quarantine period. The proposed work obtained a precision of 94.0%, specificity of 98.50%, the sensitivity of 94.54%, F1 score of 93.32%, and MCC is 92.08%.
Fig. 7.
Performance evaluation along with infection visualization of the proposed work on follow-up patients from the Indian dataset.
5.5. Ablation study
The proposed methodology is investigated for the following queries: (1) How would different pre-trained CNN models affect the performance of multi-class classification of Siamese based network? (2) What is the impact of variation in P-shots on performance evaluation parameters of ResNet18 base encoder.
Table 4 puts forth the brief details of performance comparative analysis of the Siamese-based sub-networks on the considered dataset. It can be inferred that > 90% accuracy is achieved with the VGG19, DarkNets, and ResNet18. However, DarkNets and VGG-19 have higher trainable parameters that increase the complexity of parameters in metric space for similarity index. Thus, ResNet-18 based pre-trained model is a suitable encoder for the proposed work. The proposed ResNet18 encoder performance is evaluated with variation in N-ways and P-shots-based multi-class classification framework. Table 5 puts forth the details of global performance parameters with 5 classes and a P-shots learning framework. It is concluded that 5 class 15 shot training networks provide the optimum performance on the considered dataset.
Table 4.
Comparative analysis of global accuracy on the training dataset with 5-ways 2-shot Siamese network with different pre-trained encoder networks (Top-3 classification accuracy values are highlighted in bold).
Encoder | Layers | Parameters (in millions) |
Size of the input image | |
---|---|---|---|---|
ResNet18 | 71 | 11.7 | 224 × 224 | 92.75% |
ResNet50 | 177 | 25.6 | 224 × 224 | 88.10% |
ResNet101 | 347 | 44.6 | 224 × 224 | 89.23% |
DarkNet19 | 64 | 20.8 | 256 × 256 | 91.20% |
DarkNet53 | 184 | 41.6 | 256 × 256 | 95.83% |
DenseNet201 | 708 | 20 | 224 × 224 | 86.40% |
VGG19 | 47 | 144 | 224 × 224 | 94.80% |
Inceptionv3 | 315 | 23.9 | 299 × 299 | 85.20% |
SqueezeNet | 68 | 1.24 | 227 × 227 | 84.10% |
GoogleNet | 144 | 7 | 224 × 224 | 87.20% |
Xception | 170 | 22.9 | 299 × 299 | 77.78% |
Table 5.
Comparative analysis of the P-ways M-shot Siamese network ResNet18 based encoder for multi-class classification of COVID-19 infection severity levels.
Techniques |
Performance parameters (%) |
||||||
---|---|---|---|---|---|---|---|
P | M | ||||||
5 | 15 | 95.68 | 91.40 | 88.64 | 88.60 | 89.04 | 87.06 |
5 | 15 | 97.07 | 94.58 | 98.27 | 92.70 | 92.75 | 91.59 |
5 | 15 | 98.07 | 95.66 | 98.83 | 94.90 | 95.10 | 94.03 |
Table 6 put forth the comparative analysis of the proposed methodology with other existing techniques available in literature on the considered MosMed dataset with 5 ways and 15 shots classification framework. The proposed methodology detects the numerous severities of COVID-19 infection with an accuracy of 98.07% in 0.8 seconds per chest CT scan. The few-shot learning with ResNet50 pre-trained network without augmentation obtained the classification accuracy of 87.23% on MosMed dataset in 15 seconds duration per scan [20]. The COVID-19 infection severity is classified with an accuracy of 96.20% in 12 s per scan using DenseNet [59].
Table 6.
Comparative analysis of the proposed techniques with the exiting state-of-the-art methods on considered MosMed dataset for multi-class classification (5 ways 15 shot).
5.6. Limitations and future scope
Earlier we have worked on binary classification of chest CT scans of COVID19 patients from normal patients using deep learning-based techniques [63], [32]. For this, numerous statistical data augmentation techniques are implemented. But the earlier reported work is not able to detect the COVID-19 infection severity on smaller dataset. The proposed McS-Nets addresses the issue of COVID-19 infection severity classification with limited data availability in each class. It provides the accurate prediction of the disease in mean time with the minimum number of training samples. However, the proposed work is having certain limitations associated with it: (a) performance parameters are < 90% for the infection level with less than 25% infection or few GGO is visible, (b) it is costly as multiple CT scans are required for follow-up patients (average 2–3 scans during 14 days of tenure of COVID-19 infection first detected).
In future work, the proposed technique is to be implemented in a real-time environment for COVID-19 infection diagnosis. The proposed framework is based on image similarity; thus, techniques need to be explored on various variants of COVID-19 infection. Further, the proposed methodology will be implemented on the following conditions: (a) different variants of COVID-19 infections, (b) blood samples and other clinical parameters will also be used for COVID-19 infection identification. To make the proposed technique more effective, contrastive learning-based segmentation is to be explored. Also, a cloud computing-based framework is to be used for wider clinical applications.
6. Conclusions
The proposed work addresses the COVID-19 severity infection diagnosis using chest CT scans with limited samples available from each class. The COVID-19 infection in lungs is categorized as: 1. Zero — 0% infection (CT-0), 2. Low — < 25% infection (CT-1), 3. Intermediate — 25%–50% infection (CT-2), 4. High — 50%–75% infection (CT-3), 5. Extremely high infection — > 75% infection (CT-4). The proposed work learned different severity of infection features in the lung area in comparison to the best performing deep learning-based encoder network. In the proposed 15-shot 5-ways McS-Net an average of > 95% of accuracy, sensitivity, and specificity is obtained on testing data. The Grad-CAM-based functionality is used to investigate the infection map learned by the proposed networks. Further, the proposed work is evaluated on the Indian dataset with limited CT scan slices in each class and obtains average accuracy of 97.06%, a sensitivity of 94.54%. Thus, the proposed technique is useful in the infection severity identification on the follow-up low dose chest CT scan of the patients.
CRediT authorship contribution statement
Sakshi Ahuja: Conceptualization, Methodology, Software, Writing – original draft. Bijaya Ketan Panigrahi: Project administration, Supervision, Investigation, Visualization. Nilanjan Dey: Project administration, Methodology, Supervision, Investigation, Visualization. Arpit Taneja: Project administration, Medical assistance, Visualization. Tapan Kumar Gandhi: Supervision, Formal analysis, Investigation, Validation.
Declaration of Competing Interest
The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.
References
- 1.WHO, 2021. URL http://www.emro.who.int/health-topics/corona-virus/about-covid-19.html. (Accessed 26 December 2021).
- 2.W. H. Organization, 2021. URL www.who.int/emergencies/diseases/novel-coronavirus-2019. (Accessed 26 December 2021).
- 3.Nature, 2021. URL www.nature.com/articles/d41586-020-00502-w. (Accessed 26 December 2021).
- 4.Outlook, 2021. URL www.outlookindia.com/website/story/india-news-explained-how-many-mutant-coronavirus/380587. (Accessed 8 December 2021).
- 5.Callaway E. The coronavirus is mutating—Does it matter? Nature. 2020;585:174–177. doi: 10.1038/d41586-020-02544-6. [DOI] [PubMed] [Google Scholar]
- 6.Covid19india, 2021. URL www.covid19india.org/. (Accessed 26 December 2021).
- 7.ICMR, 2021. URL https://www.icmr.gov.in/index.html. (Accessed 8 December 2021).
- 8.Zhang Z. Insight into the practical performance of RT-PCR testing for SARS-COV-2 using serological data: A cohort study. Lancet Microbe. 2021;2:79–87. doi: 10.1016/S2666-5247(20)30200-7. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 9.M. Manych, 2021. URL http://www.siemens-healthineers.com/en-uk/news/mso-x-ray-imaging-for-covid-19.html. (Accessed 26 December 2021).
- 10.Ferrag M.A., Shu L., Choo K.-K.R. Fighting COVID-19 and future pandemics with the internet of things: Security and privacy perspectives. IEEE/CAA J. Autom. Sin. 2021;8(9):1477–1499. doi: 10.1109/JAS.2021.1004087. [DOI] [Google Scholar]
- 11.Fong S.e.a. Springer; 2022. Probabilistic Forecasting Model for the COVID-19 Pandemic Based on the Composite Monte Carlo Model Integrated with Deep Learning and Fuzzy System. [DOI] [Google Scholar]
- 12.Wu W., Shi J., Yu H., Wu W., Vardhanabhuti V. Tensor gradient LO norm minimization- based low-dose CT and its application to COVID-19. IEEE Trans. Instrum. Meas. 2021;70:1–12. doi: 10.1109/TIM.2021.3050190. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 13.Sharma R., Kumar M., Maheshwari S., Ray K. Evdhm-arima-based time series forecasting model and its application for COVID-19 cases. IEEE Trans. Instrum. Meas. 2021;70:1–10. doi: 10.1109/TIM.2020.3041833. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14.S. J. F, et al. Composite Monte Carlo decision-making under high uncertainty of novel coronavirus epidemic using hybridized deep learning and fuzzy rule induction. Appl. Soft Comput. 2020;93:1–14. doi: 10.1016/j.asoc.2020.106282. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15.L. Y, et al. A wireless, low-power, and miniaturized eit system for remote and long-term monitoring of lung ventilation in the isolation ward of ICU. IEEE Trans. Instrum. Meas. 2021;70:1–11. doi: 10.1109/TIM.2021.3085970. [DOI] [Google Scholar]
- 16.Wang Y., Jacobsen E.W. Inferring individual network edges - with application to target identification in gene networks. IFAC-PapersOnLine. 2020;53:63–68. doi: 10.1016/j.ifacol.2020.12.050. [DOI] [Google Scholar]
- 17.Wu H., Xu Z., Zhang J., Yan W., Ma X. 10th International Congress on Image and Signal Processing. IEEE; 2017. Face recognition based on convolution siamese networks; pp. 1–5. [DOI] [Google Scholar]
- 18.Dey S., Dutta A., Toledo J.I., Ghosh S.K., Llados J., Pal U. 2017. SigNet: Convolutional siamese network for writer independent offline signature verification; pp. 1–7. arXiv:1707.02131v2. [Google Scholar]
- 19.Wang B., Wang D. Plant leaves classification: A few-shot learning method based on siamese network. IEEE Access. 2019;7:151754–151763. doi: 10.1109/ACCESS.2019.2947510. [DOI] [Google Scholar]
- 20.Chen X., Yao L., Zhou T., Dong J., Zhang Y. Momentum contrastive learning for few- shot COVID-19 diagnosis from chest CT images. Pattern Recognit. 2021;113:1–8. doi: 10.1016/j.patcog.2021.107826. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 21.Paul A., Tang Y.-X., Shen T.C., Summers R.M. Discriminative ensemble learning for few- shot chest X-ray diagnosis. Med. Image Anal. 2021;68:1–16. doi: 10.1016/j.media.2020.101911. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 22.Wang J., Fang Z., Lang N., Yuan H., Su M.-Y., Baldi P. A multi-resolution approach for spinal metastasis detection using deep siamese neural networks. Comput. Biol. Med. 2017;84:137–146. doi: 10.1016/j.compbiomed.2017.03.024. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 23.Shorfuzzamana M., Hossain M.S. MetaCOVID: A siamese neural network framework with contrastive loss for n-shot diagnosis of COVID-19 patients. Pattern Recognit. 2020;113:1–12. doi: 10.1016/j.patcog.2020.107700. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 24.S. T, et al. COVIDGR dataset and COVID-SDNet methodology for predicting COVID-19 based on chest X-ray images. IEEE J. Biomed. Health Inf. 2020;24:3595–3605. doi: 10.1109/JBHI.2020.3037127. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 25.Karmore S., Bodhe R., Al-Turjman F., Kumar R.L., Pillai S. IoT based humanoid software for identification and diagnosis of COVID-19 suspects. IEEE Sens. J. 2020:1–8. doi: 10.1109/JSEN.2020.3030905. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 26.Singh G., Yow K.C. An interpretable deep learning model for COVID-19 detection with chest X-ray images. IEEE Access. 2021;9:85198–85208. doi: 10.1109/ACCESS.2021.3087583. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 27.Zhou C., Song J., Zhou S., Zhang Z., Xing J. COVID-19 detection based on image regrouping and ResNet-SVM using chest X-ray images. IEEE Access. 2021;9:81902–81912. doi: 10.1109/ACCESS.2021.3086229. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 28.Carvalho E.D., Silva R.R., Araújo F.H., de A.L. Rabelo R., de Carvalho Filho A.O. An approach to the classification of COVID-19 based on CT scans using convolutional features and genetic algorithms. Comput. Biol. Med. 2021;136:1–11. doi: 10.1016/j.compbiomed.2021.104744. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 29.Zhang Y., Satapathy S., Zhu L.Y., Górriz J., Wang S. A seven-layer convolutional neural network for chest CT based COVID-19 diagnosis using stochastic pooling. IEEE Sens. J. 2020:1–8. doi: 10.1109/JSEN.2020.3025855. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 30.Y. W, et al. JCS: An explainable COVID-19 diagnosis system by joint classification and segmentation. IEEE Trans. Image Process. 2021;30:3113–3126. doi: 10.1109/TIP.2021.3058783. [DOI] [PubMed] [Google Scholar]
- 31.Wang S., Fernandes S., Zhu Z., Zhang Y. AVNC: Attention-based VGG-style network for COVID-19 diagnosis by CBAM. IEEE Sens. J. 2021:1–8. doi: 10.1109/JSEN.2021.3062442. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 32.M. S, et al. Transfer learning–based ensemble support vector machine model for automated COVID-19 detection using lung computerized tomography scan data. Med. Biol. Eng. Comput. 2021;59:825–839. doi: 10.1007/s11517-020-02299-2. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 33.Attallah O., Samir A. A wavelet-based deep learning pipeline for efficient COVID-19 diagnosis via CT slices. Appl. Soft Comput. 2022;128:1–16. doi: 10.1016/j.asoc.2022.109401. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 34.Abbasi W.A., Abbas S.A., Andleeb S., ul Islam G., Ajaz S.A., et al. COVIDC: An expert system to diagnose COVID-19 and predict its severity using chest CT scans: Application in radiology. Inform. Med. Unlocked. 2021;23 doi: 10.1016/j.imu.2021.100540. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 35.Kundu R., Singh P.K., Mirjalili S., Sarkar R. COVID-19 detection from lung CT-scans using a fuzzy integral-based cnn ensemble. Comput. Biol. Med. 2021;138:1–14. doi: 10.1016/j.compbiomed.2021.104895. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 36.Lahsaini I., Daho M.E.H., Chikh M.A. Deep transfer learning based classification model for COVID-19 using chest CT-scans. Pattern Recognit. Lett. 2021;152:122–128. doi: 10.1016/j.patrec.2021.08.035. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 37.Zhu X., Song B., Shi F., Chen Y., Hu R., et al. Joint prediction and time estimation of COVID-19 developing severe symptoms using chest CT scan. Med. Image Anal. 2021;67:1–12. doi: 10.1016/j.media.2020.101824. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 38.de Moura J., Novo J., Ortega M. Fully automatic deep convolutional approaches for the analysis of COVID-19 using chest X-ray images. Appl. Soft Comput. 2022;115 doi: 10.1016/j.asoc.2021.108190. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 39.Serte S., Demirel H. Deep learning for diagnosis of COVID-19 using 3D CT scans. Comput. Biol. Med. 2021;132:1–8. doi: 10.1016/j.compbiomed.2021.104306. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 40.Abdel-Basset M., Hawash H., Moustafa N., Elkomy O.M. Two-stage deep learning framework for discrimination between COVID-19 and community-acquired pneumonia from chest CT scans. Pattern Recognit. Lett. 2021;152:311–319. doi: 10.1016/j.patrec.2021.10. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 41.Ohata E.F., Bezerra G.M., Chagas J.V.S.d., Lira Neto A.V., Albuquerque A.B., Albuquerque V.H.C.d., Reboucas Filho P.P. Automatic detection of COVID-19 infection using chest X-ray images through transfer learning. IEEE/CAA J. Autom. Sin. 2021;8(1):239–248. doi: 10.1109/JAS.2020.1003393. [DOI] [Google Scholar]
- 42.Ter-Sarkisov A. One shot model for the prediction of COVID-19 and lesions segmentation in chest CT scans through the affinity among lesion mask features. Appl. Soft Comput. 2022;116 doi: 10.1016/j.asoc.2021.108261. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 43.Elazab A., Elfattah M.A., Zhang Y. Novel multi-site graph convolutional network with supervision mechanism for COVID-19 diagnosis from X-ray radiographs. Appl. Soft Comput. 2022;114 doi: 10.1016/j.asoc.2021.108041. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 44.Ye Q., Gao Y., Ding W., Niu Z., Wang C., Jiang Y., Wang M., Fang E.F., Menpes-Smith W., Xia J., Yang G. Robust weakly supervised learning for COVID-19 recognition using multi-center CT images. Appl. Soft Comput. 2022;116 doi: 10.1016/j.asoc.2021.108291. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 45.Baghdadi N.A., Malki A., Abdelaliem S.F., Balaha H.M., Badawy M., Elhosseini M. An automated diagnosis and classification of COVID-19 from chest CT images using a transfer learning-based convolutional neural network. Comput. Biol. Med. 2022;144 doi: 10.1016/j.compbiomed.2022.105383. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 46.Gaur P., Malaviya V., Gupta A., Bhatia G., Pachori R.B., Sharma D. COVID-19 disease identification from chest CT images using empirical wavelet transformation and transfer learning. Biomed. Signal Process. Control. 2022;71 doi: 10.1016/j.bspc.2021.103076. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 47.Konar D., Panigrahi B.K., Bhattacharyya S., Dey N., Jiang R. Auto-diagnosis of COVID-19 using lung CT images with semi-supervised shallow learning network. IEEE Access. 2021;9:28716–28728. doi: 10.1109/ACCESS.2021.3058854. [DOI] [Google Scholar]
- 48.Hasija S., Akash P., Hemanth M.B., Kumar A., Sharma S. A novel approach for detection of COVID-19 and pneumonia using only binary classification from chest CT-scans. Neurosci. Inform. 2022;2 doi: 10.1016/j.neuri.2022.100069. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 49.Joshi A., Dey N., Santosh K.C. Springer; Singapore: 2020. Intelligent Systems and Methods to Com- Bat COVID-19; pp. 2191–5318. [DOI] [Google Scholar]
- 50.M. P, et al. Co-rads: A categorical ct assessment scheme for patients suspected of having COVID-19—definition and evaluation. Work. Group Dutch Radiol. Soc. 2020;296:97–104. doi: 10.1148/radiol.2020201473. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 51.H. W, et al. Frequency and distribution of chest radiographic findings in COVID-19 positive patients. Radiology. 2020;296:72–78. doi: 10.1148/radiol.2020201160. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 52.M. A. W., et al. Severity scoring of lung oedema on the chest radiograph is associated with clinical outcomes in ards. Thorax. 2018;73:840–846. doi: 10.1136/thoraxjnl-2017-211280. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 53.S. M., et al. 2017. Mosmeddata: Chest CT scans with COVID-19 related findings dataset; pp. 1–4. arXiv:2005.06465v1. [Google Scholar]
- 54.Larobina M., Murino M.L. Medical image file formats. J. Digit. Imaging. 2014;27:200–206. doi: 10.1007/s10278-013-9657-9. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 55.Fotouhi S., Asadi S., Kattan M.W. A comprehensive data level analysis for cancer diagnosis on imbalanced data. J. Biomed. Inform. 2019;90 doi: 10.1016/j.jbi.2018.12.003. [DOI] [PubMed] [Google Scholar]
- 56.Russakovsky O., Deng J., H.S, et al. Imagenet large scale visual recognition challenge. Int. J. Comput. Vis. 2015;115:211–252. [Google Scholar]
- 57.ImageNet, 2017. URL image-net.org/download.php. (Accessed 26 December 2021).
- 58.Bhapkar H.R., Mahalle P.N., Dey N., Santosh K.C. Revisited COVID-19 mortality and recovery rates: Are we missing recovery time period? J. Med. Syst. 2020;44:1–5. doi: 10.1007/s10916-020-01668-6. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 59.Y. Q, et al. Detection and severity classification of COVID-19 in CT images using deep learning. Diagnostics. 2021;11 doi: 10.3390/diagnostics11050893. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 60.Rohila V.S., Gupta N., Kaul A., Sharma D.K. Deep learning assisted COVID-19 detection using full CT-scans. Internet of Things. 2021;14 doi: 10.1016/j.iot.2021.100377. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 61.Lee E., Zheng J., Colak E., et al. Deep COVID detect: An international experience on COVID-19 lung detection and prognosis using chest CT. NPJ Digit. Med. 2021:1–11. doi: 10.1038/s41746-020-00369-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 62.Amini N., Shalbaf A. Automatic classification of severity of COVID-19 patients using texture feature and random forest based on computed tomography images. Int. J. Imaging Syst. Technol. 2022;32:102–110. doi: 10.1002/ima.22679. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 63.S. A, et al. Deep transfer learning-based automated detection of COVID-19 from lung CT scan slices. Appl. Intell. 2020;51:571–585. doi: 10.1007/s10489-020-01826-w. [DOI] [PMC free article] [PubMed] [Google Scholar]