Abstract
COVID-19 is a virus that causes upper respiratory tract and lung infections. The number of cases and deaths increased daily during the pandemic. Once it is vital to diagnose such a disease in a timely manner, the researchers have focused on computer-aided diagnosis systems. Chest X-rays have helped monitor various lung diseases consisting COVID-19. In this study, we proposed a deep transfer learning approach with novel fine-tuning mechanisms to classify COVID-19 from chest X-ray images. We presented one classical and two new fine-tuning mechanisms to increase the model’s performance. Two publicly available databases were combined and used for the study, which included 3616 COVID-19 and 1576 normal (healthy) and 4265 pneumonia X-ray images. The models achieved average accuracy rates of 95.62%, 96.10%, and 97.61%, respectively, for 3-class cases with fivefold cross-validation. Numerical results show that the third model reduced 81.92% of the total fine-tuning operations and achieved better results. The proposed approach is quite efficient compared with other state-of-the-art methods of detecting COVID-19.
Keywords: COVID-19 disease detection, Deep transfer learning, CNN, Fine-tuning, MobileNet
Introduction
COVID-19 (COronaVIrus Disease 2019-COVID-19) is a contagious disease caused by coronavirus 2 (SARS-CoV-2) that causes severe acute respiratory syndrome (Ai et al. 2020). The World Health Organization (WHO) officially reported the disease on March 11, 2020, declaring it a pandemic affecting more than 227 countries (https://www.euro.who.int/en/health-topics/health-emergencies/coronavirus-968covid19/news/news/2020/3/who-announces-covid-19-outbreak-969a-pandemic). COVID-19 has infected more than 645,630, 482 people worldwide and caused more than 6,634,816 deaths, as reported on December 14, 2022 (https://covid19.who.int/). The first clinical manifestation of SARS-CoV-2-associated COVID-19 disease in which cases were detected was pneumonia (Seibert et al. 2020). In symptomatic patients, clinical signs of the disease, such as cough, pyrexia, nasal obstruction, fatigue, and upper respiratory tract infections, usually begin in less than a week. However, the disease is fatal, with about 75% of patients experiencing severe chest symptoms consistent with pneumonia, as seen on computed tomography during hospitalization (Feng et al. 2020).
The main factors for the rapid spread of the disease are the low efficacy of vaccines with different mutations and inadequate medical testing (Kerr et al. 2021). Therefore, rapid detection and isolation of positive cases are critical to controlling the spread. Given the increase in COVID-19 cases worldwide, more attention must be paid to the containment of the outbreak. The use of reverse transcription-polymerase chain reaction (RT-PCR) is the most commonly used method to detect COVID-19 virus (Susanto et al. 2022). Rapid and accurate detection of the disease is critical to control the epidemic.
An alternative method of diagnosing COVID-19 is the analysis of chest X-rays and CT scans. The equipment required for these images is available worldwide and has been used for years to detect viral infections such as cough, fever, and shortness of breath. In addition, some classification methods have been used to detect lung abnormalities (Chaunzwa et al. 2021), such as tuberculosis, cancer, and pneumonia, that may be related to COVID-19 in chest X-rays in previous studies using deep learning (DL) methodologies. The importance of medical imaging is recognized as a decisive source of information to enable rapid diagnosis (Song et al. 2021). COVID-19 and the combination of AI and chest imaging can help describe and detect COVID-19 complications (Afshar-Oromieh et al. 2021). X-ray images play a critical role in the rapid clinical evaluation of COVID-19 (Wong et al. 2020).
Recently, medical imaging systems have used DL techniques and helped medical personnel make rapid diagnoses (Suzuki 2017). Researchers used convolutional neural network (CNN) models, one of the most popular designs, to detect disease on chest X-ray images (Maior et al. 2021). It is reported that classical computer imaging systems cannot solve complex problems such as image identification, organ recognition, bacterial colony classification, and disease identification (Abiyev and Ma’aitah 2018). However, computer imaging systems that use DL models can solve such problems with a high success rate. Deep learning approaches require powerful GPU systems and a large number of training examples to train the models. To overcome this challenge, transfer learning can be applied to pre-trained CNN models for small datasets by re-designing the last few layers of the model and fine-tuning the model. The CNN model can achieve high accuracy if appropriate hyperparameters are adjusted and efficient fine-tuning approaches are used.
CNNs, which compute weights and extract features during training, were developed to handle multidimensional data such as time series or image data. The name “convolutional” comes from using a convolution operator to solve complex operations. In addition, CNNs are commonly used in health care due to their ability to generate features from images automatically. CNNs can also learn from one task and transfer it to another through transfer learning and fine-tuning. This method has proven successful in classification tasks (Tajbakhsh et al. 2016).
In the literature, many deep learning models have been reported to detect and classify COVID-19 cases from chest X-ray images (Shorfuzzaman and Masud 2020; Degadwala et al. 2021; Iskanderani et al. 2021; Pathak et al. 2020; Kaur et al. 2021; Wang et al. 2020; Misra et al. 2020; Xu et al. 2020; Ozturk et al. 2020; Asif and Wenhui 2020; Varela-Santos and Melin 2021; Boccaletti et al. 2020; Sun and Wang 2020). These studies include different CNN models such as Resnet50, VGG16 (Shorfuzzaman and Masud 2020), MobileNetV2 (Apostolopoulos and Mpesiana 2020), DenseNet (Iskanderani et al. 2021), and several customized new CNN models such as DarkCovidNet (Ozturk et al. 2020) , InstaCovidNet (Gupta et al. 2021)).
However, one of the major problems in studying COVID-19 is the lack of sufficient and reliable data in the first quarter of the pandemic. Because of the limited number of tests, many deaths and viral infections cannot be detected during the beginning of the pandemic. The accessibility of large and high-quality datasets in the ML and DL applications plays a critical role in the reliability of the results. Therefore, researchers had to work with datasets in a limited range by combining available datasets or using data augmentation techniques. As the pandemic progressed, larger data sets were made available to the public (Mahajan et al. 2022). Thus, the most recent studies have used large datasets and ensured the accuracy of their models (Tammina 2022; Bhattacharyya et al. 2022).
In our study, we tested five different pre-trained CNN architectures, namely VGG16, DenseNet, MobileNetV2, InceptionV3, and ResNet, to find a suitable model for our problem. We trained these models in 100 epochs using 80% of the images, and the results show that MobileNetV2 is more successful. Thus, a deep transfer learning approach based on MobileNetV2 with novel fine-tuning mechanisms is proposed to detect COVID-19 diseases from chest X-ray images. We created a relatively large dataset by combining two publicly available datasets containing COVID-19 and pneumonia X-ray images, namely chest X-ray images (Pneumonia) (ChestX-RayImagesPneumonia 2022) and COVID-19 Radiography Database (Covid-19RadiographyDatabase 2022), to evaluate the proposed model. The dataset consisted of three classes: normal, pneumonia, and COVID-19.
Our approach integrates the training phases of the MobileNetV2 CNN architecture, such as data augmentation, transfer learning, and fine-tuning, into a single design. We proposed one classical and two new fine-tuning mechanisms to improve model performance. We also used data augmentation methods such as zooming and rotation to preserve the diversity of the test data. Our theory is that during the training of the network, the last layers are trained more, which increases the ability to generalize to the new problem while at the same time hiding the information learned from the previous training of the first layers, leading to more successful results. In this study, three methods were used for comparative theory implementation. In the fine-tuning phase, the first approach was classical by opening the last 50 layers and freezing the 102 remaining layers of the MobileNetV2 model. The second approach used a step function to fine-tune the model. The third approach used a predefined exponential mathematical function to fine-tune the model. We performed experimental tests with three models and compared the numerical results. The numerical results show that the proposed approach achieves promising results when using raw data without complex preprocessing steps.
The main contributions of this study can be summarized as follows:
A new deep transfer learning model was proposed based on MobileNetV2 to detect COVID-19 disease.
Two novel fine-tuning mechanisms were integrated into the proposed model.
We combined two public COVID-19 X-ray databases to evaluate our model and trained our model on this dataset using two standard training approaches, 30–70% hold-out with five different seeds and fivefold cross-validation.
Proposed model reduced 81.92% of total fine-tuning operations and achieved better results.
The remainder of the article is organized as follows. Section 2 contains a literature review. Section 3 contains information about the datasets, the proposed CNN architecture, and its sublevels. Section 4 contains the numerical results obtained from the proposed approach. Section 5 discusses the proposed approach with the works in the literature. Finally, Section 6 presents the conclusions and future work of the paper.
Related works
Deep learning (DL) algorithms and CNN models in image analysis and processing in the biomedical field have recently produced successful results (Bala et al. 2019). Moreover, various CNN-based deep neural networks achieved remarkable results in ImageNet competition (Krizhevsky et al. 2012). Thus, the researchers have focused on using DL to analyze medical images.
Diagnosis of COVID-19 by chest X-ray is associated with pneumonia symptoms. The image classification methods developed by researchers for COVID-19 or pneumonia are divided into the following categories: Machine learning methods (ML) (Kwekha-Rashid et al. 2021), CNN architectures (Minaee et al. 2020; Wang et al. 2020; Varshni et al. 2019), complex CNN models (Wang et al. 2020; Monshi et al. 2021; Khan et al. 2020), adversarial networks (Shams et al. 2020), deep fuzzy model (Song et al. 2022), and transfer learning (TL) methods (Shorfuzzaman and Masud 2020; Pathak et al. 2020; Apostolopoulos and Mpesiana 2020).
More specifically, Stephen et al. developed a DL model to identify pneumonia in chest X-ray images (Stephen et al. 2019). They used a ConvNet model to compute features from images and used these features to detect pneumonia. The dataset consists of 36 validation and 64 training data. Data augmentation and hyperparameter fine-tuning were performed to improve the model. The result was an accuracy rate of 93.73% on a small dataset. In another study, Islam et al. proposed a general model for computer-aided detection of pneumonia on chest X-ray images on a dataset comprising 5863 images of normal and pneumonia patients. Their model classified pneumonia with 97.34% accuracy (Islam et al. 2019). In Ayan et al. (2019), the authors aimed to detect pneumonia from chest X-ray images based on the VGG16 model. They used a pre-trained model and achieved an accuracy rate of 82%. Varshni et al. applied deep learning models such as ResNet, VGG16, and Xception, extracted some features and used these features to classify pneumonia on chest X-rays using Random Forest, SVM, Naive Bayes classifiers, and K-Nearest Neighbors (Varshni et al. 2019). In addition, they used hyperparameter optimization in their model to improve the results.
Concerning the COVID-19 studies, Wang et al. used a DL approach to compute features and achieved a correct detection rate of 79.3% on a database consisting of 325 COVID-19 and 740 pneumonia images (Wang et al. 2021). Apostolopoulos et al. proposed a model to distinguish between normal, viral, and COVID-19 cases. They implemented a model using TL techniques and achieved success rates of 98.66%, 96.46%, and 96.78%, respectively (Apostolopoulos and Mpesiana 2020). In Ucar and Korkmaz (2020), the authors used a DL model to categorize coronavirus-related infections based on X-ray images. They applied Bayesian optimization-based fine-tuning to their model to calculate their results (Ucar and Korkmaz 2020).
In Nour et al. (2020), the authors also proposed a classification model using a Bayesian optimization algorithm and performing TL. In another study implementing a 3-class COVID-19 detection model, the authors used CT images for disease detection and achieved a success rate of 86.7% with a model segmentation-based approach (Xu et al. 2020). Khan et al. integrated pre-trained Xception into a model for automatically classifying COVID-19 on X-rays. Their study achieved an accuracy rate of 87.5% (Khan et al. 2020). In Rehman et al. (2020), the authors developed a DL approach to implement a CNN model to distinguish viral pneumonia, bacterial pneumonia, and normal images. They also used their model to diagnose COVID-19 and achieved a correct classification rate of 98.75%. Narin et al. proposed a model employing five pre-trained CNN models, namely InceptionV3, ResNet50, ResNet101, ResNet152, and Inception-ResNetV2, on chest X-ray images. The proposed approach achieved an accuracy rate of 98.00% (Narin et al. 2021). In Toğaçar et al. (2020), the authors proposed a MobileNetV2-based DL approach to detect COVID-19 infection from X-ray images. They used stacked fuzzy colored and original images to increase their model’s success and achieved an accuracy rate of 97.06%. They also employed the Social Mimic Optimization algorithm to optimize their parameters. Gupta et al. used several pre-trained CNN models to identify COVID-19 and pneumonia by extracting features from healthy chest X-rays. The model had a 99.08% success rate (Gupta et al. 2021). In another study using CNN to calculate features, researchers developed a model that combined two CNN models to classify chest X-ray images and attained a classification success rate of 91.40% (Rahimzadeh and Attar 2020).
Luz et al. presented an EfficientNet-based model to analyze X-ray images. They trained their model on 183 COVID-19 samples and achieved an accuracy rate of 93.9% (Luz et al. 2021). Punn and Agarwal introduced a DNN for coronavirus symptom detection. They used 108 COVID-19 cases in the database and achieved an average classification rate of 97% (Punn and Agarwal 2021).
Feature extraction and deep network design used for diagnosing COVID-19 infection are very effective in the results. Previous studies show that TL and fine-tuning models of DL architectures are widely used to detect diseases such as COVID-19 and pneumonia. Most of the work has been done using data augmentation operations such as image enlargement and resizing. Some reported approaches used hybrid models that combine classical machine learning algorithms with DL approaches to detect coronavirus-associated infections on chest images.
Material and methods
This section briefly discusses datasets, data augmentation, preprocessing, transfer learning, and fine-tuning approaches. In addition, the proposed DL methods are described in detail. A general block diagram of the study is shown in Fig. 1.
Fig. 1.
Block Diagram of proposed approach
We used pre-trained CNN models, VGG16, InceptionV3, MobileNetV2, ResNet50, DenseNet, and EfficientNet, to find the most appropriate design for our problem in this study. These models were trained on 80% of the entire dataset for 100 epochs. Table 1 shows the results of this test. As can be seen from the table, MobileNetV2 was the most effective architecture. Therefore, we determined MobileNetV2 as the main design.
Table 1.
Achievements of pre-trained deep CNN models for COVID-19 disease classification before any fine-tuning
| Model | Training loss | Training acc. | Val. loss | Val. acc. |
|---|---|---|---|---|
| VGG16 | 0.9279 | 0.5938 | 0.9064 | 0.6250 |
| InceptionV3 | 0.0857 | 0.9722 | 0.0730 | 0.9375 |
| MobileNetV2 | 0.0970 | 0.9657 | 0.1183 | 0.9549 |
| ResNet50 | 0.0445 | 0.9826 | 0.2045 | 0.9488 |
| DenseNet | 0.0598 | 0.9896 | 0.1710 | 0.9375 |
| EfficientNet | 0.3335 | 0.9890 | 0.8310 | 0.8346 |
Bold text represents the best validation accuracy
Datasets
The main challenge in training and validating the proposed model is the lack of publicly available labeled datasets. Most studies used small datasets (Shorfuzzaman and Masud 2020; Kaur et al. 2021; Rehman et al. 2020), which does not guarantee that the proposed model is fully trained. Thus, to increase the proposed model’s generalization ability and create a more robust system, it is necessary to work with a larger database. Most of the public COVID-19 datasets have two classes. We combined the two publicly available datasets to increase the number of images and added pneumonia as a third class in this work. The first dataset, the chest X-ray dataset, contains images in two classes (pneumonia and normal) and consists of 5,863 X-ray images (ChestX-RayImagesPneumonia 2022). The second dataset, the COVID-19 Radiography Database, contains normal, viral pneumonia, and COVID-19 images. The dataset consists of 1341 normal, 219 COVID-19, and 1345 viral pneumonia chest X-ray images (Covid-19RadiographyDatabase 2022). Figure 2 shows some sample images from these datasets.
Fig. 2.
Sample images of our database
Data augmentation and preprocessing
Deep CNNs have performed remarkably well on many image classification tasks; however, these models rely on large datasets to avoid over-fitting (Shorten and Khoshgoftaar 2019). We need a large dataset to develop a robust and successful DL classification model, and this is not always possible. Data augmentation methods are used, a set of techniques that improve the size of training datasets. It also provides a diversity of classes in the dataset. Data augmentation was applied to the proposed model to increase the amount of data, avoid over-fitting, and create a more robust model. These methods generate new images by interacting with different variations in the visual properties of images that can significantly change them, such as image rotation, random horizontal flipping, and zooming. The dataset contains chest X-ray images, some of which are 1007x1024 pixels. We need to reduce the size of the images to fit the input shape of the tested models. The input shape of the MobilenetV2 is 224x224 pixels. This study only used the resize operator as a preprocessing process.
MobileNetV2 deep learning model
MobileNetV2 is a CNN architecture that aims to perform well on mobile devices. It is based on a different structure from other CNN structures where its connections are between the bottleneck layers. In addition, the intermediate expansion layer uses deep folds to filter out nonlinear features. The MobileNetV2 architecture includes 32 layers of initial convolution followed by 19 bottleneck layers. The design details of the MobileNetV2 model are shown in Figure 3. In this study, we propose a MobileNetV2 design that uses two novel fine-tuning approaches to identify COVID-19 X-ray images.
Fig. 3.
MobileNetv2 architecture
MobileNetV2 design has some advantages over the other DL designs. For small datasets, it is hard to train the system and the visual classification task becomes prone to over-fitting. The MobileNetV2 architecture hinders this effect, preventing over-fitting, and it is a fast and successful architecture that optimizes memory consumption with a low margin of error. Moreover, MobileNetV2 design provides fast execution of transactions while making experimentation and parameter optimization (Huo et al. 2022).
Transfer learning
Training a CNN model from scratch is an excessively time-consuming and computationally expensive process (Alzubaidi et al. 2021). Therefore, the common approach is to use pre-trained CNN models on datasets. TL is applying the gained knowledge. If training examples are insufficient in a classification problem, pre-trained examples of TL are used. There are two methods of applying a pre-trained model in TL. The first method is to use the pre-trained CNN model as a feature extractor and for classification; the last fully connected layer(s) is changed according to the number of classes in the data set. In the other method, fine-tune the pre-trained CNN model and its retraining of all or part of the layers with specific methods. As a result, a changed architectural design is employed for the new classification task.
Fine-tuning
Fine-tuning, a TL technique, focuses on saving the information obtained while solving a classification task and applying this knowledge to different approaches to a related problem. The major difference between fine-tuning and TL is that only the weights of the newly added classifier layers are optimized in TL, while the whole model is optimized in fine-tuning.
In fine-tuning, unfreeze some of the top layers of a frozen convolutional base model and simultaneously train the last layers of the convolutional base model and the newly added classifier layers. Fine-tuning allows the higher-order feature representations in the convolutional base model to be more relevant for the specific task. In most convolutional networks, the last layers are more specialized than the beginning layers. When you go deeper, the features of the pre-trained model are progressively more specific to the pre-trained dataset. Fine-tuning aims to adjust these specific features to run with the new data rather than overwrite the generic knowledge to adapt to the new classification task. Thus, it is crucial to open how many of the last layers and to train them on how many epochs.
In addition, fine-tuning helps achieve the goal and is an excellent combination for COVID-19 detection, where time to delivery and availability of training data is critical. This method uses pre-trained models to perform the tasks of DL, saving time and afford. It uses the features developed by different algorithms to pass on the learned parameters or information. DL gives decent results when more data can be used, but fine-tuned TL model can achieve the same results with a smaller dataset.
Proposed fine-tuned MobileNetV2 model
In our study, we implemented the MobileNetV2 network (https://www.tensorflow.org/api_docs/python/tf/keras/applications/mobilenet_v2/MobileNetV2) to classify COVID-19 X-ray images. There are two kinds of blocks in MobileNetV2. One is the residual block with stride 1 and the other block is the non-residual block with stride 2, which is used for downsizing (Sandler et al. 2018). Details of the MobileNetV2 model are given in Fig. 3.
MobileNetV2 has 155 layers, including the classification layer. We used this model to transfer information from a related task that has already been achieved. Our proposed model includes 154 pre-trained network layers (convolutional base) and two additional layers, one at the beginning for preprocessing and one at the end for classification of the new task.
As seen in Fig. 4, the classification process was performed by entering the inputs through the layers obtained in the fine-tuning process.
Fig. 4.
Fine-Tuned MobileNetV2 Model
In our approach, we trained the entire model for 50 epochs before fine-tuning. In the first fine-tuning, we opened the last 50 layers of the convolutional base model and created new training loops to train the entire model for 80 epochs, as shown in Fig. 5 (yellow bars). In the second fine-tuning, we started by opening the last layers of the convolutional base model using a step function. In this approach, we reduced the number of opened layers from the last 50 layers by five for every eight cycles, as shown in Fig. 5 (green bars).
Fig. 5.
Fine-tuning number of epoch and layer numbers for Models 1–3 (color figure online)
In our last approach, unlike the other two approaches, we determined the number of epochs and which layers to be opened according to the result using a predefined exponential equation instead of performing the fine-tuning process in a certain decreasing or increasing order. We used the following (Eq. 1) equation to calculate these values:
| 1 |
where x is between 1 and 50 corresponding to defining unfrozen top layers of the base model and y is the calculated value for fine-tuning to define the number of epochs. This equation was determined by empirical observation. Therefore, another exponential decreasing equation can be used for that purpose.
According to this equation, the whole system was trained 80 epochs in the last layers and exponentially decreased to 1 epoch in the layer, as seen in Fig. 5 (blue bars). Thus, our model can preserve more pre-trained generic information. The last layers of CNN models can have special learned properties and the first layers of the CNN models often learn more about generic properties, such as edges, shapes, and textures. This equation provides us with an exponentially decreasing number of training cycles from the last layer to a predetermined depth during the training process. As can be seen in Fig. 5 that in our three approaches, we fine-tuned models at 4000, 2200, and 723 training operations in total for Models 1, 2, and 3, respectively. We used the following training parameters: 0.0001 as the learning rate and Adam as the optimizer.
Evaluation metrics
We used the following evaluation metrics: accuracy, recall, precision, and f1-score as in Eqs. (2)–5.
| 2 |
| 3 |
| 4 |
| 5 |
where True Positive (TP) is the case where the output of the algorithm outputs YES when the actual state is YES, False Positive (FP) is the case where the output of the algorithm outputs NO when the actual state is YES, True Negative (TN) is the case where the output of the algorithm outputs NO when the actual state is NO, False Negative (FN) is the case where the output of the algorithm YES when the actual state is NO.
Results
In this paper, we implemented three different fine-tuning approaches for classifications of three different classes, namely normal, COVID-19, and viral pneumonia, on X-ray images. We combined two datasets containing 9457 images. We evaluated our model using two common evaluation methods, separating the data 30–70% randomly test-train splits using different seeds and repeating five times, and fivefold cross-validation to validate our models. Experimental tests were conducted on a PC with an Intel I7 6700hq 2.60GHz CPU, NVIDIA GTX970M having 3GB GPU, and 16GB RAM. The proposed approach was implemented using the Keras-Tensorflow library. We evaluated our model using the performance metrics given in Eqs. (2)–(5). Table 3 shows the pre-fine-tuning and after fine-tuning accuracy results of three models. Models 1, 2, and 3 demonstrate proposed classical, step, and exponential fine-tuning models, respectively. We used 50 epochs of training to calculate pre-fine-tuning metrics and 80 epochs of training to calculate after fine-tuning metrics. The number of epochs and learning rate parameters were determined empirically. Previous studies showed that using Adam as an optimizer achieved promising results (Bera and Shrivastava 2020). Thus, Adam was chosen as the optimizer. In addition, we defined the batch size as 32 and the learning rate as 0.0001. All parameters of the model are summarized in Table 2.
Table 3.
Pre-fine-tuning and After fine-tuning accuracy and loss values
| Model | Pre-fine-tune acc. | Loss | After fine-tune acc. | Loss |
|---|---|---|---|---|
| Model 1 | 0.9409 | 0.1223 | 0.9588 | 0.0969 |
| Model 2 | 0.9429 | 0.1335 | 0.9739 | 0.0675 |
| Model 3 | 0.9425 | 0.1268 | 0.9781 | 0.0417 |
Table 2.
Model parameters
| Parameter | Value |
|---|---|
| Learning rate | 0.0001 |
| Optimizer | Adam |
| Epochs | 50+80 |
| Batch size | 32 |
As can be seen from Table 3 that before fine-tuning, models gave about 94% accuracy rates. Although these models are the same in the pre-tuning stage, the results are slightly different because of the random separation of datasets and the small number of training epochs. Results after the fine-tuning show that the proposed fine-tuning mechanisms achieved better accuracies than the classical fine-tuning mechanisms. Furthermore, these models were fine-tuned at 80, 44, and 14.46 epochs on average for models 1, 2, and 3, respectively. Model 3 decreased 81.92% of total fine-tuning epochs and still achieved better results.
Table 4 shows the results of the experiments based on different evaluation metrics: Accuracy, recall, precision, and F1-score for the 30–70% hold-out evaluation model. The tests were repeated five times using different seeds and mean, standard deviation (Std. Dev.), and the best and worst values of evaluation metrics were calculated. When the standard deviation was examined, the proposed model 3 was the most robust. It has outperformed the other models in terms of all evaluation metrics.
Table 4.
Numerical results for proposed models using 30–70% hold-out for five different seeds
| Metrics | Model 1 | Model 2 | Model 3 |
|---|---|---|---|
| Accuracy | |||
| Mean | 0.9261 | 0.9583 | 0.9666 |
| Std. Dev. | 0.0079 | 0.0076 | 0.0017 |
| Best | 0.9358 | 0.9608 | 0.9689 |
| Worst | 0.9143 | 0.9499 | 0.9644 |
| Recall | |||
| Mean | 0.9274 | 0.9583 | 0.9665 |
| Std. Dev. | 0.098 | 0.0052 | 0.0017 |
| Best | 0.9400 | 0.9630 | 0.9671 |
| Worst | 0.9140 | 0.9500 | 0.9643 |
| Precision | |||
| Mean | 0.9275 | 0.9582 | 0.9666 |
| Std. Dev. | 0.0093 | 0.0052 | 0.0017 |
| Best | 0.9401 | 0.9631 | 0.9672 |
| Worst | 0.9143 | 0.9496 | 0.9644 |
| F1-Score | |||
| Mean | 0.9275 | 0.9582 | 0.9666 |
| Std. Dev. | 0.0097 | 0.0051 | 0.0017 |
| Best | 0.9401 | 0.9630 | 0.9690 |
| Worst | 0.9143 | 0.9606 | 0.9643 |
We also employed a fivefold cross-validation technique to evaluate the proposed model and to compare the results to the results of other studies. The results of cross-validation of each model are shown in Tables 5, 6 and 7. As can be seen from the tables, Model 3 performed better results and achieved fivefold mean accuracy, recall, precision, and f1-scores of 97.61%, 97.60%, 97.59%, and 97.60%, respectively.
Table 5.
Numerical results for Model 1 using fivefold cross-validation
| K-Fold | Accuracy | Recall | Precision | F1-score |
|---|---|---|---|---|
| Fold 1 | 0.9571 | 0.9571 | 0.9568 | 0.9569 |
| Fold 2 | 0.9582 | 0.9582 | 0.9582 | 0.9582 |
| Fold 3 | 0.9587 | 0.9587 | 0.9582 | 0.9583 |
| Fold 4 | 0.9557 | 0.9577 | 0.9550 | 0.9753 |
| Fold 5 | 0.9513 | 0.9513 | 0.9511 | 0.9511 |
| Mean | 0.9562 | 0.9566 | 0.9558 | 0.9559 |
Table 6.
Numerical results for Model 2 using fivefold cross-validation
| K-fold | Accuracy | Recall | Precision | F1-score |
|---|---|---|---|---|
| Fold 1 | 0.9635 | 0.9635 | 0.9634 | 0.9634 |
| Fold 2 | 0.9582 | 0.9582 | 0.9579 | 0.9580 |
| Fold 3 | 0.9603 | 0.9603 | 0.9599 | 0.9599 |
| Fold 4 | 0.9640 | 0.9640 | 0.9636 | 0.9638 |
| Fold 5 | 0.9592 | 0.9592 | 0.9592 | 0.9592 |
| Mean | 0.9610 | 0.9610 | 0.9608 | 0.9608 |
Table 7.
Numerical results for Model 3 using fivefold cross-validation
| K-fold | Accuracy | Recall | Precision | F1-score |
|---|---|---|---|---|
| Fold 1 | 0.9756 | 0.9756 | 0.9756 | 0.9756 |
| Fold 2 | 0.9725 | 0.9725 | 0.9725 | 0.9725 |
| Fold 3 | 0.9793 | 0.9793 | 0.9792 | 0.9792 |
| Fold 4 | 0.9777 | 0.9777 | 0.9776 | 0.9776 |
| Fold 5 | 0.9751 | 0.9751 | 0.9750 | 0.9751 |
| Mean | 0.9761 | 0.9760 | 0.9759 | 0.9760 |
Figure 6 presents the average confusion matrices for Models 1–3 for a 30–70% hold-out data split using random seeds of five repetitions. From Fig. 6a, we can note that while 1062.2 COVID-19, 387.2 normal, and 1179 pneumonia samples were accurately classified, 21.6 COVID-19, 96.2 normal, and 92.8 pneumonia samples were misclassified using Model 1. Thus, the correct classification rates of COVID-19 samples were 98.09%, while 80.09% for normal and 92.70% for pneumonia samples. As shown in Fig. 6b, 1076 COVID-19, 429 normal, and 1215.8 pneumonia samples were classified correctly, and 7.8 COVID-19, 54.4 normal, and 56 pneumonia cases were misclassified using Model 2. As a result, 99.28%, 88.74%, and 95.59% correct classification rates of COVID-19, normal, and pneumonia were achieved, respectively. It can also be noted in Fig. 6c that 1076 COVID-19, 438.4 normal, and 1230 pneumonia samples were classified correctly, and 7.8 COVID-19, 45 normal, and 41.8 pneumonia cases were misclassified using Model 3. Therefore, the correct classification rates of COVID-19, normal, and pneumonia were 99.28%, 90.69%, and 96.71%, respectively.
Fig. 6.
Calculated confusion matrices of Models 1–3 using 30–70% hold-out
Figure 7 presents the mean confusion matrices of fivefold cross-validation for Models 1–3. From Fig. 7a, we can note that while 717.4 COVID-19, 275 normal, and 816.2 pneumonia samples were accurately classified, 5.8 COVID-19, 40.2 normal, and 36.8 pneumonia samples were misclassified using Model 1. Thus, the correct classification rates of COVID-19 samples were 99.72%, while 87.62% for normal and 94.23% for pneumonia samples. It is shown in Fig. 7b, 718 COVID-19, 278.6 normal, and 821.2 pneumonia samples were classified correctly, and 5.2 COVID-19, 37.6 normal, and 31.8 pneumonia cases were misclassified using Model 2. As a result, 99.28%, 88.10%, and 96.27% correct classification rates of COVID-19, normal, and pneumonia were achieved, respectively. It can also be noted in Fig. 7c that 832.8 COVID-19, 292.2 normal, and 832.8 pneumonia samples were classified correctly, and 5.8 COVID-19, 23 normal, and 20.2 pneumonia cases were misclassified using Model 3. Thus, the correct classification rates of COVID-19, normal, and pneumonia were 99.20%, 92.76%, and 97.63%, respectively.
Fig. 7.
Fivefold cross-validation mean confusion matrices of Models 1–3
In addition to these evaluation metrics, the Friedman test, a popular nonparametric statistical test used to rank the algorithms based on their results without specifying any statistical difference, was performed for fivefold cross-validation results. We achieved chi-square test statistics of 8.0 and a p value of 0.0183. Friedman’s test results show that the p value is less than 0.05. Thus, there are significant differences in the methods. We also employed the Nemenyi post hoc test, used to determine the statistical difference with method pairs, to evaluate the performance of the proposed methods. Table 8 shows the analysis of the Nemenyi tests on the methods using a significance level of 5%. It can be seen from the table that the results of Model 1 and Model 3 have significant differences, and Model 3 achieves better results.
Table 8.
Statistical test results
| Model | 1 | 2 | 3 |
|---|---|---|---|
| 1 | 1.0000 | 0.3341 | 0.0129 |
| 2 | 0.3341 | 1.0000 | 0.3341 |
| 3 | 0.0129 | 0.3341 | 1.0000 |
Figures 8, 9, and 10 demonstrate the training process of fine-tuned Models 1–3 using 30–70% hold-out method, respectively. The green vertical line in the plots shows the 50th epoch, where fine-tuning started. We previously stated that we trained the classification layer added to the base model for 50 epochs and fine-tuned the whole model for additional 80 epochs in a total of 130 epochs. As shown in Fig. 8, validation accuracy flattened about 94% before fine-tuning, and after fine-tuning, it increased to about 95%.
Fig. 8.

The training process of Model 1
Fig. 9.

The training process of Model 2
Fig. 10.

The training process of Model 3
Colored vertical lines in Fig. 9 show stages of step fine-tuning. It is clearly seen that fine-tuning using Model 2 increased validation accuracy from 94% to about 97%. Beginning of the fine-tuning, validation accuracy fluctuated strongly. In later epochs, it settled down to better rates. It was also observed in Fig. 10 that after fine-tuning, validation accuracy increased sharply in early epochs.
Discussion
This study aims to develop a CNN-based COVID-19 diagnostic model using new fine-tuning mechanisms. It aims to bring novel approaches to the literature by incorporating different proposals into the methodological design. Most recent studies on this topic using DL methods are summarized in terms of accuracy metrics in Table 9.
Table 9.
Comparison of the proposed models with previous studies
| Study | Methodology | Accuracy (%) |
|---|---|---|
| Rehman et al. (2020) | ResNet101, MobileNet | 98.75 |
| Shorfuzzaman and Masud (2020) | VGG16, ResNet50, Xception, MobileNet, and DenseNet121 | 99.26 |
| Degadwala et al. (2021) | FT CNN | 90.70 |
| Minaee et al. (2020) | ResNet, SqueezeNet, and DenseNet121 | 98.00 |
| Iskanderani et al. (2021) | DenseNet | 96.25 |
| Pathak et al. (2020) | ResNet32 | 96.22 |
| Apostolopoulos and Mpesiana (2020) | MobileNetV2 | 96.78 |
| Kaur et al. (2021) | AlexNet | 99.52 |
| Wang et al. (2020) | ResNet101 and ResNet152 | 96.10 |
| Misra et al. (2020) | ResNet18 | 93.90 |
| Punn and Agarwal (2021) | ResNet, Inception-v3, and NASNetLarge | 97.00 |
| Ucar and Korkmaz (2020) | Bayes-SqueezeNet(CNN) | 98.30 |
| Narin et al. (2021) | ResNets and Inception | 96.10 |
| Lee et al. (2020) | VGG16 | 95.00 |
| Ismael and Şengür (2021) | ResNet50 | 92.60 |
| Wang et al. (2020) | COVID-Net | 94.00 |
| Nishio et al. (2020) | VGG16 | 86.30 |
| Monshi et al. (2021) | CovidX-RayNet | 95.82 |
| Das et al. (2021) | VGG16 | 97.67 |
| Rahaman et al. (2020) | VGG19 | 89.30 |
| Moujahid et al. (2020) | VGG19 | 96.97 |
| Khan et al. (2020) | Coro-Net | 95.00 |
| Ohata et al. (2020) | DenseNet201 | 95.60 |
| Manokaran et al. (2021) | DenseNet201 | 92.19 |
| Garg et al. (2020) | DenseNet121 | 94.00 |
| Naronglerdrit et al. (2021) | MobileNet | 96.76 |
| Xu et al. (2020) | ResNet + Location Attention | 86.70 |
| Ozturk et al. (2020) | DarkCovidNet | 87.02 |
| Asif and Wenhui (2020) | InceptionV3 | 96.00 |
| Luz et al. (2021) | EfficientNet | 93.90 |
| Bargshady et al. (2022) | CycleGAN | 94.20 |
| Rahimzadeh and Attar (2020) | Xception + ResNet50V2 | 91.40 |
| Model 1 | MobileNetV2+Classical fine-tuning | 95.62 |
| Model 2 | MobileNetV2+Step fine-tuning | 96.10 |
| Model 3 | MobileNetV2+Exponential fine-tuning | 97.61 |
As can be seen in Table 9, many studies have been conducted so far that include the CNN architecture. The major advantage of the CNN architecture is that it allows continuous learning. Three approaches are distinguished in the literature: trained from scratch DL models, TL approaches that use pre-trained models with fine-tuning and hybrid techniques that combine DL models with classical machine learning methods.
Some studies achieved high success rates in classification using a pre-trained CNN models (Iskanderani et al. 2021; Pathak et al. 2020; Kaur et al. 2021; Misra et al. 2020; Apostolopoulos and Mpesiana 2020; Luz et al. 2021; Lee et al. 2020; Ismael and Şengür 2021; Nishio et al. 2020; Das et al. 2021; Rahaman et al. 2020; Moujahid et al. 2020; Ohata et al. 2020; Manokaran et al. 2021; Garg et al. 2020; Naronglerdrit et al. 2021; Asif and Wenhui 2020). It is seen that different pre-trained CNN model designs attained varied results on the same data set (Rehman et al. 2020; Shorfuzzaman and Masud 2020; Minaee et al. 2020; Wang et al. 2020; Narin et al. 2021; Punn and Agarwal 2021; Rahimzadeh and Attar 2020). In Rehman et al. (2020) and (Iskanderani et al. 2021), the authors used binary classification (COVID-19, normal) and achieved more successful results.
To summarize these results, Table 9 includes various studies using CNN models that offer promising results. However, these studies use diverse CNN architectures, training samples, the number of classes, and design parameters. Thus, it is unfair to compare these studies directly. It can be seen from the table that some studies with decent results either have fewer training samples (Shorfuzzaman and Masud 2020; Kaur et al. 2021; Rehman et al. 2020) or have a binary classification (Minaee et al. 2020; Rehman et al. 2020). It is common knowledge that the model trained using a small number of training data results in a poor approximation. The model trained using fewer classification classes achieves better results. In other words, if the model contains a small amount of data or fewer classification classes, it will result in an uncertain and high variance estimate compared to its actual performance (Barbedo 2018).
Conclusions
The COVID-19 virus and pandemic might be called the disease one of the most significant challenges in human history. Many groups, including medical communities and scientific researchers, have attempted to find viable solutions to combat the disease. As a result of numerous studies, advances in artificial intelligence and medical imaging have brought hope for anomaly detection in medical images. Thus, many researchers have developed DL-based COVID-19 diagnosis systems. However, there were still some gaps to be developed.
In this work, we attempted to fill the gaps and presented a new TL-based deep CNN model that includes novel fine-tuning approaches to identify COVID-19. We tested three fine-tuning models, two of which differ from the classical fine-tuning approaches, providing decent COVID-19 detection using chest X-ray images. The first approach is the classical fine-tuning approach in which the last 50 layers of the CNN model are trained at 80 epochs, and the remaining layers are frozen. Unlike the classical fine-tuning approach, the second proposed approach relies on training the last 50 layers of the CNN model in descending order using a step function instead of training at once. Thus, the model achieved a higher success rate than the first approach by minimizing data loss and increasing the number of transferred features in fine-tuning operations. In the last fine-tuning approach, we performed training of the proposed model by determining the number of epochs and which layers would unfreeze for training according to a predetermined exponential function. The proposed second and third models attained promising results and achieved the correct classification rate of COVID-19 as 99.28% and 99,20%, respectively, using fivefold cross-validation. In addition, Model 3 achieved this result using only 18.07% of 4000 total fine-tuning training operations, as seen in Fig. 5. Our study offers a robust and reliable model that can be used as a decision support system to detect COVID-19 disease. Thus, computer-aided diagnosis systems can be created, and using those systems can reduce the X-ray evaluation time of the experts. The numerical results show that training more of the bottom layers of the model has the ability to generalize to the new problem and increase the classification results. Our approach could be accepted as an alternative COVID-19 diagnostic tool or contribute to the further development of diagnostic methods. In addition, the proposed fine-tuning mechanisms can also be applied to other medical applications, such as breast cancer and tumor detection.
For future work, meta-heuristics-based hyperparameter tuning (Kıymaç and Kaya 2023) can be implemented to optimize the model and to get better results. In addition, we plan to extend the proposed methodology to some other relevant pulmonary diseases. A lightweight DL model can be trained from scratch to classify COVID-19 cases and implemented on mobile devices for smart health applications.
Funding
Not applicable.
Data Availibility
The datasets analyzed during the current study are available in the chest X-ray (ChestX-RayImagesPneumonia 2022) and COVID-19 Radiography Database repositories (Covid-19RadiographyDatabase 2022), https://www.kaggle.com/paultimothymooney/chest-xray-pneumonia, https://www.kaggle.com/tawsifurrahman/covid19-radiography-database.
Declarations
Conflict of interest
All of the authors declare that they have no conflict of interest.
Human and animal rights
This article does not contain any studies with human participants or animals performed by any of the authors.
Footnotes
This article has been retracted. Please see the retraction notice for more detail:https://doi.org/10.1007/s00500-024-10013-9
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Change history
8/7/2024
This article has been retracted. Please see the Retraction Notice for more detail: 10.1007/s00500-024-10013-9
References
- Abiyev RH, Ma’aitah MKS (2018) Deep convolutional neural networks for chest diseases detection. J Healthc Eng. 10.1155/2018/4168538 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Afshar-Oromieh A, Prosch H, Schaefer-Prokop C, Bohn KP, Alberts I, Mingels C, Thurnher M, Cumming P, Shi K, Peters A et al (2021) A comprehensive review of imaging findings in covid-19-status in early 2021. Eur J Nucl Med Mol Imaging. 10.1007/s00259-021-05375-3 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Ai T, Yang Z, Hou H, Zhan C, Chen C, Lv W, Tao Q, Sun Z, Xia L (2020) Correlation of chest CT and RT-PCR testing for coronavirus disease 2019 (covid-19) in china: a report of 1014 cases. Radiology 296(2):E32. 10.1148/radiol.2020200642 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Alzubaidi L, Zhang J, Humaidi AJ, Al-Dujaili A, Duan Y, Al-Shamma O, Santamaría J, Fadhel MA, Al-Amidie M, Farhan L (2021) Review of deep learning: Concepts, cnn architectures, challenges, applications, future directions. J big Data 8(1):1. 10.1186/s40537-021-00444-8 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Apostolopoulos ID, Mpesiana TA (2020) Covid-19: automatic detection from x-ray images utilizing transfer learning with convolutional neural networks. Phys Eng Sci Med 43(2):635. 10.1007/s13246-020-00865-4 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Asif S, Wenhui Y (2020) Automatic detection of covid-19 using x-ray images with deep convolutional neural networks and machine learning. medRxiv. 10.1109/ICCC51575.2020.9344870 [Google Scholar]
- Ayan E, Ünver HM (2019) 2019 scientific meeting on electrical-electronics & biomedical engineering and computer science (EBBT) (Ieee), pp. 1–5. 10.1109/EBBT.2019.8741582
- Bala SA, Kant S, Kumar K (2019) Impact of deep learning in medical imaging: a systematic new proposed model. Int J Recent Technol Eng. 10.35940/ijrte.C1019.1083S219 [Google Scholar]
- Barbedo JGA (2018) Impact of dataset size and variety on the effectiveness of deep learning and transfer learning for plant disease classification. Comput Electron Agric 153:46. 10.1016/j.compag.2018.08.013 [Google Scholar]
- Bargshady G, Zhou X, Barua PD, Gururajan R, Li Y, Acharya UR (2022) Application of cyclegan and transfer learning techniques for automated detection of covid-19 using x-ray images. Pattern Recogn Lett 153:67. 10.1016/j.patrec.2021.11.020 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Bera S, Shrivastava VK (2020) Analysis of various optimizers on deep convolutional neural network model in the application of hyperspectral remote sensing image classification. Int J Remote Sens 41(7):2664. 10.1080/01431161.2019.1694725 [Google Scholar]
- Bhattacharyya A, Bhaik D, Kumar S, Thakur P, Sharma R, Pachori RB (2022) A deep learning based approach for automatic detection of covid-19 cases using chest x-ray images. Biomed Signal Process Control 71:103182. 10.1016/j.bspc.2021.103182 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Boccaletti S, Ditto W, Mindlin G, Atangana A (2020) Modeling and forecasting of epidemic spreading: the case of covid-19 and beyond. Chaos Solitons Fractals 135:109794. 10.1016/j.chaos.2020.109794 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Chaunzwa TL, Hosny A, Xu Y, Shafer A, Diao N, Lanuti M, Christiani DC, Mak RH, Aerts HJ (2021) Deep learning classification of lung cancer histology using CT images. Sci Rep 11(1):1. 10.1038/s41598-021-84630-x [DOI] [PMC free article] [PubMed] [Google Scholar]
- ChestX-RayImagesPneumonia. Chest X-Ray Images Pneumonia, howpublished = “https://www.kaggle.com/paultimothymooney/chest-xray-pneumonia”, note = Accessed: 2022-1-3 (2022)
- Covid-19RadiographyDatabase. Covid-19 Radiography Database, howpublished = “https://www.kaggle.com/tawsifurrahman/covid19-radiography-database”, note = Accessed: 2022-1-3 (2022)
- Das AK, Kalam S, Kumar C, Sinha D (2021) Tlcov-an automated covid-19 screening model using transfer learning from chest x-ray images. Chaos Solitons Fractals 144:110713. 10.1016/j.chaos.2021.110713 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Degadwala S, Vyas D, Dave H (2021) 2021 international conference on artificial intelligence and smart systems (ICAIS) (IEEE), pp. 609–613. 10.1109/ICAIS50930.2021.9395864
- Feng Z, Yu Q, Yao S, Luo L, Zhou W, Mao X, Li J, Duan J, Yan Z, Yang M et al (2020) Early prediction of disease progression in covid-19 pneumonia patients with chest CT and clinical characteristics. Nat Commun 11(1):1. 10.1038/s41467-020-18786-x [DOI] [PMC free article] [PubMed] [Google Scholar]
- Garg T, Garg M, Mahela OP, Garg AR (2020) Convolutional neural networks with transfer learning for recognition of covid-19: a comparative study of different approaches. AI 1(4):586. 10.3390/ai1040034 [Google Scholar]
- Gupta A, Gupta S, Katarya R et al (2021) Instacovnet-19: a deep learning classification model for the detection of covid-19 patients using chest x-ray. Appl Soft Comput 99:106859. 10.1016/j.asoc.2020.106859 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Huo H, Yu Y, Liu Z (2022) Facial expression recognition based on improved depthwise separable convolutional network. Multimed Tools Appl. 10.1007/s11042-022-14066-6 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Iskanderani AI, Mehedi IM, Aljohani AJ, Shorfuzzaman M, Akther F, Palaniswamy T, Latif SA, Latif A, Alam A (2021) Artificial intelligence and medical internet of things framework for diagnosis of coronavirus suspected cases. J Healthc Eng. 10.1155/2021/3277988 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Islam SR, Maity SP, Ray AK, Mandal M (2019) 2019 IEEE Canadian Conference of Electrical and Computer Engineering (CCECE) (IEEE), pp. 1–4. 10.1109/CCECE.2019.8861969
- Ismael AM, Şengür A (2021) Deep learning approaches for covid-19 detection based on chest x-ray images. Expert Syst Appl 164:114054. 10.1016/j.eswa.2020.114054 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Kaur M, Kumar V, Yadav V, Singh D, Kumar N, Das NN (2021) Metaheuristic-based deep covid-19 screening model from chest x-ray images. J Healthc Eng. 10.1155/2021/8829829 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Kerr JR, Freeman AL, Marteau TM, van der Linden S (2021) Effect of information about covid-19 vaccine effectiveness and side effects on behavioural intentions: two online experiments. Vaccines 9(4):379. 10.3390/vaccines9040379 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Khan AI, Shah JL, Bhat MM (2020) Coronet: A deep neural network for detection and diagnosis of covid-19 from chest x-ray images. Comput Methods Programs Biomed 196:105581. 10.1016/j.cmpb.2020.105581 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Kıymaç E, Kaya Y (2023) A novel automated cnn arrhythmia classifier with memory-enhanced artificial hummingbird algorithm. Expert Syst Appl 213:119162 [Google Scholar]
- Krizhevsky A, Sutskever I, Hinton GE (2012) Imagenet classification with deep convolutional neural networks. Adv Neural Inf Process Syst 25:1097. 10.1145/3065386 [Google Scholar]
- Kwekha-Rashid AS, Abduljabbar HN, Alhayani B (2021) Coronavirus disease (covid-19) cases analysis using machine-learning applications. Appl Nanosci. 10.1007/s13204-021-01868-7 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Lee KS, Kim JY, Et Jeon, Choi WS, Kim NH, Lee KY (2020) Evaluation of scalability and degree of fine-tuning of deep convolutional neural networks for covid-19 screening on chest x-ray images using explainable deep-learning algorithm. J Pers Med 10(4):213. 10.3390/jpm10040213 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Luz E, Silva P, Silva R, Silva L, Guimarães J, Miozzo G, Moreira G, Menotti D (2021) Towards an effective and efficient deep learning model for covid-19 patterns detection in x-ray images. Res Biomed Eng. 10.1007/s42600-021-00151-6 [Google Scholar]
- Mahajan A, Pawar V, Punia V, Vaswani A, Gupta P, Bharadwaj K, Salunke A, Palande SD, Banderkar K, Apparao M (2022) Deep learning-based covid-19 triage tool: an observational study on an x-ray dataset. Cancer Res Stat Treat 5(1):19. 10.4103/crst.crst_162_21 [Google Scholar]
- Maior CB, Santana JM, Lins ID, Moura MJ (2021) Convolutional neural network model based on radiological images to support covid-19 diagnosis: evaluating database biases. PLoS ONE 16(3):e0247839. 10.1371/journal.pone.0247839 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Manokaran J, Zabihollahy F, Hamilton-Wright A, Ukwatta E (2021) Detection of covid-19 from chest x-ray images using transfer learning. J Med Imaging 8(S1):017503. 10.1117/1.JMI.8.S1.017503 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Minaee S, Kafieh R, Sonka M, Yazdani S, Soufi GJ (2020) Deep-covid: predicting covid-19 from chest x-ray images using deep transfer learning. Med Image Anal 65:101794. 10.1016/j.media.2020.101794 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Misra S, Jeon S, Lee S, Managuli R, Jang IS, Kim C (2020) Multi-channel transfer learning of chest x-ray images for screening of covid-19. Electronics 9(9):1388. 10.3390/electronics9091388 [Google Scholar]
- Monshi MMA, Poon J, Chung V, Monshi FM (2021) Covidxraynet: optimizing data augmentation and cnn hyperparameters for improved covid-19 detection from cxr. Comput Biol Med 133:104375. 10.1016/j.compbiomed.2021.104375 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Moujahid H, Cherradi B, Al-Sarem M, Bahatti L (2020) International conference of reliable information and communication technology (Springer), pp. 148–159. 10.1007/978-3-030-70713-2_16
- Narin A, Kaya C, Pamuk Z (2021) Automatic detection of coronavirus disease (covid-19) using x-ray images and deep convolutional neural networks. Pattern Anal Appl. 10.1007/s10044-021-00984-y [DOI] [PMC free article] [PubMed] [Google Scholar]
- Naronglerdrit P, Mporas I, Sheikh-Akbari A (2021) Data science for COVID-19 (Elsevier), pp. 255–273. 10.1016/B978-0-12-824536-1.00031-9
- Nishio M, Noguchi S, Matsuo H, Murakami T (2020) Automatic classification between covid-19 pneumonia, non-covid-19 pneumonia, and the healthy on chest x-ray image: combination of data augmentation methods. Sci Rep 10(1):1. 10.1038/s41598-020-74539-2 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Nour M, Cömert Z, Polat K (2020) A novel medical diagnosis model for covid-19 infection detection based on deep features and Bayesian optimization. Appl Soft Comput 97:106580. 10.1016/j.asoc.2020.106580 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Ohata EF, Bezerra GM, Chagas JVS, Neto AVL, Albuquerque AB, de Albuquerque VHC, Reboucas Filho PP (2020) Automatic detection of covid-19 infection using chest x-ray images through transfer learning. IEEE/CAA J Autom Sin 8(1):239. 10.1109/JAS.2020.1003393 [Google Scholar]
- Ozturk T, Talo M, Yildirim EA, Baloglu UB, Yildirim O, Acharya UR (2020) Automated detection of covid-19 cases using deep neural networks with x-ray images. Comput Biol Med 121:103792. 10.1016/j.compbiomed.2020.103792 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Pathak Y, Shukla PK, Tiwari A, Stalin S, Singh S (2020) Deep transfer learning based classification model for covid-19 disease. Irbm. 10.1016/j.irbm.2020.05.003 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Punn NS, Agarwal S (2021) Automated diagnosis of covid-19 with limited posteroanterior chest x-ray images using fine-tuned deep neural networks. Appl Intell 51(5):2689. 10.1007/s10489-020-01900-3 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Rahaman MM, Li C, Yao Y, Kulwa F, Rahman MA, Wang Q, Qi S, Kong F, Zhu X, Zhao X (2020) Identification of covid-19 samples from chest x-ray images using deep learning: a comparison of transfer learning approaches. J Xray Sci Technol 28(5):1. 10.3233/XST-200715 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Rahimzadeh M, Attar A (2020) A modified deep convolutional neural network for detecting covid-19 and pneumonia from chest x-ray images based on the concatenation of xception and resnet50v2. Inf Med Unlocked 19:100360. 10.1016/j.imu.2020.100360 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Rehman A, Naz S, Khan A, Zaib A, Razzak I (2020) Improving coronavirus (covid-19) diagnosis using deep transfer learning. MedRxiv. 10.1007/978-981-16-7618-5_3 [Google Scholar]
- Sandler M, Howard A, Zhu M, Zhmoginov A, Chen LC (2018) Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 4510–4520. 10.48550/arXiv.1801.04381
- Seibert FS, Toma D, Bauer F, Paniskaki K, Anft M, Rohn BJ, Wang S, Racovitan D, Babel N, Westhoff TH (2020) Detection of SARS-CoV-2 pneumonia: two case reports. J Med Case Reports 14(1):1. 10.1186/s13256-020-02551-1 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Shams M, Elzeki O, Abd Elfattah M, Medhat T, Hassanien AE (2020) Big data analytics and artificial intelligence against COVID-19: innovation vision and approach (Springer), pp. 147–162. 10.1007/978-3-030-55258-9_9
- Shorfuzzaman M, Masud M (2020) On the detection of covid-19 from chest x-ray images using cnn-based transfer learning. Comput Mater Contin. 10.32604/cmc.2020.011326 [Google Scholar]
- Shorten C, Khoshgoftaar TM (2019) A survey on image data augmentation for deep learning. J Big Data 6(1):1. 10.1186/s40537-019-0197-0 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Song Y, Zheng S, Li L, Zhang X, Zhang X, Huang Z, Chen J, Wang R, Zhao H, Zha Y et al (2021) Deep learning enables accurate diagnosis of novel coronavirus (covid-19) with CT images. IEEE/ACM Trans Comput Biol Bioinf. 10.1109/TCBB.2021.3065361 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Song L, Liu X, Chen S, Liu S, Liu X, Muhammad K, Bhattacharyya S (2022) A deep fuzzy model for diagnosis of covid-19 from ct images. Appl Soft Comput 122:108883. 10.1016/j.asoc.2022.108883 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Stephen O, Sain M, Maduh UJ, Jeong DU (2019) An efficient deep learning approach to pneumonia classification in healthcare. J Healthc Eng. 10.1155/2019/4180949 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Sun T, Wang Y (2020) Modeling covid-19 epidemic in heilongjiang province, china. Chaos Solitons Fractals 138:109949. 10.1016/j.chaos.2020.109949 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Susanto AD, Agustin H, Taufik M, Rahman MA, Hidayat M (2022) Accuracy of volatile organic compound (VOC) detection in exhaled breath compared to reverse-transcriptase polymerase chain reaction (RT-PCR) for diagnosis of covid-19: an evidence-based case report. Arch Clin Infect Dis. 10.5812/archcid-119263 [Google Scholar]
- Suzuki K (2017) Overview of deep learning in medical imaging. Radiol Phys Technol 10(3):257. 10.1007/s12194-017-0406-5 [DOI] [PubMed] [Google Scholar]
- Tajbakhsh N, Shin JY, Gurudu SR, Hurst RT, Kendall CB, Gotway MB, Liang J (2016) Convolutional neural networks for medical image analysis: Full training or fine tuning? IEEE Trans Med Imaging 35(5):1299. 10.1109/TMI.2016.2535302 [DOI] [PubMed] [Google Scholar]
- Tammina S (2022) ICDSMLA 2020 (Springer), pp. 431–447. 10.1007/978-981-16-3690-5_37
- Tensorflow Page, howpublished = “https://www.tensorflow.org/api_docs/python/tf/keras/applications/mobilenet_v2/MobileNetV2”, note = Accessed: 2021-10-5
- Toğaçar M, Ergen B, Cömert Z (2020) Covid-19 detection using deep learning models to exploit social mimic optimization and structured chest x-ray images using fuzzy color and stacking approaches. Comput Biol Med 121:103805. 10.1016/j.compbiomed.2020.103805 [DOI] [PMC free article] [PubMed]
- Ucar F, Korkmaz D (2020) Covidiagnosis-net: deep bayes-squeezenet based diagnosis of the coronavirus disease 2019 (covid-19) from x-ray images. Med Hypotheses 140:109761. 10.1016/j.mehy.2020.109761 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Varela-Santos S, Melin P (2021) A new approach for classifying coronavirus covid-19 based on its manifestation on chest x-rays using texture features and neural networks. Inf Sci 545:403. 10.1016/j.ins.2020.09.041 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Varshni D, Thakral K, Agarwal L, Nijhawan R, Mittal A (2019) 2019 IEEE International Conference on Electrical, Computer and Communication Technologies (ICECCT) (IEEE), pp. 1–7. 10.1109/ICECCT.2019.8869364
- Virtual press conference on COVID-19, howpublished = “https://www.euro.who.int/en/health-topics/health-emergencies/coronavirus-covid19/news/news/2020/3/who-announces-covid-19-outbreak-a-pandemic”. Accessed: 2022-23-1
- Wang L, Lin ZQ, Wong A (2020) Covid-net: a tailored deep convolutional neural network design for detection of covid-19 cases from chest x-ray images. Sci Rep 10(1):1. 10.1038/s41598-020-76550-z [DOI] [PMC free article] [PubMed] [Google Scholar]
- Wang S, Kang B, Ma J, Zeng X, Xiao M, Guo J, Cai M, Yang J, Li Y, Meng X et al (2021) A deep learning algorithm using ct images to screen for corona virus disease (covid-19). Eur Radiol. 10.1007/s00330-021-07715-1 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Wang N, Liu H, Xu C (2020) 2020 IEEE 10th international conference on electronics information and emergency communication (ICEIEC) (IEEE), pp. 281–284. 10.1109/ICEIEC49280.2020.9152329
- WHO Coronavirus (COVID-19) Dashboard, howpublished = “https://covid19.who.int/”. Accessed: 2022-14-12
- Wong HYF, Lam HYS, Fong AHT, Leung ST, Chin TWY, Lo CSY, Lui MMS, Lee JCY, Chiu KWH, Chung TWH et al (2020) Frequency and distribution of chest radiographic findings in patients positive for covid-19. Radiology 296(2):E72. 10.1148/radiol.2020201160 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Xu X, Jiang X, Ma C, Du P, Li X, Lv S, Yu L, Ni Q, Chen Y, Su J et al (2020) A deep learning system to screen novel coronavirus disease 2019 pneumonia. Engineering 6(10):1122. 10.1016/j.eng.2020.04.010 [DOI] [PMC free article] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Data Availability Statement
The datasets analyzed during the current study are available in the chest X-ray (ChestX-RayImagesPneumonia 2022) and COVID-19 Radiography Database repositories (Covid-19RadiographyDatabase 2022), https://www.kaggle.com/paultimothymooney/chest-xray-pneumonia, https://www.kaggle.com/tawsifurrahman/covid19-radiography-database.







