Abstract
Background
The COVID-19 outbreak has spread rapidly and hospitals are overwhelmed with COVID-19 patients. While analysis of nasal and throat swabs from patients is the main way to detect COVID-19, analyzing chest images could offer an alternative method to hospitals, where health care personnel and testing kits are scarce. Deep learning (DL), in particular, has shown impressive levels of performance when analyzing medical images, including those related to COVID-19 pneumonia.
Objective
The goal of this study was to perform a systematic review with a meta-analysis of relevant studies to quantify the performance of DL algorithms in the automatic stratification of COVID-19 patients using chest images.
Methods
A search strategy for use in PubMed, Scopus, Google Scholar, and Web of Science was developed, where we searched for articles published between January 1 and April 25, 2020. We used the key terms “COVID-19,” or “coronavirus,” or “SARS-CoV-2,” or “novel corona,” or “2019-ncov,” and “deep learning,” or “artificial intelligence,” or “automatic detection.” Two authors independently extracted data on study characteristics, methods, risk of bias, and outcomes. Any disagreement between them was resolved by consensus.
Results
A total of 16 studies were included in the meta-analysis, which included 5896 chest images from COVID-19 patients. The pooled sensitivity and specificity of the DL models in detecting COVID-19 were 0.95 (95% CI 0.94-0.95) and 0.96 (95% CI 0.96-0.97), respectively, with an area under the receiver operating characteristic curve of 0.98. The positive likelihood, negative likelihood, and diagnostic odds ratio were 19.02 (95% CI 12.83-28.19), 0.06 (95% CI 0.04-0.10), and 368.07 (95% CI 162.30-834.75), respectively. The pooled sensitivity and specificity for distinguishing other types of pneumonia from COVID-19 were 0.93 (95% CI 0.92-0.94) and 0.95 (95% CI 0.94-0.95), respectively. The performance of radiologists in detecting COVID-19 was lower than that of the DL models; however, the performance of junior radiologists was improved when they used DL-based prediction tools.
Conclusions
Our study findings show that DL models have immense potential in accurately stratifying COVID-19 patients and in correctly differentiating them from patients with other types of pneumonia and normal patients. Implementation of DL-based tools can assist radiologists in correctly and quickly detecting COVID-19 and, consequently, in combating the COVID-19 pandemic.
Keywords: COVID-19, SARS-CoV-2, pneumonia, artificial intelligence, deep learning
Introduction
COVID-19 is a serious global infectious disease and is spreading at an unprecedented level worldwide [1,2]. The World Health Organization declared this infectious disease a public health emergency of international concern and then declared it a pandemic. SARS-CoV-2 is even more contagious than SARS-CoV or Middle East respiratory syndrome coronavirus and is sometimes undetected due to people having asymptomatic or mild symptoms [3,4]. Earlier detection paired with aggressive public health steps, such as social distancing and isolation of suspected or sick patients, can help tackle the crisis [5]. Presently, reverse transcription–polymerase chain reaction (RT-PCR), gene sequencing, and analysis of blood specimens are considered the gold standard methods for detecting COVID-19; however, the performance of these methods (∼73% sensitivity for nasal swabs and ∼61% for throat swabs) is not satisfactory [6,7]. Since hospitals are overwhelmed by COVID-19 patients, those with severe acute respiratory illness are given priority over others with mild symptoms. Therefore, a large number of undiagnosed patients may lead to a serious risk of cross-infection.
Chest radiography imaging (eg, x-ray and computed tomography [CT] scan) is often used as an effective tool for the quick diagnosis of pneumonia [8,9]. The CT scan images of COVID-19 patients show multilobar involvement and peripheral airspace, mostly ground-glass opacities [10,11]. Moreover, asymmetric patchy or diffuse airspace opacities have also been reported in patients with SARS-CoV-2 infection [12]. These changes in CT scan images can be easily interpreted by a trained or experienced radiologist. Automatic classification of COVID-19 patients, however, has huge benefits, such as increasing efficiency, wide coverage, reducing barriers to access, and improving patient outcomes. Several studies demonstrated the application of deep learning (DL) techniques to identify and detect novel COVID-19 using radiography images [13,14].
Herein, we report the results of a comprehensive systematic review of DL algorithm studies that investigated the performance of DL algorithms for COVID-19 classification from chest radiography imaging. Our main objective was to quantify the performance of DL methods for COVID-19 classification, which might encourage health care policy makers to implement DL-based automated tools in the real-world clinical setting. DL-based automated tools can help reduce radiologists’ workload, as DL can help maintain diagnostic radiology support in real time and with increased sensitivity.
Methods
Experimental Approach
The PRISMA (Preferred Reporting Items for Systematic Reviews and Meta-Analyses) guidelines, which are based on the Cochrane Handbook for Systematic Reviews of Interventions, were used to conduct this study [15].
Literature Search
We searched electronic databases, such as PubMed, Scopus, Google Scholar, and Web of Science, for articles published between January 1 and April 25, 2020. We developed a search strategy using combinations of the following Medical Subject Headings: “COVID-19,” or “coronavirus,” or “SARS-CoV-2,” or “novel corona,” or “2019-ncov,” and “deep learning,” or “artificial intelligence,” or “automatic detection.” Reference lists of the retrieved articles and relevant reviews were also checked for additional eligible articles.
Eligibility Criteria
During the first screening, two authors (MMI and TNP) assessed the title and abstract of each article and excluded irrelevant articles. To include eligible articles, those two authors examined the full text of the articles and evaluated whether they fulfilled the inclusion criteria of this study. Disagreement during this selection process was resolved by consensus or, if necessary, the main investigator (YCL) was consulted. We included articles if they met the following criteria: (1) were published in English, (2) were published in a peer-reviewed journal, (3) assessed performance of a DL model to detect COVID-19, and (4) provided a clear description of the methodology and the total number of images. We excluded studies if they were published in preprint repositories or if they were published in the form of a review or a letter to the editor.
Data Extraction and Synthesis
Two authors (MMI and TNP) independently screened all titles and abstracts of retrieved articles. The most relevant studies were selected based on the predefined selection criteria. Any disagreement during the screening process was resolved by discussion with the other authors; unsettled issues were settled by discussion with the study supervisor (YCL). The two authors who conducted the first screening cross-checked studies for duplication by comparing author names, publication dates, and journal names. They excluded all duplicate studies. Afterward, they collected data from the selected studies, such as author name, publication year, location, model description, total number of images, total number of COVID-19 cases and images, imaging modality, total number of patients, sensitivity, specificity, accuracy, area under the receiver operating characteristic curve (AUROC), and database.
Risk of Bias Assessment
The Quality Assessment of Diagnostic Accuracy Studies-2 (QUADAS-2) tool was used to assess the quality of the selected studies [16]. The QUADAS-2 scale comprises four domains: patient selection, index test, reference standard, and flow and timing. The first three domains are used to evaluate the risk of bias in terms of concerns regarding applicability. The overall risk of bias was categorized into three groups: low, high, and unclear risk of bias.
Statistical Analysis
Meta-DiSc, version 1.4, was used to calculate the evaluation metrics of the DL model. The software was also used to (1) perform statistical pooling from each study and (2) assess the homogeneity with a variety of statistics, including chi-square and I2. The sensitivity and specificity with 95% CIs in distinguishing between COVID-19 patients, patients with other types of pneumonia, and normal patients were calculated. The pooled receiver operating characteristic (ROC) curve was plotted and the area under the curve (AUC) was calculated with 95% CIs based on the DerSimonian-Laird random effects model method. The diagnostic odds ratio (DOR) was calculated by the Moses constant of the linear model. Diagnostic tests where the DOR is constant, regardless of the diagnostic threshold, have symmetrical curves around the sensitivity-specificity line. In these situations, it is possible to combine DORs using the DerSimonian-Laird method to estimate the overall DOR and, hence, to determine the best-fitting ROC curve [17]. The mathematical equation is given below:
When the DOR changes with the diagnostic threshold, the ROC curve is asymmetrical. To fit the DOR variation based on a different threshold, the Moses-Shapiro-Littenberg method was used. It consists of observing the relationship by fitting the straight line:
D = a + bS (2) |
where D is the log of DOR and S is a measure of threshold given by the following:
Estimates of parameters a and b and their standard errors and covariance were obtained by the ordinary or weighted least squares method using the NAG Library for C (The Numerical Algorithms Group).
The ROC curve is the AUC that summarized the diagnostic performance as a single number: an AUC close to 1 is considered a perfect curve and an AUC close to 0.5 is considered poor [18]. The AUC is computed by numeric integration of the curve equation by the trapezoidal method [19]. The Q* index is defined by the point where sensitivity and specificity are equal, which is the point closest to the ideal top-left corner of the ROC curve space. It was calculated by the following:
Moreover, the standard error of the AUROC was calculated by following equation:
The standard error of Q* was calculated by following equation:
Results
Selection Criteria
Figure 1 shows the process of identifying relevant DL studies. A total of 562 studies were retrieved by searching electronic databases and by reviewing their reference lists. We excluded 435 duplicate studies and an additional 104 studies that did not fulfill the selection criteria. We reviewed 23 full-text studies and further excluded 7 studies because of the reasons shown in Figure 1. Finally, we included 16 studies in the meta-analysis [13,14,20-33].
Characteristics of Included Studies
Among the 16 DL-based COVID-19 detection studies, we identified 5896 digital images for COVID-19 patients and 645,825 images for non-COVID-19 patients, including those with other types of viral pneumonia and normal patients. Included studies used DL algorithms, such as convolutional neural networks, MobileNetV2, and COVNet, for stratifying COVID-19 patients with higher accuracy. The range of accuracy for detecting COVID-19 correctly was 76.00 to 99.51. A total of 8 studies used CT images and 8 studies used x-ray images. The characteristics of the included studies in the meta-analysis are shown in Table 1 [13,14,20-33].
Table 1.
Author | Modality | Method | Images, n | COVID-19 images, n | Sensitivity | Specificity | Accuracy |
Apostolopoulos and Mpesiana [14] | X-ray | MobileNetV2 | 1428 | 224 | 98.66 | 96.46 | 99.18 |
Butt et al [13] | Computed tomography (CT) | Convolutional neural network (CNN) | 618 | 219 | 98.20 | 92.20 | —a |
Apostolopoulos et al [21] | X-ray | MobileNetV2 | 3905 | 463 | 97.36 | 99.42 | 96.78 |
Li et al [25] | CT | COVNet | 4356 | 127 | 90.00 | 95.00 | — |
Ucar and Korkmaz [29] | X-ray | CNN | 4608b | 1536b | — | 99.13 | 98.30 |
Ozturk et al [26] | X-ray | CNN and DarkNet | 1186 | 108 | 95.13 | 95.30 | 98.08 |
Bai et al [24] | CT | EfficientNet | 1186 | 521 | 95.00 | 96.00 | 96.00 |
Zhang at al [33] | CT | DeepLabv3 | 617,775 | — | 94.93 | 91.13 | 92.49 |
El Asnaoui and Chawki [20] | X-ray | Inception ResNet V2 | 6087 | 231 | 92.11 | 96.06 | — |
Ardakani et al [22] | CT | ResNet-101 | 1020 | 510 | 100 | 99.02 | 99.51 |
Pathak et al [27] | CT | CNN | 852 | 413 | 91.45 | 94.77 | 93.01 |
Wu et al [32] | CT | ResNet50 | 495 | 368 | 81.10 | 61.50 | 76.00 |
Toğaçar et al [28] | X-ray | SqueezeNet | 458 | 295 | 100 | 100 | 100 |
Waheed at al [30] | X-ray | ACGANc | 1124 | 403 | 90.00 | 97.00 | 95.00 |
Khan et al [23] | X-ray | Xception | 1251 | 284 | 99.30 | 98.60 | 99.00 |
Wang et al [31] | CT | DenseNet | 5372 | 102 | 80.39 | 76.66 | 78.32 |
Wang et al [31] | CT | DenseNet | 5372 | 92 | 79.35 | 81.16 | 80.12 |
aNot reported.
bAugmented images.
cACGAN: auxiliary classifier generative adversarial network.
Model Performance
Based on the 16 studies, the performance of the DL algorithms for detecting COVID-19 was determined and is summarized in Table 2 [22,24,33]. The pooled sensitivity and specificity of the DL methods for detecting COVID-19 was 0.95 (95% CI 0.94-0.95) and 0.96 (95% CI 0.96-0.97), respectively, with a summary ROC (SROC) of 0.98 (Figure 2). The pooled sensitivity and specificity are shown in Figure 3.
Table 2.
Class and method | Data sets, n | Sensitivity (95% CI) | Specificity (95% CI) | Positive likelihood ratio (95% CI) | Negative likelihood ratio (95% CI) | AUROCa | Accuracy |
|
||||||||||||||||||
COVID-19 | ||||||||||||||||||||||||||
|
Deep learning model | 17 | 0.95 (0.94-0.95) | 0.96 (0.96-0.97) | 19.02 (12.83-28.19) | 0.06 (0.04-0.10) | 0.98 | —b |
|
|||||||||||||||||
|
Radiologists (Bai et al [24]) | |||||||||||||||||||||||||
|
|
Total | 6 | 0.79 (0.64-0.89) | 0.88 (0.78-0.94) | — | — | — | 0.85 |
|
||||||||||||||||
|
|
Juniorc | 3 | 0.80 (0.72-0.87) | 0.88 (0.83-0.92) | — | — | — | — |
|
||||||||||||||||
|
|
Seniord | 3 | 0.78 (0.70-0.85) | 0.87 (0.82-0.91) | — | — | — | — |
|
||||||||||||||||
|
|
Junior + AIe | — | 0.88 (0.81-0.93) | 0.93 (0.89-0.96) | — | — | — | — |
|
||||||||||||||||
|
|
Senior + AI | — | 0.88 (0.81-0.93) | 0.89 (0.84-0.93) | — | — | — | — |
|
||||||||||||||||
|
Radiologists (Zhang et al [33]) | |||||||||||||||||||||||||
|
|
Total | 8 | 0.75 (0.65-0.84) | 0.90 (0.86-0.94) | — | — | — | — |
|
||||||||||||||||
|
|
Junior | 4 | 0.65 (0.48-0.79) | 0.89 (0.81-0.94) | — | — | — | 0.82 |
|
||||||||||||||||
|
|
Senior | 4 | 0.85 (0.70-0.94) | 0.91 (0.85-0.96) | — | — | — | 0.90 |
|
||||||||||||||||
|
|
Junior + AI | — | 0.80 (0.64-0.90) | 0.94 (0.88-0.97) | — | — | — | 0.90 |
|
||||||||||||||||
|
Radiologist (Ardakani et al [22]; senior) | 1 | 0.89 (0.81-0.94) | 0.83 (0.74-0.89) | — | — | — | — |
|
|||||||||||||||||
Other types of pneumonia: deep learning model | 7 | 0.93 (0.92-0.94) | 0.95 (0.94-0.95) | 22.45 (12.86-39.19) | 0.06 (0.03-0.13) | 0.98 | — |
|
||||||||||||||||||
Normal: deep learning model | 6 | 0.95 (0.94-0.96) | 0.98 (0.97-0.98) | 47.47 (20.70-108.86) | 0.04 (0.02-0.08) | 0.99 | — |
|
aAUROC: area under the receiver operating characteristic curve.
bNot reported.
cJunior radiologists have 5 to 15 years of experience.
dSenior radiologists have 15 to 25 years of experience.
eAI: artificial intelligence.
DL methods were able to correctly distinguish other types of pneumonia from COVID-19 with an SROC of 0.98 (sensitivity: 0.93, 95% CI 0.92-0.94; specificity: 0.95, 95% CI 0.94-0.95). The positive likelihood, negative likelihood, and DOR were 22.45 (95% CI 12.86-39.19), 0.06 (95% CI 0.03-0.13), and 461.81 (95% CI 134.96-1580.24), respectively. Moreover, the DL model showed good performance for correctly stratifying normal patients, with an SROC of 0.99 (sensitivity: 0.95, 95% CI 0.94-0.96; specificity: 0.98, 95% CI 0.97-0.98). The positive likelihood, negative likelihood, and DOR were 47.47 (95% CI 20.70-108.86), 0.04 (95% CI 0.02-0.08), and 1524.81 (95% CI 625.29-3718.34), respectively.
Performance of Radiologists
Overview
A total of 3 studies compared the performance of DL models with radiologists [22,24,33]. Zhang et al [33] included 8 radiologists with 5 to 25 years of experience; they were categorized into two groups: junior radiologists had 5 to 15 years of experience and senior radiologists had 15 to 25 years of experience. Bai et al [24] compared DL model performance with 6 radiologists; 3 of them had 10 years of experience (ie, junior) and 3 had 20 years of experience (ie, senior). Finally, Ardakani et al [22] compared the performance of DL models with 1 senior radiologist, who had 15 years of experience. The performance of 15 radiologists in detecting COVID-19 was evaluated; the pooled sensitivity and specificity for detecting COVID-19 ranged from 0.75 to 0.89 and from 0.83 to 0.90, respectively. With the assistance of DL-based artificial intelligence (AI) tools, the performance of the junior radiologists improved: sensitivity improved by 0.08 to 0.15 and specificity improved by 0.05.
Sensitivity Analysis
A total of 8 studies evaluated the performance of DL algorithms for detecting COVID-19 using x-ray photographs. The pooled sensitivity and specificity of DL algorithms for detecting COVID-19 were 0.96 (95% CI 0.95-0.97) and 0.97 (95% CI 0.97-0.98), respectively, with an SROC of 0.99. Moreover, 8 studies assessed the performance of DL algorithms for classifying COVID-19 using CT images. The pooled sensitivity and specificity were 0.94 (95% CI 0.94-0.95) and 0.95 (95% CI 0.95-0.96), respectively, with an SROC of 0.96 (see Figures S1-S12 in Multimedia Appendix 1).
Risk of Bias and Applicability
In this meta-analysis, we also assessed heterogeneous findings that originated from included studies based on the QUADAS-2 tool (see Table 3 [13,14,20-33]). The risk of bias for patient selection was unclear for 16 studies. All studies had an unclear risk of bias for flow and timing and for the index test. Moreover, all studies had a high risk of bias for the reference standard. In the case of applicability, all studies had a low risk of bias for patient selection. However, the risk of index test and the applicability concern for the reference standard were uncertain.
Table 3.
Study | Risk of bias (high, low, or unclear) | Applicability concerns | ||||||
|
Patient selection | Index test | Reference standard | Flow and timing | Patient selection | Index test | Reference standard | |
Apostolopoulos and Mpesiana [14] | High | Unclear | High | Unclear | Low | Unclear | Unclear | |
Butt et al [13] | High | Unclear | High | Unclear | Low | Unclear | Unclear | |
Apostolopoulos et al [21] | High | Unclear | High | Unclear | Low | Unclear | Unclear | |
Li et al [25] | High | Unclear | High | Unclear | Low | Unclear | Unclear | |
Ucar and Korkmaz [29] | High | Unclear | High | Unclear | Low | Unclear | Unclear | |
Ozturk et al [26] | High | Unclear | High | Unclear | Low | Unclear | Unclear | |
Bai et al [24] | High | Unclear | High | Unclear | Low | Unclear | Unclear | |
Zhang at al [33] | High | Unclear | High | Unclear | Low | Unclear | Unclear | |
El Asnaoui and Chawki [20] | High | Unclear | High | Unclear | Low | Unclear | Unclear | |
Ardakani et al [22] | High | Unclear | High | Unclear | Low | Unclear | Unclear | |
Pathak et al [27] | High | Unclear | High | Unclear | Low | Unclear | Unclear | |
Wu et al [32] | High | Unclear | High | Unclear | Low | Unclear | Unclear | |
Toğaçar et al [28] | High | Unclear | High | Unclear | Low | Unclear | Unclear | |
Waheed at al [30] | High | Unclear | High | Unclear | Low | Unclear | Unclear | |
Khan et al [23] | High | Unclear | High | Unclear | Low | Unclear | Unclear | |
Wang et al [31] | High | Unclear | High | Unclear | Low | Unclear | Unclear |
Discussion
Principal Findings
In this study, we evaluated the performance of the DL model regarding detection of COVID-19 automatically using chest images to assist with proper diagnosis and prognosis. The findings of our study showed that the DL model achieved high sensitivity and specificity (95% and 96%, respectively) when detecting COVID-19. The pooled SROC value of both COVID-19 and other types of pneumonia was 98%. The performance of the DL model was comparable to that of experienced radiologists, whose clinical experience was at least 10 years, and the model could improve the performance of junior radiologists.
Clinical Implications
The rate of COVID-19 cases has been mounting day by day; therefore, it is important to quickly and accurately diagnose patients so that we may combat this pandemic. However, screening an increased number of chest images is challenging for the radiologists, and the number of trained radiologists is not sufficient, especially in underdeveloped and developing countries [34]. The recent success of DL applications in imaging analysis of CT scans, as well as x-ray imaging in automatic segmentation and classification in the radiology domain, has encouraged health care providers and researchers to exploit the advancement of deep neural networks in other applications [35]. DL models have been trained to assist radiologists in achieving higher interrater reliability during their years of experience in clinical practice.
Since the start of the COVID-19 pandemic, efforts have been made by AI researchers and AI modelers to help radiologists in the rapid diagnosis of COVID-19 in order to combat the COVID-19 pandemic [33,36]. Developing an accurate, automated AI COVID-19 detection tool is deemed as essential in reducing unnecessary waiting times, shortening screening and examination times, and improving performance. Moreover, such a tool could help to reduce radiologists’ workloads and allow them to respond to emergency situations rapidly and in a cost-effective manner [25]. RT-PCR is considered the gold standard detection method; however, findings of our study showed that chest CT could be used as a reliable and rapid approach for screening of COVID-19. Our findings also showed that the DL model was able to discriminate COVID-19 from other types of pneumonia with high a sensitivity and specificity, which is a challenging task for radiologists [32].
Strengths and Limitations
Our study has several strengths. First, this is the first meta-analysis that evaluated the performance of a DL model to classify COVID-19 patients. Second, we considered only peer-reviewed articles to be included in our study because articles that are not peer reviewed might contain bias. Third, we compared the performance of the DL model with that of senior and junior radiologists, which would be helpful for policy makers in considering an automated classification system in real-world clinical settings in order to speed up routine examination.
However, our study also has some limitations that need to be addressed. First, only 16 studies were used to evaluate the performance of the model; inclusion of more studies may have provided more specific findings. Second, some studies included similar data sets, which may have created some bias, but the researchers in those studies had optimized algorithms to improve performance. Third, two different kinds of digital photographs (ie, CT scan and x-ray) were used to develop and evaluate the performance of the DL model in classifying COVID-19; however, the performance of the DL model was almost the same in both cases. Finally, none of the studies included external validation; therefore, model performance could vary if those models were implemented in other clinical settings.
Future Perspective
The primary objectives of prediction models are the quick screening of COVID-19 patients and to help physicians make appropriate decisions. Misdiagnosis could have a destructive effect on society, as COVID-19 could spread from infected people to healthy people. Therefore, it is important to select a target population among which this automated tool could serve a clinical need; it is also important to select a representative data set on which the model could be trained, developed, and validated internally and externally. All the studies included in this meta-analysis had a high risk of bias for patient selection and reference standards. Moreover, generalizability was lacking in the newly developed classification models. Models without proper evidence and with a lack of external validation are not appropriate for clinical practice because they might cause more harm than good. Since the number of cases is mounting each day and COVID-19 is spreading to all continents, it is therefore important to develop a model to assist in the quick and efficient screening of patients during the COVID-19 pandemic. This could encourage clinicians and policy makers to prematurely implement prediction models without sufficient documentation and validation. All studies showed promising discrimination in their training, testing, and validation cohorts, but future studies should focus on external validation and comparing their findings to other data sets. Interpretability of DL systems is more important to a health care professional than to an AI expert. Proper interpretation and explanation of algorithms will more likely be acceptable to physicians. More clinical research is needed to determine the tangible benefits for patients in terms of the high performance of the model. High sensitivity and specificity do not necessarily represent clinical efficacy, and the higher value of the AUROC is not always the best metric to exhibit clinical applicability. All papers should follow standard guidelines and they should present positive and negative predictive values in order to be able to make a fair comparison. Although all of the included studies used a significant amount of data to compare model performance to that of the radiologists, they used only retrospective data to train the models, which might result in worse performance in real-world clinical settings, as data complexity is different. Therefore, prospective evaluation is needed in future studies before considering implementation in clinical settings. AI models always consist of potential flaws, including the inapplicability of new data, reliability, and bias. Generalization of the model is important for presenting the real performance because the rate of sensitivity and specificity varied across the studies (0.79 to 1.00 and 0.62 to 1.00, respectively). A higher number of false negatives will make the situation worse and will waste health care resources.
Conclusions
Our study showed that the DL model had immense potential to distinguish COVID-19 patients, with high sensitivity and specificity, from patients with other types of pneumonia and normal patients. DL-based tools could assist radiologists in the fast screening of COVID-19 and in classifying potential high-risk patients, which could have clinical significance for the early management of patients and could optimize medical resources. A higher number of false negatives could have a devastating effect on society; therefore, it is crucial to test the performance of models with other, unknown data sets. Retrospective evaluation and reliable interpretation are warranted to consider the application of AI models in real-world clinical settings.
Acknowledgments
This research is sponsored, in part, by the Ministry of Education (MOE) (grants MOE 108-6604-001-400 and DP2-109-21121-01-A-01) and the Ministry of Science and Technology (MOST) (grants MOST 108-2823-8-038-002- and 109-2222-E-038-002-MY2).
Abbreviations
- AI
artificial intelligence
- AUC
area under the curve
- AUROC
area under the receiver operating characteristic curve
- CT
computed tomography
- DL
deep learning
- DOR
diagnostic odds ratio
- MOE
Ministry of Education
- MOST
Ministry of Science and Technology
- PRISMA
Preferred Reporting Items for Systematic Reviews and Meta-Analyses
- QUADAS-2
Quality Assessment of Diagnostic Accuracy Studies-2
- ROC
receiver operating characteristic
- RT-PCR
reverse transcription–polymerase chain reaction
- SROC
summary receiver operating characteristic
Appendix
Supplementary figures.
Footnotes
Conflicts of Interest: None declared.
References
- 1.Sun J, He W, Wang L, Lai A, Ji X, Zhai X, Li G, Suchard MA, Tian J, Zhou J, Veit M, Su S. COVID-19: Epidemiology, evolution, and cross-disciplinary perspectives. Trends Mol Med. 2020 May;26(5):483–495. doi: 10.1016/j.molmed.2020.02.008. http://europepmc.org/abstract/MED/32359479. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 2.Park M, Cook AR, Lim JT, Sun Y, Dickens BL. A systematic review of COVID-19 epidemiology based on current evidence. J Clin Med. 2020 Mar 31;9(4):967. doi: 10.3390/jcm9040967. https://www.mdpi.com/resolver?pii=jcm9040967. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 3.The Novel Coronavirus Pneumonia Emergency Response Epidemiology Team The epidemiological characteristics of an outbreak of 2019 novel coronavirus diseases (COVID-19) — China, 2020. China CDC Wkly. 2020 Feb 14;2(8):113–122. doi: 10.46234/ccdcw2020.032. http://weekly.chinacdc.cn/en/article/doi/10.46234/ccdcw2020.032. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 4.Rothan HA, Byrareddy SN. The epidemiology and pathogenesis of coronavirus disease (COVID-19) outbreak. J Autoimmun. 2020 May;109:102433. doi: 10.1016/j.jaut.2020.102433. http://europepmc.org/abstract/MED/32113704. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 5.Lipsitch M, Swerdlow DL, Finelli L. Defining the epidemiology of Covid-19 — Studies needed. N Engl J Med. 2020 Mar 26;382(13):1194–1196. doi: 10.1056/nejmp2002125. [DOI] [PubMed] [Google Scholar]
- 6.Kojima N, Turner F, Slepnev V, Bacelar A, Deming L, Kodeboyina S, Klausner J D. Self-collected oral fluid and nasal swab specimens demonstrate comparable sensitivity to clinician-collected nasopharyngeal swab specimens for the detection of SARS-CoV-2. Clin Infect Dis. 2020 Oct 19;:ciaa1589. doi: 10.1093/cid/ciaa1589. http://europepmc.org/abstract/MED/33075138. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 7.Rhoads DD, Cherian SS, Roman K, Stempak LM, Schmotzer CL, Sadri N. Comparison of Abbott ID Now, DiaSorin Simplexa, and CDC FDA emergency use authorization methods for the detection of SARS-CoV-2 from nasopharyngeal and nasal swabs from individuals diagnosed with COVID-19. J Clin Microbiol. 2020 Jul 23;58(8):1–2. doi: 10.1128/jcm.00760-20. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8.Jones BP, Tay ET, Elikashvili I, Sanders JE, Paul AZ, Nelson BP, Spina LA, Tsung JW. Feasibility and safety of substituting lung ultrasonography for chest radiography when diagnosing pneumonia in children: A randomized controlled trial. Chest. 2016 Jul;150(1):131–138. doi: 10.1016/j.chest.2016.02.643. https://linkinghub.elsevier.com/retrieve/pii/S0012-3692(16)01263-0. [DOI] [PubMed] [Google Scholar]
- 9.Ye X, Xiao H, Chen B, Zhang S. Accuracy of lung ultrasonography versus chest radiography for the diagnosis of adult community-acquired pneumonia: Review of the literature and meta-analysis. PLoS One. 2015;10(6):e0130066. doi: 10.1371/journal.pone.0130066. https://dx.plos.org/10.1371/journal.pone.0130066. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 10.Bai HX, Hsieh B, Xiong Z, Halsey K, Choi JW, Tran TML, Pan I, Shi L, Wang D, Mei J, Jiang X, Zeng Q, Egglin TK, Hu P, Agarwal S, Xie F, Li S, Healey T, Atalay MK, Liao W. Performance of radiologists in differentiating COVID-19 from non-COVID-19 viral pneumonia at chest CT. Radiology. 2020 Aug;296(2):E46–E54. doi: 10.1148/radiol.2020200823. http://europepmc.org/abstract/MED/32155105. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 11.Ai T, Yang Z, Hou H, Zhan C, Chen C, Lv W, Tao Q, Sun Z, Xia L. Correlation of chest CT and RT-PCR testing for coronavirus disease 2019 (COVID-19) in China: A report of 1014 cases. Radiology. 2020 Aug;296(2):E32–E40. doi: 10.1148/radiol.2020200642. http://europepmc.org/abstract/MED/32101510. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 12.Shi H, Han X, Jiang N, Cao Y, Alwalid O, Gu J, Fan Y, Zheng C. Radiological findings from 81 patients with COVID-19 pneumonia in Wuhan, China: A descriptive study. Lancet Infect Dis. 2020 Apr;20(4):425–434. doi: 10.1016/s1473-3099(20)30086-4. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 13.Notice of retraction: Butt C, Gill J, Chun D, Babu BA. Deep learning system to screen coronavirus disease 2019 pneumonia. Appl Intell. 2020 Apr 22;:1–7. doi: 10.1007/s10489-020-01714-3. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14.Apostolopoulos ID, Mpesiana TA. Covid-19: Automatic detection from x-ray images utilizing transfer learning with convolutional neural networks. Phys Eng Sci Med. 2020 Jun;43(2):635–640. doi: 10.1007/s13246-020-00865-4. http://europepmc.org/abstract/MED/32524445. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15.Moher D, Liberati A, Tetzlaff J, Altman DG, PRISMA Group Preferred reporting items for systematic reviews and meta-analyses: The PRISMA statement. PLoS Med. 2009 Jul 21;6(7):e1000097. doi: 10.1371/journal.pmed.1000097. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 16.Whiting PF, Rutjes AWS, Westwood ME, Mallett S, Deeks JJ, Reitsma JB, Leeflang MMG, Sterne JAC, Bossuyt PMM, QUADAS-2 Group QUADAS-2: A revised tool for the quality assessment of diagnostic accuracy studies. Ann Intern Med. 2011 Oct 18;155(8):529–536. doi: 10.7326/0003-4819-155-8-201110180-00009. [DOI] [PubMed] [Google Scholar]
- 17.Moses LE, Shapiro D, Littenberg B. Combining independent studies of a diagnostic test into a summary ROC curve: Data-analytic approaches and some additional considerations. Stat Med. 1993 Jul 30;12(14):1293–1316. doi: 10.1002/sim.4780121403. [DOI] [PubMed] [Google Scholar]
- 18.Islam MM, Yang H, Poly TN, Jian W, Jack Li YC. Deep learning algorithms for detection of diabetic retinopathy in retinal fundus photographs: A systematic review and meta-analysis. Comput Methods Programs Biomed. 2020 Jul;191:105320. doi: 10.1016/j.cmpb.2020.105320. [DOI] [PubMed] [Google Scholar]
- 19.Walter SD. Properties of the summary receiver operating characteristic (SROC) curve for diagnostic test data. Stat Med. 2002 May 15;21(9):1237–1256. doi: 10.1002/sim.1099. [DOI] [PubMed] [Google Scholar]
- 20.El Asnaoui K, Chawki Y. Using x-ray images and deep learning for automated detection of coronavirus disease. J Biomol Struct Dyn. 2020 May 22;:1–12. doi: 10.1080/07391102.2020.1767212. http://europepmc.org/abstract/MED/32397844. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 21.Apostolopoulos ID, Aznaouridis SI, Tzani MA. Extracting possibly representative COVID-19 biomarkers from x-ray images with deep learning approach and image data related to pulmonary diseases. J Med Biol Eng. 2020 May 14;40:1–8. doi: 10.1007/s40846-020-00529-4. http://europepmc.org/abstract/MED/32412551. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 22.Ardakani AA, Kanafi AR, Acharya UR, Khadem N, Mohammadi A. Application of deep learning technique to manage COVID-19 in routine clinical practice using CT images: Results of 10 convolutional neural networks. Comput Biol Med. 2020 Jun;121:103795. doi: 10.1016/j.compbiomed.2020.103795. http://europepmc.org/abstract/MED/32568676. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 23.Khan AI, Shah JL, Bhat MM. CoroNet: A deep neural network for detection and diagnosis of COVID-19 from chest x-ray images. Comput Methods Programs Biomed. 2020 Nov;196:105581. doi: 10.1016/j.cmpb.2020.105581. http://europepmc.org/abstract/MED/32534344. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 24.Bai HX, Wang R, Xiong Z, Hsieh B, Chang K, Halsey K, Tran TML, Choi JW, Wang DC, Shi LB, Mei J, Jiang XL, Pan I, Zeng QH, Hu PF, Li YH, Fu FX, Huang RY, Sebro R, Yu QZ, Atalay MK, Liao WH. Artificial intelligence augmentation of radiologist performance in distinguishing COVID-19 from pneumonia of other origin at chest CT. Radiology. 2020 Sep;296(3):E156–E165. doi: 10.1148/radiol.2020201491. http://europepmc.org/abstract/MED/32339081. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 25.Li L, Qin L, Xu Z, Yin Y, Wang X, Kong B, Bai J, Lu Y, Fang Z, Song Q, Cao K, Liu D, Wang G, Xu Q, Fang X, Zhang S, Xia J, Xia J. Using artificial intelligence to detect COVID-19 and community-acquired pneumonia based on pulmonary CT: Evaluation of the diagnostic accuracy. Radiology. 2020 Aug;296(2):E65–E71. doi: 10.1148/radiol.2020200905. http://europepmc.org/abstract/MED/32191588. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 26.Ozturk T, Talo M, Yildirim EA, Baloglu UB, Yildirim O, Acharya UR. Automated detection of COVID-19 cases using deep neural networks with x-ray images. Comput Biol Med. 2020 Jun;121:103792. doi: 10.1016/j.compbiomed.2020.103792. http://europepmc.org/abstract/MED/32568675. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 27.Pathak Y, Shukla P, Tiwari A, Stalin S, Singh S, Shukla P. Deep transfer learning based classification model for COVID-19 disease. Ing Rech Biomed. 2020 May 20;:1–7. doi: 10.1016/j.irbm.2020.05.003. http://europepmc.org/abstract/MED/32837678. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 28.Toğaçar M, Ergen B, Cömert Z. COVID-19 detection using deep learning models to exploit social mimic optimization and structured chest x-ray images using fuzzy color and stacking approaches. Comput Biol Med. 2020 Jun;121:103805. doi: 10.1016/j.compbiomed.2020.103805. http://europepmc.org/abstract/MED/32568679. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 29.Ucar F, Korkmaz D. COVIDiagnosis-Net: Deep Bayes-SqueezeNet based diagnosis of the coronavirus disease 2019 (COVID-19) from x-ray images. Med Hypotheses. 2020 Jul;140:109761. doi: 10.1016/j.mehy.2020.109761. http://europepmc.org/abstract/MED/32344309. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 30.Waheed A, Goyal M, Gupta D, Khanna A, Al-Turjman F, Pinheiro PR. CovidGAN: Data augmentation using auxiliary classifier GAN for improved Covid-19 detection. IEEE Access. 2020;8:91916–91923. doi: 10.1109/access.2020.2994762. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 31.Wang S, Zha Y, Li W, Wu Q, Li X, Niu M, Wang M, Qiu X, Li H, Yu H, Gong W, Bai Y, Li L, Zhu Y, Wang L, Tian J. A fully automatic deep learning system for COVID-19 diagnostic and prognostic analysis. Eur Respir J. 2020 Aug;56(2):2000775. doi: 10.1183/13993003.00775-2020. http://erj.ersjournals.com:4040/lookup/pmidlookup?view=long&pmid=32444412. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 32.Wu X, Hui H, Niu M, Li L, Wang L, He B, Yang X, Li L, Li H, Tian J, Zha Y. Deep learning-based multi-view fusion model for screening 2019 novel coronavirus pneumonia: A multicentre study. Eur J Radiol. 2020 Jul;128:109041. doi: 10.1016/j.ejrad.2020.109041. https://linkinghub.elsevier.com/retrieve/pii/S0720-048X(20)30230-8. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 33.Zhang K, Liu X, Shen J, Li Z, Sang Y, Wu X, Zha Y, Liang W, Wang C, Wang K, Ye L, Gao M, Zhou Z, Li L, Wang J, Yang Z, Cai H, Xu J, Yang L, Cai W, Xu W, Wu S, Zhang W, Jiang S, Zheng L, Zhang X, Wang L, Lu L, Li J, Yin H, Wang W, Li O, Zhang C, Liang L, Wu T, Deng R, Wei K, Zhou Y, Chen T, Yiu-Nam Lau J, Fok M, He J, Lin T, Li W, Wang G. Clinically applicable AI system for accurate diagnosis, quantitative measurements, and prognosis of COVID-19 pneumonia using computed tomography. Cell. 2020 Sep 03;182(5):1360. doi: 10.1016/j.cell.2020.08.029. http://europepmc.org/abstract/MED/32888496. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 34.Liedenbaum MH, Bipat S, Bossuyt PMM, Dwarkasing RS, de Haan MC, Jansen RJ, Kauffman D, van der Leij C, de Lijster MS, Lute CC, van der Paardt MP, Thomeer MG, Zijlstra IA, Stoker J. Evaluation of a standardized CT colonography training program for novice readers. Radiology. 2011 Feb;258(2):477–487. doi: 10.1148/radiol.10100019. [DOI] [PubMed] [Google Scholar]
- 35.Ker J, Wang L, Rao J, Lim T. Deep learning applications in medical image analysis. IEEE Access. 2018;6:9375–9389. doi: 10.1109/access.2017.2788044. [DOI] [Google Scholar]
- 36.Hemdan E, Shouman M. COVIDX-Net: A framework of deep learning classifiers to diagnose COVID-19 in x-ray images. arXiv. Preprint posted online on March 24, 2020. https://arxiv.org/pdf/2003.11055. [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Supplementary Materials
Supplementary figures.