Skip to main content
PLOS ONE logoLink to PLOS ONE
. 2020 Jul 28;15(7):e0236621. doi: 10.1371/journal.pone.0236621

Deep transfer learning artificial intelligence accurately stages COVID-19 lung disease severity on portable chest radiographs

Jocelyn Zhu 1, Beiyi Shen 1, Almas Abbasi 1, Mahsa Hoshmand-Kochi 1, Haifang Li 1, Tim Q Duong 1,*
Editor: Dilbag Singh2
PMCID: PMC7386587  PMID: 32722697

Abstract

This study employed deep-learning convolutional neural networks to stage lung disease severity of Coronavirus Disease 2019 (COVID-19) infection on portable chest x-ray (CXR) with radiologist score of disease severity as ground truth. This study consisted of 131 portable CXR from 84 COVID-19 patients (51M 55.1±14.9yo; 29F 60.1±14.3yo; 4 missing information). Three expert chest radiologists scored the left and right lung separately based on the degree of opacity (0–3) and geographic extent (0–4). Deep-learning convolutional neural network (CNN) was used to predict lung disease severity scores. Data were split into 80% training and 20% testing datasets. Correlation analysis between AI-predicted versus radiologist scores were analyzed. Comparison was made with traditional and transfer learning. The average opacity score was 2.52 (range: 0–6) with a standard deviation of 0.25 (9.9%) across three readers. The average geographic extent score was 3.42 (range: 0–8) with a standard deviation of 0.57 (16.7%) across three readers. The inter-rater agreement yielded a Fleiss’ Kappa of 0.45 for opacity score and 0.71 for extent score. AI-predicted scores strongly correlated with radiologist scores, with the top model yielding a correlation coefficient (R2) of 0.90 (range: 0.73–0.90 for traditional learning and 0.83–0.90 for transfer learning) and a mean absolute error of 8.5% (ranges: 17.2–21.0% and 8.5%-15.5, respectively). Transfer learning generally performed better. In conclusion, deep-learning CNN accurately stages disease severity on portable chest x-ray of COVID-19 lung infection. This approach may prove useful to stage lung disease severity, prognosticate, and predict treatment response and survival, thereby informing risk management and resource allocation.

Introduction

Coronavirus Disease 2019 (COVID-19) is an infectious disease that can cause severe respiratory illness [1, 2]. First reported in Wuhan, China in December 2019 [3], COVID-19 was declared a pandemic on Mar 11, 2020 (https://www.who.int). More than 2 million people have been infected and more than 120,000 have died of COVID-19 (https://coronavirus.jhu.edu, Apr 14, 2020). These numbers are predicted to continue to rise in the foreseeable future with a high likelihood of a second wave and recurrence. COVID-19 has already overwhelmed many hospitals in many countries.

Portable chest x-rays (CXR) has become an indispensable imaging tool to imaging patients with contagious diseases because it is informative and the equipment can be readily disinfected. CXR provides both the extent and severity of lung infection, and is widely used to quantify disease severity in many lung diseases [4]. The hallmarks of COVID-19 lung infection on CXR include bilateral, peripheral hazy opacity and airspace consolidation [5]. Given the anticipated shortage of intensive care unit (ICU) beds and mechanical ventilators in many hospitals, CXR has the potential to play a critical role in decision-making to determine which patients to put on a mechanical ventilator, monitor disease progression and treatment effects during mechanical ventilation, and determine when it is safe to extubate. Similarly, computed tomography (CT), which offer better sensitivity than CXR, has been used in imaging suspected or confirmed COVID-19 patients, albeit mostly in China early on in the pandemic [69]. However, CT is less convenient, not practical in ICU setting, and prone to cross contamination and, thus, it is not widely used in COVID-19 infection or similar contagious disease in the United States.

Deep-learning artificial intelligence (AI) has become increasingly popular for analyzing diagnostic images [10, 11]. AI has the potential to facilitate disease diagnosis, diseases severity staging, and longitudinal monitoring of disease progression. One common machine-learning algorithm is the convolutional neural network (CNN) [12, 13], which takes an input image, learns important features in the image such as size or intensity, and saves these parameters as weights and bias to differentiate types of images [14, 15]. CNN architecture is ideally suited for analyzing images.

The majority of machine learning algorithms to date are trained to solve specific tasks, working in isolation. Models have to be rebuilt from scratch if the feature-space distribution changes. Transfer learning overcomes the isolated learning paradigm by utilizing knowledge acquired for one task to solve related ones. Transfer learning in AI is particularly important for small sample size data because the pre-trained weights enables more efficient training and better performance [16, 17].

There are many AI algorithms that have already been developed for CXR applications (see review [18]) and these AI algorithms can be repurposed to study COVID-19 lung infection. A few studies have already reported AI applications to classify COVID-19 versus non-COVID-19 lung infection using CXR [1922] and CT [2326]. Many of these referenced publications here are non-peer reviewed pre-prints. These studies aimed to classify COVID-19 versus non-COVID-19 images, not to stage lung disease severity. Classification of COVID-19 versus non-COVID-19 images for the purpose of diagnosis made by AI unfortunately has poor specificity. The reverse-transcription polymerase chain reaction test of a nasopharyngeal or oropharyngeal swab is still needed for definitive diagnosis. By contrast, deep-learning AI is well suited to stage disease severity. To our knowledge, there have been no studies to date that use deep-learning AI to stage disease severity on CXR of COVID-19 lung infection.

The goal of this study was to employ deep-learning CNN to stage disease severity of COVID-19 infection on CXR (as opposed to classify COVID-19 versus non-COVID-19 lung infection). The ground truths were the disease severity scores determined by expert chest radiologists. Comparison was made with traditional and transfer learning. This approach has the potential to provide frontline physicians an efficient and accurate means to triage patients, assess risk, allocate resources, as well as to monitor disease progression and treatment response. This approach is timely and urgent because of the anticipated widespread shortage of ICU beds and mechanical ventilators.

Materials and methods

Data sources

The diagram of patient selection and experimental design is shown in Fig 1. This is a retrospective study using publicly available de-identified data. Images were downloaded from https://github.com/ieee8023/covid-chestxray-dataset [27] on Mar 30, 2020. The original download contained 250 images of COVID-19 and SARS. Only anterior-posterior and posterior-anterior CXR from COVID-19 were included in this study, resulting in final sample size of 131 CXRs from 84 COVID-19 positive patients (51 males and 29 females; 4 with missing information). The mean and standard deviation of age was 55.1±14.9 years old for men and 60.1±14.3 years old for women.

Fig 1. Diagram of patient selection and experimental design.

Fig 1

Radiologist scoring

Radiological scoring is widely used to stage disease severity in many lung diseases [4]. To establish the disease severity score, a group of 6 chest radiologists with at least 20 years of experience worked together to reach consensus by evaluating two dozen images of portable CXRs of COVID-19 patients from Stony Brook University Hospital (these testing CXRs were not used in this study). Our scoring system was adapted from those by Warren et al. [28] and Wong et al [5]. Two board-certified chest radiologists with 20+ years of experience and one radiology resident reviewed image quality and scored the CXR for disease severity using the following criteria based on degree of opacity and geographical extent. The degree of opacity score of 0–3 was assigned to each of the right and left lung as: 0 = no opacity; 1 = ground glass opacity; 2 = consolidation; 3 = white-out. The right and left lung were scored separately and were added together. The geographical extent score of 0–4 was assigned to each of the right and left lung depending on the extent of involvement with ground glass opacity or consolidation: 0 = no involvement; 1 = <25%; 2 = 25–50%; 3 = 50–75%; 4 = >75% involvement. The right and left lung were scored separately and were added together. Sum of the two types of scores (0–14) and the product of the two types scores (0–48) were computed. The average of all radiologist scores were used for analysis.

CNN

A convolutional neural network (Fig 2A) [12, 13] with an additional regression layer was utilized to predict disease severity scores of CXR on a graded scale. The images were normalized, converted to RGB images, resized to 64x64 pixel images, and separated into 80% training and 20% testing datasets. This model architecture included Convolutional, Activation, Batch Normalization, Max Pooling, Dense layers, and a final Dense regression layer to fit the model to predict continuous values. Specifically, for convolutional neural networks, deep learning models are well-suited to learn complex functions, and optimize performance of regularization methods. Batch normalization layers served to stabilize the learning process by standardizing inputs, and standard rectified linear activation layers for CNNs were used to optimize performance. In general, the number of layers and nodes depends heavily on the input data and prediction goals, therefore heuristics were used to generate a standard CNN model for image analysis, the model was then fine-tuned based on performance on the testing dataset. This resulted in three 3x3 kernel sized convolutional layers. L2 regularization and Dropout layers, using a parametric probability of 0.3 as determined by analyzing performance results, were used to prevent overfitting. An optimal batch size of 8 was determined by five-fold cross validation, and the model was trained for 110 epochs. The loss function was measured by mean squared error as the model predicted continuous values, and the learning rate was set to 0.001, as it proved to perform more positively than the standard 0.01 rate in early training, and the Adam optimizer proved the most successful in minimizing loss.

Fig 2.

Fig 2

CNN Model architecture for traditional training (A) and transfer learning method (B). Model takes in a CXR image as an input. The outputs of the prediction were different severity scores.

Transfer learning

Given the limited dataset, a transfer learning method was also explored to optimize prediction. A VGG16 model was loaded with sample weights trained off the ImageNet dataset [29]. To be compatible with VGG16, the data was normalized, converted to RGB images, and resized into 224x224 pixel images. The dataset was also split into 80% training and 20% testing. Only the top layers were trained to prevent overfitting, and all convolutional layers were frozen. Standard VGG16 architecture was utilized, with the exception of the fully connected layers (Fig 2B). Once out of the convoluted layers, the model output was flattened, then passed through three Dense layers, including a regression Dense layer. For this transfer learning method, cross validation revealed an optimal batch size of 8 and only 10 epochs were necessary to train the data, as opposed to the 110 needed for the traditional training method. Adam optimizer and 0.001 learning rate were similarly employed in this model.

Statistical methods and performance evaluation

The means and standard deviations of the radiologist scores, and the Fleiss’ Kappa inter-rater agreement were calculated. Correlation analysis between AI-predicted versus radiologist severity scores were analyzed with slopes, intercepts, correlation coefficients (R2), p-values, and mean absolute errors (MAE) reported. Results using traditional and transfer learning were compared. All results are reported as means ± standard deviations. A p<0.05 was taken to be statistically significant unless otherwise specified. Note that the prediction of graded values precluded receiver operating curve (ROC) analysis.

Results

The final sample size consisted of 131 CXR from 84 COVID-19 positive patients (51M and 29F; 4 missing information). The men were on average 55.1±14.9 years old and women were 60.1±14.3 years old. CXR examples demonstrating a range of severity scores which include extent and opacity score are illustrated in Fig 3. CXR of COVID-19 positive patients showed hazy opacities and/or airspace consolidation. It has a predominance of bilateral, peripheral and lower lung zone distribution. The results of three raters are summarized in Table 1. The mean opacity score was 2.52 (0–6), with a standard deviation of 0.25 (9.9%) across three readers. The mean geographic extent score was 3.42 (a range from 0 to 8), with a standard deviation of 0.57 (16.7%) across three readers. The low standard deviations across three raters suggest good agreement. Fleiss’ Kappa for inter-rater agreement was 0.45 (z = 24.4) for opacity score and 0.71 (z = 29.5) for geographic extent score, indicating good to excellent agreement. The mean of sum of scores was 5.84 (0–14) and mean of product of scores was 11.86 (0–48). Fig 4 shows the histograms of radiologist scores. The product score was slightly skewed toward lower values because numbers multiplied by a number close to zero tended to be of lower values.

Fig 3. Examples showing a range of severity scores which include extent and opacity scores.

Fig 3

The two numbers for each are for the right and left lung. Arrows indicate region of Ground glass opacities or/and airspace consolidations.

Table 1. Individual, sum and product of degree of opacity scores and geographic extent scores.

Values in parentheses are ranges for the corresponding scores.

Geographic extent score (0–8) Opacity score (0–6) Sum (0–14) Product (0–48)
Mean 3.41 2.52 5.84 11.86
Standard deviation across 3 readers 0.78 0.25 na na
Fleiss’ κ 0.45 0.71 na na

Fig 4. Histograms of radiologist severity scores.

Fig 4

G: geographic extent, O: degree of opacity, S: sum of the two score types, P: product of the two score types.

The data were separated into 80% training and 20% testing datasets. The loss function decreased and prediction accuracy improved with increasing epochs for both training and validation datasets. For the traditional training method, the loss function typically converged at training epochs of 30–50 epochs, and five-fold validation confirmed optimal hyperparameters of training on 110 epochs. For the transfer learning method employing a VGG16 model, 10 epochs were sufficient to train the model with high predictive ability; more epochs led to overfitting and increased computational time.

Fig 5 shows the AI-predicted scores versus radiologist severity scores of the “test” dataset, obtained for different types of scores using traditional and transfer learning. Overall, there were strong correlations between the AI-predicted and radiologist scores. The slopes, intercepts, R2, p values, and mean absolute errors (MAE) with traditional and transfer learning are summarized in Table 2. The slopes were generally (but not consistently across different scores) closer to unity, the intercepts were generally closer to zero, and R2 are generally closer to unity for the transfer learning compared to traditional learning. The top model yielded a R2 = 0.90 (range: 0.73–0.90 for traditional learning and 0.83–0.90 for transfer learning). The p-values for the transfer learning were consistently smaller than to those of the traditional learning.

Fig 5.

Fig 5

Scatterplots of AI-predicted versus radiologist severity scores on CXR using (A) traditional training method by CNN and (B) transfer learning method by CNN. G: geographic extent, O: degree of opacity, S: sum of the two score types, P: product of the two score types. The lines are linear regressions. See Table 2 for slopes, intercepts, R2, and p values.

Table 2. Summary of slope, intercepts, R2, and p values of the correlation analysis, and the Mean Absolute Error (MAE) of AI-predicted and radiologist scores with traditional and transfer learning.

The % was obtained by dividing MAE by the corresponding maximum score.

Slope Intercept R2 p-value MAE (%)
Traditional learning Geographic extent score (0–8) 0.73 2.16 0.73 0.02 1.51 (18.9%)
Opacity score (0–6) 0.77 0.07 0.90 0.001 1.26 (21.0%)
Sum (0–14) 0.67 0.71 0.76 0.0003 2.57(18.4%)
Product (0–48) 0.58 -1.60 0.88 0.06 8.26 (17.2%)
Transfer learning Geographic extent score (0–8) 0.60 0.99 0.83 <0.0001 0.91(11.4%)
Opacity score (0–6) 0.84 1.48 0.87 <0.0001 0.93 (15.5%)
Sum (0–14) 0.68 0.45 0.90 0.001 2.12 (15.1%)
Product (0–48) 0.80 1.37 0.85 <0.0001 4.10 (8.5%)

The MAEs for the transfer learning are consistently smaller than to those of the traditional learning. The top model yielded a MAE of 8.5% (range: 17.2% to 21.3% for traditional learning, and 8.5% to 15.5% for transfer learning). Transfer learning generally performed better than those traditional learning.

Discussion

This study tested the hypothesis that deep-learning convolutional neural networks accurately stage disease severity on portable chest x-rays using radiologists’ severity scores as ground truths associated with COVID-19 lung infection. Disease severity scores by three radiologists yielded good inter-rater agreement. AI-predicted scores were highly correlated with radiologist scores. Transfer learning improved efficiency and shorten computational time without compromising performance.

Although there are already over 5500 publications on COVID-19 on PubMed (keyword “COVID-19” Apr 21, 2020), only a few studies have reported AI applications to classify COVID-19 versus non-COVID-19 lung infection on CXR [1922] and CT [2326]. All of these studies (including many non-peer-reviewed pre-prints cited here) aimed to classify COVID-19 versus non-COVID-19 images, not to stage disease severity. A major challenge for AI in the application to disease diagnosis is that training datasets were limited to a few lung diseases, resulting in low generalizability. Amongst these studies, most compare CXR of COVID-19 infection with those of bacterial pneumonia and/or normal CXR. More importantly, AI diagnosis of disease in general, and COVID-19 infection in particular, based on CXR has poor specificity because many pathologies have similar radiographic appearance as other infections and diseases on CXR. Definitive diagnosis of COVID-19 infection still requires a test by reverse transcription polymerase chain reaction of a nasopharyngeal or oropharyngeal swab.

Our study differs from these previous studies in that we used AI of CXR to stage disease severity on a graded scale in positive COVID-19 patients. Deep-learning AI, specifically a convolutional neural network, is well suited to extract information from CXR and stage disease severity by training using chest radiologist determination of disease severity scores. We evaluated individual and combination scores of severities as it is unknown which type of scores more accurately reflect lung infection disease severity. Our findings showed that most scores yielded similar correlation coefficients. One possible explanation is that both the opacity score and geographic extent yielded similar information regarding disease severity. Further studies are needed to determine which is superior. It is worth noting that performance by receiver-operating curve (ROC) analysis (such as area under the curve (AUC), accuracy, sensitivity and specificity) cannot be used for continuous variables. Instead, correlation analysis and mean square error analysis were performed.

There were strong correlations between the AI-predicted and radiologist scores, with the top model yielding a R2 = 0.90 and a mean absolute error of 8.5%. This is remarkable performance. The ideal intercept should be zero and ideal slope should be unity. The intercepts were generally close to zero. However, all the slopes were consistently below the line of unity. A possible explanation is that the score distributions were skewed by low score values. As the models attempted to minimize loss, it tended to underestimate higher scores, resulting in slope less than unity. Other explanations are possible. Further studies are needed.

Another novelty of our study is that we compared traditional and transfer learning. Transfer learning approach is expected to yield good performance with small datasets. Most of the performance indices (such as R2, p values, and MAEs) were better with transfer learning compared to traditional learning. The training time with transferring learning was shorter without compromising performance.

Although radiology reports are informative, they are usually qualitative. A quantitative score of disease severity of portable CXR afforded by AI in an accurate and efficient manner should prove useful in clinical settings of COVID-19 circumstance. With CXR being an indispensable imaging tool in the management of COVID-19 lung infection, radiologists need an efficient and accurate means to stage disease severity and monitor disease progression and treatment response. Furthermore, with the anticipated widespread shortage of ICU beds and ventilators, frontline physicians need an efficient and accurate means to triage patients, assess risk and allocate resources. This AI approach to stage disease severity is ideally suited to tackle these challenges associated with the COVID-19 pandemics or the like.

This proof-of-concept pilot study has several limitations. First, this retrospective study was conducted on an open multi-institutional dataset with a small sample size. This topic is timely and only small dataset are currently available at the time of this writing. These findings need to be replicated on larger sample size with multi-institutional data. It may also be of important to investigate multiple longitudinal time points to evaluate disease progression. Second, we did not study the correlation of radiographic score severity with clinical disease severity or clinical outcome which were unavailable. Although radiographic score has been widely used as a surrogate marker of disease severity in a variety of lung diseases (see review [4]), it is unknown at this time if radiographic scoring reflects functional or clinical outcome in the case of COVID-19. Third, a variety of radiographic scoring systems has been used in other studies [5, 28], each has its advantages and disadvantages. A more sophisticated score system could be explored. Future studies should also consider incorporating non-imaging data (such as demographics, co-morbidities, vitals, blood biomarkers).

In conclusion, deep-learning convolutional neural networks accurately stage lung disease severity on portable chest x-rays associated with COVID-19 lung infection. This approach has the potential to be used to prognosticate, stage disease severity, monitor disease progression and treatment response, which in turn could inform risk management and resource allocation associated with the COVID-19 pandemic.

Data Availability

The complete dataset can be found at www.kaggle.com/dataset/89ab2075b0a23026a0a1dda064441a6d05d38dc1e71b5d04d2c3e5a0196b89a8.

Funding Statement

unfunded studies.

References

  • 1.Hui DS, IA E, Madani TA, Ntoumi F, Kock R, Dar O, et al. The continuing 2019-nCoV epidemic threat of novel coronaviruses to global health—The latest 2019 novel coronavirus outbreak in Wuhan, China. Int J Infect Dis. 2020;91:264–6. Epub 2020/01/19. 10.1016/j.ijid.2020.01.009 . [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 2.Lu H, Stratton CW, Tang YW. Outbreak of pneumonia of unknown etiology in Wuhan, China: The mystery and the miracle. J Med Virol. 2020;92(4):401–2. Epub 2020/01/18. 10.1002/jmv.25678 . [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3.Li Q, Guan X, Wu P, Wang X, Zhou L, Tong Y, et al. Early Transmission Dynamics in Wuhan, China, of Novel Coronavirus-Infected Pneumonia. N Engl J Med. 2020;382(13):1199–207. Epub 2020/01/30. 10.1056/NEJMoa2001316 . [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4.Taylor E, Haven K, Reed P, Bissielo A, Harvey D, McArthur C, et al. A chest radiograph scoring system in patients with severe acute respiratory infection: a validation study. BMC Med Imaging. 2015;15:61 Epub 2015/12/31. 10.1186/s12880-015-0103-y [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5.Wong HYF, Lam HYS, Fong AH, Leung ST, Chin TW, Lo CSY, et al. Frequency and Distribution of Chest Radiographic Findings in COVID-19 Positive Patients. Radiology. 2019:201160 Epub 2020/03/29. 10.1148/radiol.2020201160 . [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6.Salehi S, Abedi A, Balakrishnan S, Gholamrezanezhad A. Coronavirus Disease 2019 (COVID-19): A Systematic Review of Imaging Findings in 919 Patients. AJR Am J Roentgenol. 2020:1–7. Epub 2020/03/17. 10.2214/AJR.20.23034 . [DOI] [PubMed] [Google Scholar]
  • 7.Li K, Wu J, Wu F, Guo D, Chen L, Fang Z, et al. The Clinical and Chest CT Features Associated with Severe and Critical COVID-19 Pneumonia. Invest Radiol. 2020. Epub 2020/03/03. 10.1097/RLI.0000000000000672 . [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.Wang Y, Dong C, Hu Y, Li C, Ren Q, Zhang X, et al. Temporal Changes of CT Findings in 90 Patients with COVID-19 Pneumonia: A Longitudinal Study. Radiology. 2020:200843 Epub 2020/03/20. 10.1148/radiol.2020200843 . [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9.Chung M, Bernheim A, Mei X, Zhang N, Huang M, Zeng X, et al. CT Imaging Features of 2019 Novel Coronavirus (2019-nCoV). Radiology. 2020;295(1):202–7. Epub 2020/02/06. 10.1148/radiol.2020200230 . [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10.LeCun Y, Bengio Y, Hinton G. Deep learning. Nature. 2015;521(7553):436–44. Epub 2015/05/29. 10.1038/nature14539 . [DOI] [PubMed] [Google Scholar]
  • 11.Ehteshami Bejnordi B, Mullooly M, Pfeiffer RM, Fan S, Vacek PM, Weaver DL, et al. Using deep convolutional neural networks to identify and classify tumor-associated stroma in diagnostic breast biopsies. Mod Pathol. 2018;31(10):1502–12. Epub 2018/06/15. 10.1038/s41379-018-0073-z . [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12.Krizhevsky A, Sutskever I, Hinton GE. Imagenet classification with deep convolutional neural networks. Advances in neural information processing systems. 2012:1097–105. [Google Scholar]
  • 13.Hinton GE, Srivastava N, Krizhevsky A, Sutskever I, Salakhutdinov RR. Improving neural networks by preventing co-adaptation of feature detectors. arxiv: 12070580v1. 2012. [Google Scholar]
  • 14.Simonyan K, Zisserman A. Very deep convolutional networks for large-scale image recognition. ICLR; 2015. [Google Scholar]
  • 15.Arel I, Rose DC, Karnowski TP. Deep machine learning-a new frontier in artificial intelligence research. IEEE Computational Intelligence Magazine. 2010;5(4):13–8. [Google Scholar]
  • 16.Samala RK, Chan HP, Hadjiiski LM, Helvie MA, Cha KH, Richter CD. Multi-task transfer learning deep convolutional neural network: application to computer-aided diagnosis of breast cancer on mammograms. Phys Med Biol. 2017;62(23):8894–908. Epub 2017/10/17. 10.1088/1361-6560/aa93d4 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17.Byra M, Styczynski G, Szmigielski C, Kalinowski P, Michalowski L, Paluszkiewicz R, et al. Transfer learning with deep convolutional neural network for liver steatosis assessment in ultrasound images. Int J Comput Assist Radiol Surg. 2018;13(12):1895–903. Epub 2018/08/11. 10.1007/s11548-018-1843-2 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18.Harris M, Qi A, Jeagal L, Torabi N, Menzies D, Korobitsyn A, et al. A systematic review of the diagnostic accuracy of artificial intelligence-based computer programs to analyze chest x-rays for pulmonary tuberculosis. PLoS One. 2019;14(9):e0221339 Epub 2019/09/04. 10.1371/journal.pone.0221339 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19.Hall LO, Paul R, Goldgof DB, & Goldgof GM. Finding covid-19 from chest x-rays using deep learning on a small datase. arXiv:200402060. [Google Scholar]
  • 20.Rahimzadeh M, Attar A. A modified deep convolutional neural network for detecting COVID-19 and pneumonia from chest X-ray images based on the concatenation of Xception and ResNet50V2. Informatics in Medicine Unlocked. 2020;19:100360 10.1016/j.imu.2020.100360 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21.Apostolopoulos ID, Mpesiana TA. Covid-19: automatic detection from X-ray images utilizing transfer learning with convolutional neural networks. Physical and Engineering Sciences in Medicine. 2020, in press. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22.Narin A, Kaya C, Pamuk Z. Automatic Detection of Coronavirus Disease (COVID-19) Using X-ray Images and Deep Convolutional Neural Networks. arXiv:200310849. 2020. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 23.Singh D, Kumar V, Kaur M. Classification of COVID-19 patients from chest CT images using multi-objective differential evolution–based convolutional neural networks. European Journal of Clinical Microbiology & Infectious Diseases. 2020, in press. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 24.Dadario AMV, Paiva JPQ, Chate RC, Machado BS, Szarf G. Regarding "Artificial Intelligence Distinguishes COVID-19 from Community Acquired Pneumonia on Chest CT". Radiology. 2020:201178 Epub 2020/04/04. 10.1148/radiol.2020201178 . [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25.Li L, Qin L, Xu Z, Yin Y, Wang X, Kong B, et al. Artificial Intelligence Distinguishes COVID-19 from Community Acquired Pneumonia on Chest CT. Radiology. 2020:200905 Epub 2020/03/20. 10.1148/radiol.2020200905 . [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 26.Wang L, Wong A. COVID-Net: A Tailored Deep Convolutional Neural Network Design for Detection of COVID-19 Cases from Chest Radiography Images. arXiv:200309871v1. 2020. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27.Cohen JP, Morrison P, Dao L. COVID-19 Image Data Collection. arXiv:200311597v1. 2020. [Google Scholar]
  • 28.Warren MA, Zhao Z, Koyama T, Bastarache JA, Shaver CM, Semler MW, et al. Severity scoring of lung oedema on the chest radiograph is associated with clinical outcomes in ARDS. Thorax. 2018;73(9):840–6. Epub 2018/06/16. 10.1136/thoraxjnl-2017-211280 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 29.Simonyan K, Zisserman A. Very Deep Convolutional Networks for Large-Scale Image Recognition. arXiv:14091556. 2014. [Google Scholar]

Decision Letter 0

Dilbag Singh

5 Jun 2020

PONE-D-20-16042

Deep learning artificial intelligence accurately stages COVID-19 disease severity on portable chest radiographs

PLOS ONE

Dear Dr. Duong,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

Please submit your revised manuscript by Jul 20 2020 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: http://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols

We look forward to receiving your revised manuscript.

Kind regards,

Dilbag Singh

Academic Editor

PLOS ONE

Additional Editor Comments:

I think authors should add more recent literature from the PLOS One journal.

Journal requirements:

When submitting your revision, we need you to address these additional requirements.

1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at

https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and

https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf

2. We note that you have included the phrase “data not shown” in your manuscript. Unfortunately, this does not meet our data sharing requirements. PLOS does not permit references to inaccessible data. We require that authors provide all relevant data within the paper, Supporting Information files, or in an acceptable, public repository. Please add a citation to support this phrase or upload the data that corresponds with these findings to a stable repository (such as Figshare or Dryad) and provide and URLs, DOIs, or accession numbers that may be used to access these data. Or, if the data are not a core part of the research being presented in your study, we ask that you remove the phrase that refers to these data.

3. We note that you have stated that you will provide repository information for your data at acceptance. Should your manuscript be accepted for publication, we will hold it until you provide the relevant accession numbers or DOIs necessary to access your data. If you wish to make changes to your Data Availability statement, please describe these changes in your cover letter and we will update your Data Availability statement to reflect the information you provide.

4. Thank you for stating the following in your Competing Interests section: 'No'

a. Please complete your Competing Interests statement to state any Competing Interests. If you have no competing interests, please state "The authors have declared that no competing interests exist.", as detailed online in our guide for authors at http://journals.plos.org/plosone/s/submit-now

 b. This information should be included in your cover letter; we will change the online submission form on your behalf.

Please know it is PLOS ONE policy for corresponding authors to declare, on behalf of all authors, all potential competing interests for the purposes of transparency. PLOS defines a competing interest as anything that interferes with, or could reasonably be perceived as interfering with, the full and objective presentation, peer review, editorial decision-making, or publication of research or non-research articles submitted to one of the journals. Competing interests can be financial or non-financial, professional, or personal. Competing interests can arise in relationship to an organization or another person. Please follow this link to our website for more details on competing interests: http://journals.plos.org/plosone/s/competing-interests

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

Reviewer #2: Yes

Reviewer #3: Yes

**********

2. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

Reviewer #2: Yes

Reviewer #3: Yes

**********

3. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: Yes

Reviewer #3: Yes

**********

4. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: No

Reviewer #3: Yes

**********

5. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: 1. There are some grammatical and typo mistakes in the entire manuscript such as:

“predict continuous values. (Figure 1).”

described by Wong et al (7)

2. Use consistent words such as either COVID19 or COVID-19 in entire manuscript.

3. Please write suitable background of Deep-learning convolutional neural network (CNN) and deep transfer learning?

4. Add more mathematical background of CNN.

5. Add suitable diagrammatic flow in the paper.

6. Authors have use CNN or deep transfer learning? If deep transfer learning is used, then title should be improved.

7. Recently, many studies have been used for diagnosis of COVID-19, so, I suggest authors to add a related work about those studies such as: 10.1148/radiol.2020200432,10.1007/s10096-020-03901-z, 10.1142/S0217984919500222, 10.1049/iet-syb.2019.0116, 10.1148/radiol.2020200642, 10.1148/radiol.2020200463, 10.1148/radiol.2020200370

8. Do improve abstract such as kappa is simply used as k. Also, authors should use present sentences rather than was were.

9. Figure 1 should be improved it seems directly copied from some source.

Reviewer #2: In this paper deep-learning convolutional neural networks has been implemented for novel COVID-19 . Problem has a great significance. The content is written well and the results are promising. The paper needs some improvement in terms of the following recommendations:

1. Main objective need to be explained in a specific way.

2. Introduction is week, need to include some more similar modeling approaches and related work on other nonlinear modeling approaches should be extended with the following papers:

Apostolopoulos, I. D., & Mpesiana, T. A. (2020). Covid-19: automatic detection from x-ray images utilizing transfer learning with convolutional neural networks. Physical and Engineering Sciences in Medicine, 1.

Singh, D., Kumar, V., & Kaur, M. (2020). Classification of COVID-19 patients from chest CT images using multi-objective differential evolution–based convolutional neural networks. European Journal of Clinical Microbiology & Infectious Diseases, 1-11.

Rahimzadeh, M., & Attar, A. (2020). A modified deep convolutional neural network for detecting COVID-19 and pneumonia from chest X-ray images based on the concatenation of Xception and ResNet50V2. Informatics in Medicine Unlocked, 100360.

Hall, L. O., Paul, R., Goldgof, D. B., & Goldgof, G. M. (2020). Finding covid-19 from chest x-rays using deep learning on a small dataset. arXiv preprint arXiv:2004.02060.

3. Methodology need to be written in more detail with respect of each step involved. Statistical methods and performance evaluation is also required more information.

4. Results need to be compared with exiting works.

Reviewer #3: This paper is nicely written and idea is novel, but I have following recommendations

The choice of the number of layers and neurons per layer given in section III A is not argued.

The database is weak to draw a conclusion.

Taking into account the uncertainties given in the various analyzes, the results are comparable. In addition, the efficiencies must be revised downwards.

Please explain how this manuscript advances this field of research and/or contributes something new to the literature.

**********

6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

Reviewer #3: No

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

PLoS One. 2020 Jul 28;15(7):e0236621. doi: 10.1371/journal.pone.0236621.r002

Author response to Decision Letter 0


10 Jun 2020

Pls see attachment for easier read.

Dear Reviewers,

We thank you for your constructive comments. Please find the attached point-by-point response below. We believe we have addressed all comments positively. We have also carefully edited throughout. We have revised/removed statement of “data not shown.” Note that we also added the mean absolute error as another measure.

Reviewer #1:

1. There are some grammatical and typo mistakes in the entire manuscript such as:

“predict continuous values. (Figure 1).”

described by Wong et al (7)

Thank you. Corrected. It was incorrectly stated, it should have been Wang et al.

2. Use consistent words such as either COVID19 or COVID-19 in entire manuscript.

Thank you. Corrected.

3. Please write suitable background of Deep-learning convolutional neural network (CNN) and deep transfer learning?

The following is added.

Deep-learning artificial intelligence (AI) has become increasingly popular for analyzing diagnostic images [10, 11]. AI has the potential to facilitate disease diagnosis of disease, diseases severity staging, and longitudinal monitoring of disease progression. One common machine-learning algorithm is the convolutional neural network (CNN) [12, 13], which takes an input image, learns important features in the image such as size or intensity, and saves these parameters as weights and bias to differentiate types of images [14, 15]. CNN architecture is ideally suited for analyzing images.

The majority of machine learning algorithms to date are trained to solve specific tasks, working in isolation. Models have to be rebuilt from scratch if the feature-space distribution changes. Transfer learning overcomes the isolated learning paradigm by utilizing knowledge acquired for one task to solve related ones. Transfer learning in AI is particularly important for small sample size data because the pre-trained weights enables more efficient training and better performance [16, 17].

There are many AI algorithms that have already been developed for CXR applications (see review [18]) and these AI algorithms can be repurposed to study COVID-19 lung infection. A few studies have already reported AI applications to classify COVID-19 versus non-COVID-19 lung infection using CXR [19-22] and CT [23-26]. Many of these publications are pre-print, non-peer reviewed publications. These studies aimed to classify (not stage) COVID-19 versus non-COVID-19 images. Classification (i.e., for the purpose of diagnosis) of COVID-19 versus non-COVID-19 made by AI unfortunately has poor specificity. The reverse-transcription polymerase chain reaction test of a nasopharyngeal or oropharyngeal swab is still needed for definitive diagnosis. By contrast, deep-learning AI is well suited to stage disease severity. To our knowledge, there have been no studies to date that use deep-learning AI to stage disease severity on CXR of COVID-19 lung infection.

The goal of this study was to employ deep-learning CNN to stage disease severity of COVID-19 infection on CXR (as opposed to classify COVID-19 versus non-COVID-19 lung infection). The ground truths were the disease severity scores determined by expert chest radiologists. Comparison was made with traditional and transfer learning. This approach has the potential to provide frontline physicians an efficient and accurate means to triage patients, assess risk, allocate resources, as well as to monitor disease progression and treatment response. This approach is timely and urgent because of the anticipated widespread shortage of ICU beds and mechanical ventilators.

4. Add more mathematical background of CNN.

This paper deals with an innovative CNN application to stage lung disease severity based on portable CXRs of COVID-19 patients. The CNN algorithm and transferring methods are standard. We feel that to add a mathematical background of CNN would be repeating as the literature has already described it in a much more eloquent and precise manner. We only provide a brief qualitative description and cited references instead.

5. Add suitable diagrammatic flow in the paper.

A flowchart diagram for analysis is added.

6. Authors have use CNN or deep transfer learning? If deep transfer learning is used, then title should be improved.

The new title is:

Deep transfer learning methods in artificial intelligence accurately stages COVID-19 lung disease severity on portable chest radiographs

7. Recently, many studies have been used for diagnosis of COVID-19, so, I suggest authors to add a related work about those studies such as: 10.1148/radiol.2020200432, 10.1007/s10096-020-03901-z, 10.1142/S0217984919500222, 10.1049/iet-syb.2019.0116, 10.1148/radiol.2020200642, 10.1148/radiol.2020200463, 10.1148/radiol.2020200370

Thank you for the references. They are cited. Our study employed deep learning to stage disease severity of COVID-19 infection on CXR (as opposed to classify COVID-19 versus non-COVID-19 lung infection that are in the literature). To our knowledge, there have been no studies to date that use deep-learning AI to stage disease severity on CXR of COVID-19 lung infection.

8. Do improve abstract such as kappa is simply used as k.

Also, authors should use present sentences rather than was were.

Thank you. ��is used more often in the clinical literature so we spelled it out instead. We changed to present tense where appropriate but past tense is used more often in the clinical literature.

9. Figure 1 should be improved it seems directly copied from some source.

This figure was created by us. It was not copied.

Reviewer #2: In this paper deep-learning convolutional neural networks has been implemented for novel COVID-19. Problem has a great significance. The content is written well and the results are promising. The paper needs some improvement in terms of the following recommendations:

1. Main objective need to be explained in a specific way.

Thank you. Pls see the response to Reviewer 1, comment #3 for the extensively revised Introduction.

2. Introduction is week, need to include some more similar modeling approaches and related work on other nonlinear modeling approaches should be extended with the following papers:

Apostolopoulos, I. D., & Mpesiana, T. A. (2020). Covid-19: automatic detection from x-ray images utilizing transfer learning with convolutional neural networks. Physical and Engineering Sciences in Medicine, 1.

Singh, D., Kumar, V., & Kaur, M. (2020). Classification of COVID-19 patients from chest CT images using multi-objective differential evolution–based convolutional neural networks. European Journal of Clinical Microbiology & Infectious Diseases, 1-11.

Rahimzadeh, M., & Attar, A. (2020). A modified deep convolutional neural network for detecting COVID-19 and pneumonia from chest X-ray images based on the concatenation of Xception and ResNet50V2. Informatics in Medicine Unlocked, 100360.

Hall, L. O., Paul, R., Goldgof, D. B., & Goldgof, G. M. (2020). Finding covid-19 from chest x-rays using deep learning on a small dataset. arXiv preprint arXiv:2004.02060.

Pls see the response to Reviewer 1, comment #3 for the extensively revised Introduction.

Thank you for the references. They are cited.

3. Methodology need to be written in more detail with respect of each step involved. Statistical methods and performance evaluation is also required more information.

Thank you. We have expanded the Methods in multiple places and included a flowchart.

4. Results need to be compared with exiting works.

Thank you. We expanded the discussion of the results generally. There are currently no studies to date that use deep-learning AI to stage disease severity on CXR of COVID-19 lung infection. There are no similar works to compare quantitatively.

Reviewer #3: This paper is nicely written and idea is novel, but I have following recommendations

The choice of the number of layers and neurons per layer given in section III A is not argued.

Thank you. The following is added.

CNN: A convolutional neural network (Figure 2A)[12, 13] with an additional regression layer was utilized to predict disease severity scores of CXR on a graded scale. The images were normalized, converted to RGB images, resized to 64x64 pixel images, and separated into 80% training and 20% testing datasets. This model architecture included Convolutional, Activation, Batch Normalization, Max Pooling, Dense layers, and a final Dense regression layer to fit the model to predict continuous values. Specifically, for convolutional neural networks, deep learning models are well-suited to learn complex functions, and optimize performance of regularization methods. Batch normalization layers served to stabilize the learning process by standardizing inputs, and standard rectified linear activation layers for CNNs were used to optimize performance. In general, the number of layers and nodes depends heavily on the input data and prediction goals, therefore heuristics were used to generate a standard CNN model for image analysis, the model was then fine-tuned based on performance on the testing dataset. This resulted in three 3x3 kernel sized convolutional layers. L2 regularization and Dropout layers, using a parametric probability of 0.3 as determined by analyzing performance results, were used to prevent overfitting. An optimal batch size of 8 was determined by five-fold cross validation, and the model was trained for 110 epochs. The loss function was measured by mean squared error as the model predicted continuous values, and the learning rate was set to 0.001, as it proved to perform more positively than the standard 0.01 rate in early training, and the Adam optimizer proved the most successful in minimizing loss.

The database is weak to draw a conclusion.

We agree that the available CXR data are of small sample size and have limited information. We emphasized these in the Limitations section. We hope that future studies will include larger multi-institutional dataset.

Taking into account the uncertainties given in the various analyzes, the results are comparable. In addition, the efficiencies must be revised downwards.

Thank you. We toned down efficiency claim.

Please explain how this manuscript advances this field of research and/or contributes something new to the literature.

Thank you. We have revised the last paragraph of the Introduction and Discussion to reflect your comments.

Attachment

Submitted filename: Response.docx

Decision Letter 1

Dilbag Singh

13 Jul 2020

Deep transfer learning artificial intelligence accurately stages COVID-19 lung disease severity on portable chest radiographs

PONE-D-20-16042R1

Dear Dr. Duong,

We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements.

Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication.

An invoice for payment will follow shortly after the formal acceptance. To ensure an efficient process, please log into Editorial Manager at http://www.editorialmanager.com/pone/, click the 'Update My Information' link at the top of the page, and double check that your user information is up-to-date. If you have any billing related questions, please contact our Author Billing department directly at authorbilling@plos.org.

If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

Kind regards,

Dilbag Singh

Academic Editor

PLOS ONE

Additional Editor Comments (optional):

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation.

Reviewer #1: All comments have been addressed

Reviewer #2: All comments have been addressed

**********

2. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

Reviewer #2: Yes

**********

3. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

Reviewer #2: Yes

**********

4. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: Yes

**********

5. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

**********

6. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: (No Response)

Reviewer #2: The recommendations given previously have been addressed successfully.

The paper is Accepted as it is.

**********

7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

Acceptance letter

Dilbag Singh

20 Jul 2020

PONE-D-20-16042R1

Deep transfer learning artificial intelligence accurately stages COVID-19 lung disease severity on portable chest radiographs

Dear Dr. Duong:

I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now with our production department.

If your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information please contact onepress@plos.org.

If we can help with anything else, please email us at plosone@plos.org.

Thank you for submitting your work to PLOS ONE and supporting open access.

Kind regards,

PLOS ONE Editorial Office Staff

on behalf of

Dr. Dilbag Singh

Academic Editor

PLOS ONE


Articles from PLoS ONE are provided here courtesy of PLOS

RESOURCES