Skip to main content
JMIR Formative Research logoLink to JMIR Formative Research
. 2023 Oct 26;7:e46905. doi: 10.2196/46905

A Hybrid Decision Tree and Deep Learning Approach Combining Medical Imaging and Electronic Medical Records to Predict Intubation Among Hospitalized Patients With COVID-19: Algorithm Development and Validation

Kim-Anh-Nhi Nguyen 1,, Pranai Tandon 2, Sahar Ghanavati 1, Satya Narayana Cheetirala 1, Prem Timsina 1, Robert Freeman 1,3, David Reich 3,4, Matthew A Levin 4,5,6, Madhu Mazumdar 1,7, Zahi A Fayad 8,9, Arash Kia 1,4
Editor: Amaryllis Mavragani
Reviewed by: Yong Fan, Casey Ta
PMCID: PMC10636624  PMID: 37883177

Abstract

Background

Early prediction of the need for invasive mechanical ventilation (IMV) in patients hospitalized with COVID-19 symptoms can help in the allocation of resources appropriately and improve patient outcomes by appropriately monitoring and treating patients at the greatest risk of respiratory failure. To help with the complexity of deciding whether a patient needs IMV, machine learning algorithms may help bring more prognostic value in a timely and systematic manner. Chest radiographs (CXRs) and electronic medical records (EMRs), typically obtained early in patients admitted with COVID-19, are the keys to deciding whether they need IMV.

Objective

We aimed to evaluate the use of a machine learning model to predict the need for intubation within 24 hours by using a combination of CXR and EMR data in an end-to-end automated pipeline. We included historical data from 2481 hospitalizations at The Mount Sinai Hospital in New York City.

Methods

CXRs were first resized, rescaled, and normalized. Then lungs were segmented from the CXRs by using a U-Net algorithm. After splitting them into a training and a test set, the training set images were augmented. The augmented images were used to train an image classifier to predict the probability of intubation with a prediction window of 24 hours by retraining a pretrained DenseNet model by using transfer learning, 10-fold cross-validation, and grid search. Then, in the final fusion model, we trained a random forest algorithm via 10-fold cross-validation by combining the probability score from the image classifier with 41 longitudinal variables in the EMR. Variables in the EMR included clinical and laboratory data routinely collected in the inpatient setting. The final fusion model gave a prediction likelihood for the need of intubation within 24 hours as well.

Results

At a prediction probability threshold of 0.5, the fusion model provided 78.9% (95% CI 59%-96%) sensitivity, 83% (95% CI 76%-89%) specificity, 0.509 (95% CI 0.34-0.67) F1-score, 0.874 (95% CI 0.80-0.94) area under the receiver operating characteristic curve (AUROC), and 0.497 (95% CI 0.32-0.65) area under the precision recall curve (AUPRC) on the holdout set. Compared to the image classifier alone, which had an AUROC of 0.577 (95% CI 0.44-0.73) and an AUPRC of 0.206 (95% CI 0.08-0.38), the fusion model showed significant improvement (P<.001). The most important predictor variables were respiratory rate, C-reactive protein, oxygen saturation, and lactate dehydrogenase. The imaging probability score ranked 15th in overall feature importance.

Conclusions

We show that, when linked with EMR data, an automated deep learning image classifier improved performance in identifying hospitalized patients with severe COVID-19 at risk for intubation. With additional prospective and external validation, such a model may assist risk assessment and optimize clinical decision-making in choosing the best care plan during the critical stages of COVID-19.

Keywords: COVID-19, medical imaging, machine learning, chest radiograph, mechanical ventilation, electronic health records, intubation, decision trees, hybrid model, clinical informatics

Introduction

Severe COVID-19 caused by SARS-CoV-2 predominantly affects the lungs due to the high affinity of the virus for the angiotensin-converting enzyme 2 receptor expressed extensively in the alveolar epithelium [1]. Approximately 14% of patients with COVID-19 required hospitalization during the initial wave of the pandemic, and the intensive care unit transfer rate ranged from 5% to 32% [2,3]. Acute hypoxemic respiratory failure, complicated by acute respiratory distress syndrome, is a frequent cause of mortality among hospitalized patients with severe COVID-19. Thus, airway and ventilation management is crucial for optimizing patient outcomes [4]. There are several guidelines for the respiratory management of SARS-CoV-2 infection, supporting the emerging consensus that noninvasive ventilation and high-flow nasal cannula are superior to invasive mechanical ventilation (IMV) for treating COVID-19 acute hypoxemic respiratory failure [5-7]. IMV, however, may ultimately be required in 8%-20% of those hospitalized with COVID-19 [8-10].

The decision to intubate a patient with COVID-19 and the timings of intubation are very challenging, and there remains significant clinical uncertainty. Currently, clinical judgment, patient’s choice, and advance directives regarding IMV are the main drivers of the decision to intubate. Clinical markers such as respiratory rate, oxygen saturation, dyspnea, arterial blood gases, and radiographic observations are the primary markers routinely being used to identify candidates for intubation [11]. There is no traditionally agreed upon numeric score or index, and while certain indices have been proposed, such as the ratio of oxygen saturation index, their use is limited to certain samples, and these indices are in the early phase of clinical validation and adoption [12]. As such, opportunity exists for multimodal artificial intelligence methods to fill this gap.

Since 2020, many published studies [13-24] have tried to use machine learning techniques to predict the need for mechanical ventilation in patients with COVID-19. The majority of these studies used only clinical variables (structured data) [13] and only 15 of them (Figure 1 [13-24]) considered chest radiographs (CXRs) as a potential modality combined with clinical variables. Figure 1 is a funnel graph showing the number of similar published studies by criteria of review. The scope and top criteria for this study are “COVID-19 intubation predictive model using CXR data.” All referenced studies were found through the following PubMed query between January 1, 2020, and February 28, 2023: (“COVID-19” OR “coronavirus disease 2019”) AND (“artificial intelligence” OR “machine learning” OR “deep learning” OR “convolutional neural network”) AND (“chest x-ray” OR “chest radiograph”) AND (“intubation” or “mechanical ventilation”). Out of the 18 studies found, 6 were out of our study’s scope (different clinical outcome prediction or review type of studies). Each study was evaluated against each criterion. No study satisfied all the criteria except our study. Our new approach not only combined CXR data and clinical variables to predict the need for mechanical ventilation but also tried to show that applying automated image segmentation and using longitudinal values of clinical observations builds the prognostic potential in patients’ clinical profiles.

Figure 1.

Figure 1

Funnel graph showing the number of similar published studies [13-24] by criteria of review.

This study evaluates a machine learning risk stratification approach to predict the need for invasive ventilation based on a broad range of potential predictors. We designed a multimodality machine learning classifier based on electronic medical record (EMR) data and picture archiving and communication system images to predict the likelihood of intubation for patients with COVID-19 on the floor up to 24 hours in advance.

Methods

Study Population and Setting

We included all adult patients (≥18 years of age) admitted to The Mount Sinai Hospital (New York, NY) between March 8, 2020, and January 29, 2021, with a confirmed COVID-19 diagnosis by real-time reverse transcription polymerase chain reaction at the time of admission. Patients who were intubated or discharged within 24 hours of admission were excluded. Figure 2 shows the flowchart of the inclusion and exclusion of the patients in this cohort. This study adhered to the TRIPOD (Transparent Reporting of a Multivariable Prediction Model for Individual Prognosis Or Diagnosis) statement [25].

Figure 2.

Figure 2

Inclusion and exclusion criteria chart. RT-PCR: reverse transcription polymerase chain reaction.

Ethics Approval

This study was undertaken at The Mount Sinai Hospital, a 1134-bed tertiary care teaching facility, and it was approved by the institutional research board (approval IRB-18-00581). All methods were performed in accordance with the relevant guidelines and regulations provided by the institutional research board, which granted a waiver of informed consent.

Data Sources

The Mount Sinai Hospital currently uses 3 main electronic health record platforms: Epic (Epic Systems), Cerner (Cerner Corporation), and Laboratory Information Systems Suite (Single Copy Cluster Soft Computer). Data are aggregated from all 3 systems into a harmonized data warehouse. We received admission-discharge-transfer events from Cerner, laboratory results from Laboratory Information Systems Suite, and clinical data (ie, vital signs and nursing assessments) from Epic. Electrocardiogram results were obtained from the MUSE cardiology information system (GE HealthCare Technologies, Inc). To assemble the CXR data set, we obtained raw DICOM (Digital Imaging and Communications in Medicine) files from the picture archiving and communication system platform (GE HealthCare Technologies, Inc). CXRs taken in supine and upright positions were included.

Label and Clinical Profile 

All inpatient encounters were annotated based on the following logic.

  1. If the intubation happened within the inpatient hospital length of stay, the label was positive, and the label time stamp was the intubation time.

  2. Otherwise, we consider that the patient was not intubated, and therefore, the label was negative, and the label time stamp was the discharge time.

Clinical profiles, including vital signs, laboratory results, nursing assessments, and electrocardiograms, were censored 24 hours before the label time stamp. The prediction window of 24 hours was chosen to provide a timely opportunity for clinical interventions, goals of care discussions, and resource planning.

CXR Processing

We included radiograph images with computed radiography and digital radiography modalities in anteroposterior or posteroanterior views. For each patient with a CXR, we used the last segmented CXR before the prediction time stamp. The images were resized to 224×224 pixels; then, their pixel intensities were rescaled to range between 0 and 255, and histogram matching and normalization were performed on the intensities. To make sure that the deep learning model did not overfit and was robust, we applied oversampling of the minority class by augmenting each image in the training set by using a random combination of right or left rotation (maximum 15°), random flipping, random translation, random blurring, and random sharpening. The region of interest in the acquired CXR was the lungs (left and right). However, they were taken with some noise surrounding the lungs, including annotated text in the corners; external devices placed on the patient; and adjacent anatomy, including the shoulders, neck, and heart. We performed image segmentation in order to retain only the lung regions of the images.

The CXRs were segmented using the U-Net model architecture [26], a fully connected convolutional neural network consisting of an encoder and a decoder. Specifically, we used the LungVAE [27] implementation of U-Net, which was trained on a publicly available CXR data set. Figure 3 shows a CXR before and after segmentation.

Figure 3.

Figure 3

Chest radiographic images before (A and C) and after (B and D) image segmentation.

Modeling and Localization Framework

Training, Testing, and Holdout Split

We randomly split the cohort into a training set (1734/2481, 70%), a test set (432/2481, 17%), and a holdout set (315/2481, 13%), with no patient overlap between the sets. Because the intubation rate was 9.6% (239/2481) over the whole cohort, there was an extreme class imbalance between the majority class (nonintubated patients) and the minority class (intubated patients). We performed random undersampling [28] on the training data set to balance the majority class until both classes were equally balanced.

Transfer Learning Approach

The segmented CXRs were then fed into a pretrained DenseNet-201 [29] model. Multimedia Appendix 1 shows the architecture of the DenseNet-201 model. The DenseNet model was pretrained on RGB images (3 input channels) from ImageNet data set and on 1000 classes. Model input and output dimensions were changed to fit a grayscale image binary classification task. We also modified the architecture by adding a linear convolutional layer with a rectified linear unit activation function and dropout and used the LogSoftmax function to obtain the final probability output.

The model prespecifications were as follows: Adam optimizer, the loss function was binary cross-entropy, and the epoch size was 50. The framework used was PyTorch (version 1.01). Both segmentation and classification model training were performed using PyTorch libraries in Python [30] and trained with graphic processing unit clusters on Amazon Web Services in a secured network.

Then, using transfer learning [31], our binary classifier was trained using 10-fold cross-validation and a grid-search algorithm to tune hyperparameters (learning rate, number of hidden units, dropout, batch size) based on the area under the receiver operating characteristic curve (AUROC). Multimedia Appendix 2 shows the search ranges and the optimal hyperparameters.

Predictors in the Image Classifier

Convolutional neural network models lack decomposability into intuitive and understandable components, making them hard to interpret. To interpret our image classifier, we used the gradient-weighted class activation mapping method [32]. This technique provides us with a way to look into what particular parts of the image influenced the whole model’s decision for a specifically assigned label. It uses the gradients of our target label (intubation) flowing into the final convolutional layer to produce a coarse localization map, highlighting the important regions in the image for predicting the label. We tested our method on our test images, but this tool is yet to be automated.

Model Fusion Classification: Combining EMR and CXR Data

When combining both data modalities—CXRs and EMR variables—different methods called fusion methods are possible [33]. The fusion model implemented here is a random forest [34] that has, as a feature vector, a concatenation of longitudinal features from the EMR (patient demographics, laboratory results, vitals, flowsheets) and the output probability from the image classifier. The final feature vector is described in Figure 4.

Figure 4.

Figure 4

Fusion model architecture. The first 3 steps of the electronic medical record pipeline refer to output data frames. DF: data frame; EMR: electronic medical record.

Sampling Strategy for EMR Features

Given the crisis nature of the pandemic, clinicians caring for this cohort collected data such as vital signs, laboratory results, electrocardiograms, and nursing assessments, based on clinical judgment and resource availability rather than standard protocols during the early phase of the crisis. Thus, to create longitudinal (time-series) data for each observational variable, we included the 3 most recent assessments available before the prediction time (Figure 5). Missing values for each variable were imputed using the median value across the cohort [35]. When less than 3 assessments are available for a particular variable, the available values are placed in the most recent time slots, and the oldest time-slot value is imputed with the cross-cohort median for that variable.

Figure 5.

Figure 5

Sampling strategy flowchart for the electronic medical record variables.

EMR Feature Selection

From a total of 56 routinely collected EMR variables from the hospital, an optimal set of 41 variables was selected for the development of the predictive models (Multimedia Appendix 3). The variables initially removed included those with 90% or higher missing values and highly correlated variables [36] (above 0.7). We then performed recursive feature elimination [37]. In this approach, a single feature is removed at each step, and the model is evaluated on the test set. The quality of the fit to the data is measured using AUROC. Variables whose removal does not significantly alter the AUROC are eliminated from the feature set.

Model Fusion Strategy

A random forest model was developed and optimized in Scala/Spark with the MLlib library [38] by using the training and test sets. It was trained using 10-fold cross-validation and a grid search algorithm to tune hyperparameters based on the AUROC on the test set to have robust evaluation.

Model Testing and Statistical Methods

For each of the developed models, performance was evaluated on the test set and on the holdout set (which was not used for model development), and the model-derived class probabilities were used to predict intubation within 24 hours with a default threshold of 0.5. Predictions less than the threshold were categorized as negative. Sensitivity, specificity, accuracy, positive predictive value, negative predictive value, F1-score, AUROC, and area under the precision recall curve (AUPRC), along with bootstrap 95% CIs, were estimated for evaluating the screening tool’s performance. Group comparisons were performed using a 2-sided Student t test or Kruskal-Wallis for continuous variables as appropriate and chi-square test for categorical variables. All analyses were performed using SciPy in Python.

Results

Study Population and Outcomes

A total of 2481 COVID-19–positive patients were included in the overall study cohort. This cohort included a higher proportion of men (1390/2481, 56%), and the median age was 62.2 years. The median duration of hospital stay was 4.9 days and ranged from 1 to 72 days. The overall rate of intubation was 9.6% (239/2481) in the whole study cohort. Table 1 shows the clinical characteristics and descriptive statistics of the cohort. Intubated patients were significantly older and more likely to be male and diabetic than the nonintubated patients.

Table 1.

Patient cohort and characteristics and statistical comparisons between intubated and nonintubated patient groups.

Characteristics Overall (N=2481) Intubated (n=239) Nonintubated (n=2242) P value
Age (years) <.001

Mean (SD) 60.4 (17.7) 64.9 (12.4) 59.9 (18.1)  

Median (min-max) 62.2 (18-120) 65.5 (20-94) 62.0 (18-120)  
Gender, mean (SD) .03

Male 1390 (56.1) 135 (64) 1237 (55.2)  

Female 1089 (43.9) 86 (36) 1003 (44.7)  

Other 2 (0.08) 0 (0) 2 (0.1)  
Race and ethnicity, mean (SD) <.001

White 819 (32.9) 72 (30.1) 746 (33.3)  

African American 456 (18.4) 24 (10) 433 (19.3)  

Hispanic 600 (24.2) 65 (27.2) 536 (23.9)  

Asian 129 (5.2) 13 (5.4) 116 (5.2)  

Other 358 (14.4) 50 (20.9) 308 (13.7)  

Unspecified 119 (4.8) 15 (6.3) 103 (4.6)  
BMI .03

Mean (SD) 29.4 (7.3) 30.5 (8.2) 29.3 (7.2)  

Median (min-max) 28.3 (12.5-69.3) 28.7 (12.4-60.5) 28.3 (12.5-69.3)  
Smoking history, mean (SD) .73

Current smoker 24 (0.9) 1 (0.4) 23 (1)  

Past smoker 558 (22.5) 48 (20.1) 510 (22.7)  

Never smoked 78 (3.1) 9 (3.8) 69 (3.1)  

Missing 1821 (73.4) 181 (75.7) 1640 (73.2)  
Hypertension, mean (SD) .07

Yes 1289 (51.9) 130 (54.4) 1159 (51.7)  

No 1013 (40.8) 79 (33) 934 (41.7)  

Missing 179 (7.2) 30 (12.6) 149 (6.6)  
Diabetes, mean (SD) <.001

Yes 854 (34.4) 112 (46.9) 742 (33.1)  

No 1448 (58.4) 97 (40.5) 1351 (60.3)  

Missing 179 (7.2) 30 (12.6) 149 (6.6)  
Chronic obstructive pulmonary disease, mean (SD) .41

Yes 399 (16.1) 41 (17.1) 358 (16)  

No 1903 (76.7) 168 (70.3) 1735 (77.4)  

Missing 179 (7.2) 30 (12.6) 149 (6.6)  
Obesity, mean (SD) <.001

Yes 445 (17.9) 64 (26.8) 381 (17)  

No 1857 (74.9) 145 (60.7) 1712 (76.4)  

Missing 179 (7.2) 30 (12.5) 149 (6.6)  
Length of stay (days) .14

Mean (SD) 6.6 (6.2) 7.2 (8.2) 6.5 (6.1)  

Median (min-max) 4.9 (1-72) 4.7 (1-72) 4.9 (1-48)  
Intensive care unit care received, mean (SD) <.001

Yes 470 (18.9) 239 (100) 231 (10.3)  

No 2011 (81.1) 0 (0) 2011 (89.7)  

Predictors in the Final Fusion Model

Hyperparameters used in the final random forest model are shown in Multimedia Appendix 2. Figure 6 summarizes the top predictive variables ordered by the Gini coefficient (the definitions of the variables in this figure are shown in Multimedia Appendix 3). Our model identified a series of features related to progressive respiratory failure (respiratory rate, oxygen saturation), markers of systemic inflammation (C-reactive protein, white blood cell count, lactate dehydrogenase), hemodynamics (systolic and diastolic blood pressures), renal failure (blood urea nitrogen, anion gap, and serum creatinine), and immune dysregulation (lymphocyte count). Respiratory rate (the earliest recorded value of the latest 3 assessments) had the highest predictive value in the random forest model, and white blood cell count was the second highest. Variables included in the final model reflected the importance of temporal changes in vital signs, markers of acid-base equilibrium and systemic inflammation, and predictors of myocardial injury and renal function. Figure 7 shows the parts of the lungs that contributed to intubation risk prediction.

Figure 6.

Figure 6

Gini coefficients of the joint fusion random forest model variables. Refer to Multimedia Appendix 3 for the definitions of the variables.

Figure 7.

Figure 7

(A) Example of a segmented lung from the last chest radiograph of an intubated patient. (B) The most important features (pixels) predicting intubation are highlighted in the class activation map calculated by gradient-weighted class activation mapping projected on the image.

Comparison of the Predictive Performance of the Models

At a prediction probability threshold of 0.5, the AUROC for the image classifier alone was 0.58 (95% CI 0.44-0.73) and the AUPRC was 0.21 (95% CI 0.08-0.38), with a positive predictive value of 14.8% (95% CI 7%-24%) on the holdout set. Table 2 shows all the performance metrics for all the models on the test set and the holdout set. Compared to the image classifier, the fusion model provided boosted performance results in the test set and the holdout set. By adding additional EMR features, the sensitivity doubled from 38.5% to 78.9%, specificity increased by nearly 10%, accuracy by 15%, positive predictive value by 104%, AUROC by 51%, F1-score by 112%, and the AUPRC by 140% in the holdout set. The AUROC graphs are shown in Figure 8 and Figure 9. The odds ratio for requiring mechanical ventilation within 48 hours of a positive prediction was 4.73 (95% CI 4.5-9.3) compared to a negative prediction and 11.2 (95% CI 10.4-12.0) for requiring mechanical ventilation at any time during admission in the holdout set.

Table 2.

Predictive performance of both the image classifier and the joint fusion classifier on the test set and the holdout set. Positive and negative predictions were assigned using the prediction probability threshold of 0.5.

Data set, model Sensitivity (95% CI) Specificity (95% CI) Accuracy (95% CI) PPVa (95% CI) NPVb (95% CI) F1-score (95% CI) AUROCc (95% CI) AUPRCd (95% CI) Unique
patients (n)
Intubation rate
Test 432 0.076

Imaging alone 0.5 (0.0-0.83) 0.776 (0.70-0.85) 0.757 (0.68-0.84) 0.103 (0.0-0.23) 0.965 (0.92-1.0) 0.160 (0.05-0.40) 0.684 (0.49-0.81) 0.124 (0.03-0.46)


Joint fusion 0.860 (0.67-1.0) 0.828 (0.78-0.89) 0.833 (0.78-0.88) 0.292 (0.16-0.43) 0.988 (0.96-1.0) 0.428 (0.27-0.58) 0.873 (0.76-0.95) 0.421 (0.19-0.64)

Holdout 315 0.117

Imaging alone 0.385 (0.15-0.64) 0.757 (0.68-0.84) 0.715 (0.63-0.79) 0.184 (0.07-0.32) 0.896 (0.82-0.95) 0.240 (0.09. 0.37) 0.577 (0.44-0.73) 0.206 (0.08-0.38)


Joint fusion 0.789 (0.59-0.96) 0.830 (0.76-0.89) 0.825 (0.76-0.88) 0.372 (0.22-0.54) 0.967 (0.93-0.99) 0.509 (0.34-0.67) 0.874 (0.80-0.94) 0.497 (0.32-0.65)

aPPV: positive predictive value.

bNPV: negative predictive value.

cAUROC: area under the receiver operating characteristic curve.

dAUPRC: area under the precision recall curve.

Figure 8.

Figure 8

Receiver operating characteristic curves of the image classifier (blue dashed line) and of the joint fusion model (blue solid line) on the test set and their respective areas under the curve and 95% CIs. AUROC: area under the receiver operating characteristic curve.

Figure 9.

Figure 9

Receiver operating characteristic curves of the image classifier (orange dashed line) and of the joint fusion model (orange solid line) on the holdout set and their respective areas under the curve and 95% CIs. AUROC: area under the receiver operating characteristic curve.

Discussion

In this study, we examined the utility of a deep learning image classifier based on routinely available CXR images along with clinical data to predict the need for IMV in patients with COVID-19. On the holdout set, the image classifier alone reached an AUROC of 0.58 and an AUPRC of 0.21; when the image probability was used in combination with structured EMR data in a random forest model, the fusion model reached an AUROC of 0.87 and an AUPRC of 0.50. Despite the relatively low AUPRC of the image classifier alone, it was still 15th in overall feature importance in the fusion model, outperforming some traditionally important clinical parameters such as creatinine levels, age, and venous blood pH. With optimization, a further increase in the feature importance of the image probabilities would be expected. The final fusion model had a negative predictive value of 97% and positive predictive value of 37% for the holdout set, which may provide significant clinical utility. This is supported by the fact that the odds ratio for intubation in patients with a positive prediction is greater than 11.

Several published reports have used deep learning of actual CXR images in combination with EMR data to predict the risk of intubation for patients admitted with COVID-19. Kwon et al [14], Aljouie et al [19], and Lee et al [39] used systematic manual scoring or manual labeling of CXR images to predict mechanical ventilation and deaths, achieving high performance; however, the utility of these approaches is limited, as it requires manual scoring by experts and cannot easily be rolled out to stressed health systems in an automated manner. Jiao et al [40] also used transfer learning on an ImageNet pretrained model to generate an image classifier used in fusion with EMR data to generate a classifier for intubation in patients with COVID-19 [40]. As in this study, the addition of EMR data boosted the image classifier performance, with the image classifier alone reaching an AUROC of 0.8, EMR alone reaching an AUROC of 0.82, and the fusion model an AUROC of 0.84. Although the addition of images only improved the AUROC of the EMR model from 0.82 to 0.84 in internal testing, on an external validation set, the addition of images improved AUROC from 0.73 to 0.79, which suggests that the images may be useful in guarding against overfitting. The differences between the image classifier and overall performance in the studies mentioned above and those in this study may be related to the higher event rate in their cohort, which diminished class imbalance (24% intubation rate in Jiao et al [40] vs 9.6% in this study) as well as potentially improved segmentation. Moreover, it suffers from manual review and hand editing of automated segmentation, which then limits clinical applicability versus using a fully automated imaging processing pipeline that this study offers.

Some studies utilized an end-to-end automated pipeline for processing radiography images and EMR data similar to that used in this study [17,22,24,41]; however, none make direct prediction of intubation and IMV in hospitalized patients. Chung et al [17] and Dayan et al [22] focused on the prediction of oxygen requirement in emergency department patients with limited data availability. Duanmu et al [41] focused on predicting the duration on IMV instead, but they are one of the very few using longitudinal data in their pipeline, suggesting that longitudinal data may bring more prognostic value than single-point data. O’Shea et al [24] had one of the highest performance end-to-end automated models, with an AUROC of 0.82 in predicting death or intubation within 7 days. However, those models are limited by the lack of image segmentation that ensure only pulmonary or thoracic features are considered in their models, use of a deep learning model to classify the degree of lung injury but not predict intubation itself, and use of a single point, that is, the first available value for each variable; therefore, they suffer from a lack of robustness that would not account for changes in the radiographs or in the patient’s clinical condition. The very long prediction window in O’Shea et al [24] (7 days vs 24 h in this study) is less amenable to clinical intervention.

The choice of a pretrained model may also be important. Kulkarni et al [18] used transfer learning using CheXNeXt, a DenseNet121 architecture model pretrained on a cohort of CXR images to identify lung pathologies as a base and reported an AUROC of 0.79 for their transfer learning model trained with only 510 images, suggesting that potentially fewer images are required when the model is pretrained on images closer to the appropriate subject matter [18].

The limitations of this study include a high-class imbalance of 9.6% (239/2481) intubation rate and a limited sample size of images. Another limitation was the changing practice pattern throughout the pandemic, as more was learned about the natural history of COVID-19, and practice patterns shifted to favor less frequent use of IMV [42]. Although there were fears of ventilator shortage or rationing of ventilators early in the pandemic; fortunately, there was no such shortage in the Mount Sinai Health System. Finally, the prediction time point for patients who were not intubated was selected to be 24 hours before discharge; this may potentially yield an optimistic performance benefit in this case, as patients are closer to recovery as opposed to deterioration and intubation. Further studies will demonstrate how much this affects performance. The strengths of this study include the use of a real-world clinical label of intubation that varied with practice patterns across the pandemic, use of a robust automated end-to-end pipeline that facilitated rapid deployment into the clinical setting, and fusion of image classifier and EMR classifier predictions in an interpretable manner such that the features most relevant to the prediction can be easily communicated to providers.

As the reach of deep learning and utilization of medical images in artificial intelligence–based clinical decision support increases, methods must be developed to combine these models with clinical data to optimize performance. Here, we demonstrate that, when linked with EMR data, an automated deep learning image classifier improved performance in identifying hospitalized patients with severe COVID-19 at risk for intubation. The image probability ranks highly among traditional clinical features in the relative importance of predictors. Further work is necessary to optimize the image classifier to yield higher performance and perform prospective and external validation. Ultimately, we seek methods that seamlessly integrate CXRs and other medical imaging with structured EMR data that enable real-time and highly accurate artificial intelligence clinical decision support systems.

Abbreviations

AUPRC

area under the precision recall curve

AUROC

area under the receiver operating characteristic curve

CXR

chest radiograph

DICOM

Digital Imaging and Communications in Medicine

EMR

electronic medical record

IMV

invasive mechanical ventilation

TRIPOD

Transparent Reporting of a Multivariable Prediction Model for Individual Prognosis Or Diagnosis

Multimedia Appendix 1

Original DenseNet-201 architecture and modified architecture implemented in the intubation risk classification use case.

Multimedia Appendix 2

Hyperparameters used in the final imaging model.

Multimedia Appendix 3

Variables included in the final fusion model and their respective data source.

Data Availability

Raw data underlying this study were generated by the Mount Sinai Health System. Derived data supporting the findings of this study are available from the corresponding author (Nguyen KAN) upon request.

Footnotes

Authors' Contributions: KANN, SG, and AK conceived the study. KANN, SG, SNC, P Timsina, and ZAF collected the data sources. KANN and SG performed modeling and experiments. KANN, P Tandon, and AK analyzed and validated the data and the results. KANN, P Tandon, and AK wrote the manuscript. KANN, P Tandon, SG, SNC, P Timsina, RF, DLR, MAL, MM, ZAF, and AK revised the manuscript for important intellectual content.

Conflicts of Interest: None declared.

References

  • 1.Liu Y, Ning Z, Chen Y, Guo M, Liu Y, Gali NK, Sun L, Duan Y, Cai J, Westerdahl D, Liu X, Xu K, Ho K, Kan H, Fu Q, Lan K. Aerodynamic analysis of SARS-CoV-2 in two Wuhan hospitals. Nature. 2020 Jun;582(7813):557–560. doi: 10.1038/s41586-020-2271-3.10.1038/s41586-020-2271-3 [DOI] [PubMed] [Google Scholar]
  • 2.Huang C, Wang Y, Li X, Ren L, Zhao J, Hu Y, Zhang L, Fan G, Xu J, Gu X, Cheng Z, Yu T, Xia J, Wei Y, Wu W, Xie X, Yin W, Li H, Liu M, Xiao Y, Gao H, Guo L, Xie J, Wang G, Jiang R, Gao Z, Jin Q, Wang J, Cao B. Clinical features of patients infected with 2019 novel coronavirus in Wuhan, China. The Lancet. 2020 Feb;395(10223):497–506. doi: 10.1016/s0140-6736(20)30183-5. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3.Guan W, Ni Z, Hu Y, Liang W, Ou C, He J, Liu L, Shan H, Lei C, Hui DS, Du B, Li L, Zeng G, Yuen K, Chen R, Tang C, Wang T, Chen P, Xiang J, Li S, Wang J, Liang Z, Peng Y, Wei L, Liu Y, Hu Y, Peng P, Wang J, Liu J, Chen Z, Li G, Zheng Z, Qiu S, Luo J, Ye C, Zhu S, Zhong N, China Medical Treatment Expert Group for COVID-19 Clinical characteristics of coronavirus disease 2019 in China. N Engl J Med. 2020 Apr 30;382(18):1708–1720. doi: 10.1056/NEJMoa2002032. https://europepmc.org/abstract/MED/32109013 . [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4.Yang J, Zheng Y, Gou X, Pu K, Chen Z, Guo Q, Ji R, Wang H, Wang Y, Zhou Y. Prevalence of comorbidities and its effects in patients infected with SARS-CoV-2: a systematic review and meta-analysis. Int J Infect Dis. 2020 May;94:91–95. doi: 10.1016/j.ijid.2020.03.017. https://linkinghub.elsevier.com/retrieve/pii/S1201-9712(20)30136-3 .S1201-9712(20)30136-3 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5.Lazzeri M, Lanza A, Bellini R, Bellofiore A, Cecchetto S, Colombo A, D'Abrosca F, Del Monaco C, Gaudiello Giuseppe, Paneroni M, Privitera E, Retucci M, Rossi V, Santambrogio M, Sommariva M, Frigerio P. Respiratory physiotherapy in patients with COVID-19 infection in acute setting: a position paper of the Italian Association of Respiratory Physiotherapists (ARIR) Monaldi Arch Chest Dis. 2020 Mar 26;90(1):1. doi: 10.4081/monaldi.2020.1285. doi: 10.4081/monaldi.2020.1285. [DOI] [PubMed] [Google Scholar]
  • 6.Costa WNDS, Miguel JP, Prado FDS, Lula LHSDM, Amarante GAJ, Righetti RF, Yamaguti WP. Noninvasive ventilation and high-flow nasal cannula in patients with acute hypoxemic respiratory failure by COVID-19: A retrospective study of the feasibility, safety and outcomes. Respir Physiol Neurobiol. 2022 Apr;298:103842. doi: 10.1016/j.resp.2022.103842. https://linkinghub.elsevier.com/retrieve/pii/S1569-9048(22)00001-5 .S1569-9048(22)00001-5 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7.Alhazzani W, Møller Morten Hylander, Arabi YM, Loeb M, Gong MN, Fan E, Oczkowski S, Levy MM, Derde L, Dzierba A, Du B, Aboodi M, Wunsch H, Cecconi M, Koh Y, Chertow DS, Maitland K, Alshamsi F, Belley-Cote E, Greco M, Laundy M, Morgan JS, Kesecioglu J, McGeer A, Mermel L, Mammen MJ, Alexander PE, Arrington A, Centofanti JE, Citerio G, Baw B, Memish ZA, Hammond N, Hayden FG, Evans L, Rhodes A. Surviving sepsis campaign: guidelines on the management of critically ill adults with coronavirus disease 2019 (COVID-19) Intensive Care Med. 2020 May;46(5):854–887. doi: 10.1007/s00134-020-06022-5. https://europepmc.org/abstract/MED/32222812 .10.1007/s00134-020-06022-5 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.Richardson S, Hirsch JS, Narasimhan M, Crawford JM, McGinn T, Davidson KW, the Northwell COVID-19 Research Consortium. Barnaby Douglas P, Becker Lance B, Chelico John D, Cohen Stuart L, Cookingham Jennifer, Coppa Kevin, Diefenbach Michael A, Dominello Andrew J, Duer-Hefele Joan, Falzon Louise, Gitlin Jordan, Hajizadeh Negin, Harvin Tiffany G, Hirschwerk David A, Kim Eun Ji, Kozel Zachary M, Marrast Lyndonna M, Mogavero Jazmin N, Osorio Gabrielle A, Qiu Michael, Zanos Theodoros P. Presenting characteristics, comorbidities, and outcomes among 5700 patients hospitalized with COVID-19 in the New York City Area. JAMA. 2020 May 26;323(20):2052–2059. doi: 10.1001/jama.2020.6775. https://europepmc.org/abstract/MED/32320003 .2765184 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9.Weissman DN, de Perio MA, Radonovich LJ. COVID-19 and risks posed to personnel during endotracheal intubation. JAMA. 2020 May 26;323(20):2027–2028. doi: 10.1001/jama.2020.6627. https://europepmc.org/abstract/MED/32338710 .2765376 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10.Goyal P, Choi JJ, Pinheiro LC, Schenck EJ, Chen R, Jabri A, Satlin MJ, Campion TR, Nahid M, Ringel JB, Hoffman KL, Alshak MN, Li HA, Wehmeyer GT, Rajan M, Reshetnyak E, Hupert N, Horn EM, Martinez FJ, Gulick RM, Safford MM. Clinical characteristics of COVID-19 in New York City. N Engl J Med. 2020 Jun 11;382(24):2372–2374. doi: 10.1056/NEJMc2010419. https://europepmc.org/abstract/MED/32302078 . [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11.Tobin MJ, Laghi F, Jubran A. Caution about early intubation and mechanical ventilation in COVID-19. Ann Intensive Care. 2020 Jun 09;10(1):78. doi: 10.1186/s13613-020-00692-6. https://europepmc.org/abstract/MED/32519064 .10.1186/s13613-020-00692-6 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12.Roca O, Caralt B, Messika J, Samper M, Sztrymf B, Hernández G, García-de-Acilu M, Frat J, Masclans JR, Ricard J. An index combining respiratory rate and oxygenation to predict outcome of nasal high-flow therapy. Am J Respir Crit Care Med. 2019 Jun 01;199(11):1368–1376. doi: 10.1164/rccm.201803-0589oc. [DOI] [PubMed] [Google Scholar]
  • 13.Arvind V, Kim JS, Cho BH, Geng E, Cho SK. Development of a machine learning algorithm to predict intubation among hospitalized patients with COVID-19. J Crit Care. 2021 Apr;62:25–30. doi: 10.1016/j.jcrc.2020.10.033. https://europepmc.org/abstract/MED/33238219 .S0883-9441(20)30755-3 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14.Kwon YJ(, Toussie D, Finkelstein M, Cedillo MA, Maron SZ, Manna S, Voutsinas N, Eber C, Jacobi A, Bernheim A, Gupta YS, Chung MS, Fayad ZA, Glicksberg BS, Oermann EK, Costa AB. Combining initial radiographs and clinical variables improves deep learning prognostication in patients with COVID-19 from the emergency department. Radiol Artif Intell. 2021 Mar;3(2):e200098. doi: 10.1148/ryai.2020200098. https://europepmc.org/abstract/MED/33928257 . [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15.Li MD, Little BP, Alkasab TK, Mendoza DP, Succi MD, Shepard JO, Lev MH, Kalpathy-Cramer J. Multi-radiologist user study for artificial intelligence-guided grading of COVID-19 lung disease severity on chest radiographs. Acad Radiol. 2021 Apr;28(4):572–576. doi: 10.1016/j.acra.2021.01.016. https://europepmc.org/abstract/MED/33485773 .S1076-6332(21)00025-8 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16.Ardestani A, Li MD, Chea P, Wortman JR, Medina A, Kalpathy-Cramer J, Wald C. External COVID-19 deep learning model validation on ACR AI-LAB: It's a brave new world. J Am Coll Radiol. 2022 Jul;19(7):891–900. doi: 10.1016/j.jacr.2022.03.013. https://europepmc.org/abstract/MED/35483438 .S1546-1440(22)00278-2 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17.Chung J, Kim D, Choi J, Yune S, Song K, Kim S, Chua M, Succi MD, Conklin J, Longo MGF, Ackman JB, Petranovic M, Lev MH, Do S. Prediction of oxygen requirement in patients with COVID-19 using a pre-trained chest radiograph xAI model: efficient development of auditable risk prediction models via a fine-tuning approach. Sci Rep. 2022 Dec 07;12(1):21164. doi: 10.1038/s41598-022-24721-5. doi: 10.1038/s41598-022-24721-5.10.1038/s41598-022-24721-5 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18.Kulkarni AR, Athavale AM, Sahni A, Sukhal S, Saini A, Itteera M, Zhukovsky S, Vernik J, Abraham M, Joshi A, Amarah A, Ruiz J, Hart PD, Kulkarni H. Deep learning model to predict the need for mechanical ventilation using chest X-ray images in hospitalised patients with COVID-19. BMJ Innov. 2021 Apr;7(2):261–270. doi: 10.1136/bmjinnov-2020-000593.bmjinnov-2020-000593 [DOI] [PubMed] [Google Scholar]
  • 19.Aljouie AF, Almazroa A, Bokhari Y, Alawad M, Mahmoud E, Alawad E, Alsehawi A, Rashid M, Alomair L, Almozaai S, Albesher B, Alomaish H, Daghistani R, Alharbi NK, Alaamery M, Bosaeed M, Alshaalan H. Early prediction of COVID-19 ventilation requirement and mortality from routinely collected baseline chest radiographs, laboratory, and clinical data with machine learning. JMDH. 2021 Jul;Volume 14:2017–2033. doi: 10.2147/jmdh.s322431. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 20.Pyrros A, Flanders AE, Rodríguez-Fernández Jorge Mario, Chen A, Cole P, Wenzke D, Hart E, Harford S, Horowitz J, Nikolaidis P, Muzaffar N, Boddipalli V, Nebhrajani J, Siddiqui N, Willis M, Darabi H, Koyejo O, Galanter W. Predicting prolonged hospitalization and supplemental oxygenation in patients with COVID-19 infection from ambulatory chest radiographs using deep learning. Acad Radiol. 2021 Aug;28(8):1151–1158. doi: 10.1016/j.acra.2021.05.002. https://europepmc.org/abstract/MED/34134940 .S1076-6332(21)00217-8 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21.Varghese BA, Shin H, Desai B, Gholamrezanezhad A, Lei X, Perkins M, Oberai A, Nanda N, Cen S, Duddalwar V. Predicting clinical outcomes in COVID-19 using radiomics on chest radiographs. Br J Radiol. 2021 Oct 01;94(1126):20210221. doi: 10.1259/bjr.20210221. https://europepmc.org/abstract/MED/34520246 . [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22.Dayan I, Roth HR, Zhong A, Harouni A, Gentili A, Abidin AZ, Liu A, Costa AB, Wood BJ, Tsai C, Wang C, Hsu C, Lee CK, Ruan P, Xu D, Wu D, Huang E, Kitamura FC, Lacey G, de Antônio Corradi Gustavo César, Nino G, Shin H, Obinata H, Ren H, Crane JC, Tetreault J, Guan J, Garrett JW, Kaggie JD, Park JG, Dreyer K, Juluru K, Kersten K, Rockenbach MABC, Linguraru MG, Haider MA, AbdelMaseeh M, Rieke N, Damasceno PF, E Silva Pedro Mario Cruz, Wang P, Xu S, Kawano S, Sriswasdi S, Park SY, Grist TM, Buch V, Jantarabenjakul W, Wang W, Tak WY, Li X, Lin X, Kwon YJ, Quraini A, Feng A, Priest AN, Turkbey B, Glicksberg B, Bizzo B, Kim BS, Tor-Díez Carlos, Lee C, Hsu C, Lin C, Lai C, Hess CP, Compas C, Bhatia D, Oermann EK, Leibovitz E, Sasaki H, Mori H, Yang I, Sohn JH, Murthy KNK, Fu L, de Mendonça Matheus Ribeiro Furtado, Fralick M, Kang MK, Adil M, Gangai N, Vateekul P, Elnajjar P, Hickman S, Majumdar S, McLeod SL, Reed S, Gräf Stefan, Harmon S, Kodama T, Puthanakit T, Mazzulli T, de Lavor VL, Rakvongthai Y, Lee YR, Wen Y, Gilbert FJ, Flores MG, Li Q. Federated learning for predicting clinical outcomes in patients with COVID-19. Nat Med. 2021 Oct;27(10):1735–1743. doi: 10.1038/s41591-021-01506-3. https://europepmc.org/abstract/MED/34526699 .10.1038/s41591-021-01506-3 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 23.Patel NJ, D'Silva KM, Li MD, Hsu TYT, DiIorio M, Fu X, Cook C, Prisco L, Martin L, Vanni KMM, Zaccardelli A, Zhang Y, Kalpathy-Cramer Jayashree, Sparks JA, Wallace ZS. Assessing the severity of COVID-19 lung injury in rheumatic diseases versus the general population using deep learning-derived chest radiograph scores. Arthritis Care Res (Hoboken) 2023 Mar;75(3):657–666. doi: 10.1002/acr.24883. https://europepmc.org/abstract/MED/35313091 . [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 24.O'Shea A, Li MD, Mercaldo ND, Balthazar P, Som A, Yeung T, Succi MD, Little BP, Kalpathy-Cramer J, Lee SI. Intubation and mortality prediction in hospitalized COVID-19 patients using a combination of convolutional neural network-based scoring of chest radiographs and clinical data. BJR Open. 2022;4(1):20210062. doi: 10.1259/bjro.20210062. https://europepmc.org/abstract/MED/36105420 . [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25.Collins GS, Reitsma JB, Altman DG, Moons KG. Transparent Reporting of a multivariable prediction model for Individual Prognosis Or Diagnosis (TRIPOD) Ann Intern Med. 2015 May 19;162(10):735–736. doi: 10.7326/l15-5093-2. [DOI] [PubMed] [Google Scholar]
  • 26.Ronneberger O, Fischer P, Brox T. U-net: Convolutional networks for biomedical image segmentation. arXiv. doi: 10.1007/978-3-319-24574-4_28. Preprint posted online on May 18, 2015. [DOI] [Google Scholar]
  • 27.Selvan R, Dam E, Detlefsen N, Rischel S, Sheng K, Nielsen M, Pai A. Lung segmentation from chest x-rays using variational data imputation. arXiv. doi: 10.48550/arXiv.2005.10052. Preprint posted online on July 7, 2020. [DOI] [Google Scholar]
  • 28.Japkowicz N. The class imbalance problem: significance and strategies. CiteSeerX. 200. [2023-10-06]. https://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.35.1693&rep=rep1&type=pdf .
  • 29.Huang G, Liu Z, Van DML, Weinberger K. Densely connected convolutional networks. Proceedings of the IEEE conference on computer vision and pattern recognition; July 21-26; Honolulu, HI, USA. 2017. pp. 4700–4708. [DOI] [Google Scholar]
  • 30.Python: A dynamic, open source programming language. Python Core Team. 2008. [2008-12-03]. https://www.python.org/
  • 31.Torrey L, Shavlik J. Handbook of Research on Machine Learning Applications and Trends: Algorithms, Methods, and Techniques. PA, USA: Information Science Reference; 2010. Transfer learning; pp. 242–264. [Google Scholar]
  • 32.Selvaraju R, Cogswell M, Das A, Vedantam R, Parikh D, Batra D. Grad-CAM: visual explanations from deep networks via gradient-based localization. 2017 IEEE International Conference on Computer Vision (ICCV); October 22-29; Venice, Italy. 2017. [DOI] [Google Scholar]
  • 33.Huang S, Pareek A, Seyyedi S, Banerjee I, Lungren MP. Fusion of medical imaging and electronic health records using deep learning: a systematic review and implementation guidelines. NPJ Digit Med. 2020;3:136. doi: 10.1038/s41746-020-00341-z. doi: 10.1038/s41746-020-00341-z.341 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 34.Breiman L. Random forests. Machine Learning. 2001 Oct;45:5–32. doi: 10.1023/A:1010933404324. doi: 10.1023/A:1010933404324. [DOI] [Google Scholar]
  • 35.Tang F, Ishwaran H. Random forest missing data algorithms. Statistical Analysis and Data Mining: The ASA Data Science Journal. 2017 Dec;10(6):363–377. doi: 10.1002/sam.11348. https://europepmc.org/abstract/MED/29403567 . [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 36.Dormann CF, Elith J, Bacher S, Buchmann C, Carl G, Carré G, Marquéz JRG, Gruber B, Lafourcade B, Leitão PJ, Münkemüller T, McClean C, Osborne PE, Reineking B, Schröder B, Skidmore AK, Zurell D, Lautenbach S. Collinearity: a review of methods to deal with it and a simulation study evaluating their performance. Ecography. 2012 May 18;36(1):27–46. doi: 10.1111/j.1600-0587.2012.07348.x. [DOI] [Google Scholar]
  • 37.Guyon I, Weston J, Barnhill S, Vapnik V. Gene selection for cancer classification using support vector machines. Machine Learning. 2002 Jan;46:389–422. doi: 10.1007/978-0-387-30162-4_415. [DOI] [Google Scholar]
  • 38.Zaharia M. MLlib: main guide—spark 2.3.0 documentation. Machine Learning Library (MLlib) Guide. 2014. May, [2014-05-26]. https://spark.apache.org/docs/2.3.0/ml-guide.html .
  • 39.Lee JH, Ahn JS, Chung MJ, Jeong YJ, Kim JH, Lim JK, Kim JY, Kim YJ, Lee JE, Kim EY. Development and validation of a multimodal-based prognosis and intervention prediction model for COVID-19 patients in a multicenter cohort. Sensors (Basel) 2022 Jul 02;22(13):5007. doi: 10.3390/s22135007. https://www.mdpi.com/resolver?pii=s22135007 .s22135007 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 40.Jiao Z, Choi JW, Halsey K, Tran TML, Hsieh B, Wang D, Eweje F, Wang R, Chang K, Wu J, Collins SA, Yi TY, Delworth AT, Liu T, Healey TT, Lu S, Wang J, Feng X, Atalay MK, Yang L, Feldman M, Zhang PJL, Liao W, Fan Y, Bai HX. Prognostication of patients with COVID-19 using artificial intelligence based on chest x-rays and clinical data: a retrospective study. Lancet Digit Health. 2021 May;3(5):e286–e294. doi: 10.1016/S2589-7500(21)00039-X. https://linkinghub.elsevier.com/retrieve/pii/S2589-7500(21)00039-X .S2589-7500(21)00039-X [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 41.Duanmu H, Ren T, Li H, Mehta N, Singer AJ, Levsky JM, Lipton ML, Duong TQ. Deep learning of longitudinal chest X-ray and clinical variables predicts duration on ventilator and mortality in COVID-19 patients. Biomed Eng Online. 2022 Oct 14;21(1):77. doi: 10.1186/s12938-022-01045-z. https://biomedical-engineering-online.biomedcentral.com/articles/10.1186/s12938-022-01045-z .10.1186/s12938-022-01045-z [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 42.Tandon P, Leibner E, Ahmed S, Acquah S, Kohli-Seth R. Comparing seasonal trends in coronavirus disease 2019 patient data at a quaternary hospital in New York City. Crit Care Explor. 2021 Apr;3(4):e0381. doi: 10.1097/CCE.0000000000000381. https://europepmc.org/abstract/MED/33937865 . [DOI] [PMC free article] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Multimedia Appendix 1

Original DenseNet-201 architecture and modified architecture implemented in the intubation risk classification use case.

Multimedia Appendix 2

Hyperparameters used in the final imaging model.

Multimedia Appendix 3

Variables included in the final fusion model and their respective data source.

Data Availability Statement

Raw data underlying this study were generated by the Mount Sinai Health System. Derived data supporting the findings of this study are available from the corresponding author (Nguyen KAN) upon request.


Articles from JMIR Formative Research are provided here courtesy of JMIR Publications Inc.

RESOURCES