Skip to main content
PLOS ONE logoLink to PLOS ONE
. 2020 Dec 16;15(12):e0244051. doi: 10.1371/journal.pone.0244051

Validation and repurposing of the MSL-COVID-19 score for prediction of severe COVID-19 using simple clinical predictors in a triage setting: The Nutri-CoV score

Omar Yaxmehen Bello-Chavolla 1,2,#, Neftali E Antonio-Villa 1,3,#, Edgar Ortiz-Brizuela 4, Arsenio Vargas-Vázquez 1,3, María Fernanda González-Lara 4, Alfredo Ponce de Leon 4, José Sifuentes-Osornio 4,5, Carlos A Aguilar-Salinas 1,6,7,*
Editor: Itamar Ashkenazi8
PMCID: PMC7743927  PMID: 33326502

Abstract

Background

During the COVID-19 pandemic, risk stratification has been used to decide patient eligibility for inpatient, critical and domiciliary care. Here, we sought to validate the MSL-COVID-19 score, originally developed to predict COVID-19 mortality in Mexicans. Also, an adaptation of the formula is proposed for the prediction of COVID-19 severity in a triage setting (Nutri-CoV).

Methods

We included patients evaluated from March 16th to August 17th, 2020 at the Instituto Nacional de Ciencias Médicas y Nutrición, defining severe COVID-19 as a composite of death, ICU admission or requirement for intubation (n = 3,007). We validated MSL-COVID-19 for prediction of mortality and severe disease. Using Elastic Net Cox regression, we trained (n = 1,831) and validated (n = 1,176) a model for prediction of severe COVID-19 using MSL-COVID-19 along with clinical assessments obtained at a triage setting.

Results

The variables included in MSL-COVID-19 are: pneumonia, early onset type 2 diabetes, age > 65 years, chronic kidney disease, any form of immunosuppression, COPD, obesity, diabetes, and age <40 years. MSL-COVID-19 had good performance to predict COVID-19 mortality (c-statistic = 0.722, 95%CI 0.690–0.753) and severity (c-statistic = 0.777, 95%CI 0.753–0.801). The Nutri-CoV score includes the MSL-COVID-19 plus respiratory rate, and pulse oximetry. This tool had better performance in both training (c-statistic = 0.797, 95%CI 0.765–0.826) and validation cohorts (c-statistic = 0.772, 95%CI 0.0.745–0.800) compared to other severity scores.

Conclusions

MSL-COVID-19 predicts inpatient COVID-19 lethality. The Nutri-CoV score is an adaptation of MSL-COVID-19 to be used in a triage environment. Both scores have been deployed as web-based tools for clinical use in a triage setting.

Introduction

The pandemic caused by the SARS-CoV2 virus, which is causative of COVID-19, has led to increased morbidity and mortality, posing challenges to healthcare systems worldwide [1]. Since its arrival in Mexico in late February 2020 to date, it has caused over 800,000 cases and more than 90,000 deaths attributable to COVID-19 [2]. Although most of these cases will remain mild or moderate, a group of patients could develop a severe to critical form of COVID-19 which will require quick medical assessment to prevent adverse clinical outcomes [3]. Most of these cases with severe to critical disease have underlying chronic cardiometabolic comorbidities (e.g., obesity, type 2 diabetes, arterial hypertension), chronic kidney disease, pulmonary obstructive disease and immunosuppression of any cause [4]. The high prevalence of obesity, type 2 diabetes and other chronic diseases, in addition with socioeconomic disparities, unequal access to prompt medical attention, and a lack of sufficient health infrastructure pose intense challenges in the public health care system in Mexico [5, 6]. As the disease continues to spread locally, the implementation of clinical tools to assess those patients with a higher risk for complications and lethality are needed [7, 8]. Recently, some predictive scores have been proposed and validated for COVID-19 in a hospital care setting, especially to predict critical illness hospitalization, intense care unit (ICU) admission, and mechanical ventilation support. However, these scores include specialized image acquisition or laboratories, which might not be available at most primary care scenarios in Mexico and elsewhere [911]. Recently, our group developed a novel mechanistic score for lethality attributable to COVID-19 (MSL-COVID-19) using age, self-reported comorbidities, and clinically suspected pneumonia, which performed adequately in a real-world scenario [12]. This tool considers the major contribution of chronic diseases to develop severe forms of COVID-19. Hence, the objective of the present study is to validate the MLS-COVID-19 to predict relevant hospitalization outcomes. Furthermore, we sought to develop an improved score based in the MLS-COVID-19 to predict disease severity including simple clinical measurements obtained in a triage setting.

Methods

Source of data and study population

This is a secondary analysis of the registry data of COVID-19 patients attended at the Instituto Nacional de Ciencias Médicas y Nutrición Salvador Zubirán (INMCNSZ). Briefly, we included patients aged >18 years with complete clinical data from March 16th to August 17th, 2020 who were evaluated at triage at INCMNSZ, a COVID-19 reference center in Mexico City, and had confirmed SARS-CoV-2 infection by RT-PCR test in respiratory samples. All patients were followed up until September 15th, 2020. All clinical procedures and measurements were approved by the INCMNSZ Research and Ethics Committee, written informed consent was waived due to the observational nature of the study. Amongst all evaluated patients within the study period, we considered consecutive patients with complete data to estimate the MSL-COVID-19 score (n = 3,007). This report adheres to the TRIPOD guidelines for development and validation of predictive models.

Outcome and timepoint setting

Clinical recovery was defined as hospital discharge based on the absence of clinical symptoms requiring inpatient management. ICU requirement was based on clinical judgment of the attending physician. Severe COVID-19 was determined as a composite event of either death, ICU admission requirement or mechanical ventilation; we estimated performance of metrics for this outcome at 7, 10, 15, 20 and 30 days, with a main focus being at 15 days given previous research indicating the disease course of mild COVID-19 cases. Follow-up time was estimated from date of symptom onset, considered as time zero of follow-up, up to last follow-up (censoring) or the composite event of severe COVID-19, which included death, requirement for invasive ventilation or ICU admission.

Predictor variables

Information was collected prospectively at the time of triage and emergency department evaluation; candidate predictor variables included predictors which were routinely measured in a triage setting without laboratory measurements. We only considered predictors available in >80% or participants. Predictors included demographic variables including age and sex, medical history of comorbidities including type 2 diabetes, obesity, chronic obstructive pulmonary disease (COPD), asthma, hypertension, immunosuppression, HIV, cardiovascular disease (CVD), chronic kidney disease (CKD), chronic liver disease (CLD), smoking habits, and current symptoms, as described elsewhere [13]. Physical examination included weight (measured in kilograms) and height (measured in meters) to estimate the body-mass index (BMI), vital signs including pulse oximetry (SpO2), respiratory rate (RR), heart rate (HR) and arterial blood pressure (BP). The Charlson Comorbidity Index (CCI), the National Early Warning Score (NEWS, NEWS2), and the quick Sequential Organ Failure Assessment (qSOFA) were also estimated to predict risk of severe COVID-19 [11, 14, 15]. Using this information, we also calculated the Mechanistic Score for COVID-19 Lethality (MSL-COVID-19), developed for prediction of COVID-19 case lethality in Mexicans using nation-wide case data (5).

Missing data

We included complete-case analysis for validation of MSL-COVID-19 under the assumption of data missing completely at random. Next, we analyzed patterns of missing data on clinical predictors to assess patterns of missing data and performed multiple imputation for variables assumed to be completely missing at random or missing at random using multivariable imputation with chained equations within the mice R package, generating five multiply imputed datasets and pooling the results of modeling using Rubin’s rules. Matrices of missing data are fully presented in S1 File.

Sample size estimation

Sample size was estimated for a time-to-event model using the guidelines proposed by Riley et al implemented with the pmsamplesize R package [16]. We considered the prior estimated c-statistic for the MSL-COVID-19 of 0.823 and the R2CS of 0.167, for an approximate event rate of 15% within the population, which accounted for 15-day risk during a mean 11.7 days of follow-up would represent 10 cases per 1,000 patient-days or an event rate of 0.010. Using all 35 candidate predictors, the minimum sample size required for new model development was estimated to be 1,706 with 19,994.3 person-days and 200 outcome events, for a minimum of 5.71 events per predictor parameter.

Statistical analysis

Outcome assessment

We compared patients with and without Severe COVID-19 using chi-squared tests for categorical variables and Student’s t-test or Mann-Whitney’s U for continuous variables depending on variable distribution. We evaluated each variable for prediction of mortality, ICU admission, requirement of mechanical invasive ventilation and/or severe disease using Kaplan-Meier curves and Cox Proportional Hazard Regression models; model assumptions were verified using Schöenfeld residuals. A multivariable model was fitted to assess for predictor independence with model selection being carried out using minimizations in the Bayesian Information Criterion (BIC).

MSL-COVID-19 model validation and predictors of severe disease

To validate the MSL-COVID-19 score we used Cox Proportional Hazard regression models on the continuous score and assessed model performance using Harrel’s c-statistic, Sommer’s Dxy score and the calibration slope in the overall sample for prediction of mortality, ICU and/or invasive ventilation requirement and severe/critical COVID-19. To estimate the optimism in these metrics, we carried out bootstrapping (B = 1,000) using the rms R package, using bias corrected accelerated (BCa) bootstrapping.

Derivation of the Nutri-CoV score

To increase precision of the MSL-COVID-19 score, we developed a two-step model whereby the score would be used as a first step and a second step would consider the previous risk and update it with clinical information. We included patients with complete clinical and physical examination data (n = 3,007) and split it into a training and validation datasets for patients admitted up to June 4th 2020 (training) and from that date up to August 17th (validation). To address overfitting and improve generalizability of our findings, we used Elastic Net Cox Regression, a regularization algorithm which handles multicollinearity and perform variable selection to increase generalizability using a λ penalization parameter and an α mixture parameter. We fitted an Elastic Net Cox Proportional Risk Regression Model using the both the MAMI and gmlnet R packages to incorporate coefficient estimation from multiply imputed data including al predictor variables depicted in Table 1; the ideal λ penalization parameter was estimated using k-fold cross-validation (k = 10) in the training dataset (S1 File). We selected the optimum alpha mixture parameter using simultaneous cross-validation for consecutive alpha mixture values ranging from 0 to 1 using 0.1 increments across each multiply imputed dataset and obtaining an average from all alpha values; to implement this, we used the cva.glmnet function of the glmnetUtils R package. We explored models including non-linear terms using restricted cubic splines, deciding the number of knots based on BIC minimization, and categorized models. We compared the performance of models with non-linear terms and categorized variables using cross-validation; since we observed only marginal decreases in performance with categorized variables, we proceeded with models using categorized predictors for simplification. Model coefficients from the Elastic Net Model were normalized to its ratio with the lowest absolute β coefficient to develop a point system and categories were developed to maximize separation assessed using Kaplan-Meier curves in the training dataset.

Table 1. Patient characteristics evaluated at triage or evaluation at the emergency department at our institution comparing severe and non-severe COVID-19 cases.
Parameter Non-severe COVID-19 (n = 2,432) Severe COVID-19 (n = 575) p-value
Age (years) 44 (33–55) 56 (47–66) <0.001
Male sex (%) 1,227 (50.5%) 403 (70.1%) <0.001
Time from symptom onset (Days) 5.0 (3.0–8.0) 7.0 (5.0–10.0) <0.001
BMI (kg/m2) 28.76±5.31 30.53±5.96 <0.001
SBP (mmHg) 123.98±19.10 126.60±17.32 0.003
DBP (mmHg) 77.62±12.00 74.47±12.56 <0.001
SpO2 (%) 89.65±7.41 68.31±15.97 <0.001
Temperature (°C) 36.76±0.75 36.99±0.76 <0.001
Heart rate (bpm) 100.48±16.96 103.45±18.46 0.003
Respiratory rate (bpm) 25.73±6.25 32.21±8.55 <0.001
Hospitalization 357 (32.8%) 237 (100.0%) <0.001
Obesity (%) 712 (29.3%) 268 (46.6%) <0.001
Overweight (%) 629 (32.6%) 215 (38.2%) <0.001
Diabetes (%) 387 (15.9%) 193 (33.6%) <0.001
COPD (%) 21 (0.9%) 10 (1.7%) 0.101
Asthma (%) 30 (3.1%) 2 (0.8%) 0.068
Immunosupression (%) 92 (3.8%) 33 (5.7%) 0.045
Hypertension (%) 440 (18.1%) 213 (37.1%) <0.001
HIV (%) 24 (1.0%) 3 (0.52%) 0.413
CVD (%) 71 (2.9%) 30 (5.2%) 0.008
CKD (%) 46 (1.8%) 16 (2.8%) 0.782
Fever (%) 1,784 (73.4%) 495 (86.1%) <0.001
Cough (%) 2,006 (82.5%) 520 (90.4%) 0.001
Headache (%) 1,909 (78.7%) 418 (73.1%) 0.004
Dyspnea (%) 901 (37.0%) 512 (89.0%) <0.001
Diarrhea (%) 757 (31.8%) 154 (29.6%) 0.345
Chest pain (%) 726 (30.5%) 211 (37.9%) <0.001
Myalgias (%) 1,512 (63.6%) 356 (64.0%) 0.833
Arthralgias (%) 1,390 (58.4%) 335 (60.5%) 0.399
Malaise (%) 1,795 (73.8%) 457 (79.5%) 0.005
Rhinorrhea (%) 929 (39.1%) 156 (28.4%) <0.001
Vomiting (%) 260 (10.9%) 73 (13.3%) 0.137
Abdominal pain (%) 147 (15.0%) 34 (14.4%) 0.809
Conjunctivitis (%) 477 (20.1%) 72 (13.1%) 0.001
Cyanosis (%) 51 (2.2%) 50 (9.15%) <0.001
qSOFA 1.0 (0.0–1.0) 1.0 (1.0–1.0) <0.001
NEWS 8.0 (7.0–9.0) 9.0 (8.0–10.0) <0.001
ABC-GOALS 3.0 (1.0–5.0) 6.0 (5.0–8.0) <0.001
MSL-COVID-19 0.0 (-5.0–4.0) 8.0 (7.0–9.0) <0.001

Abbreviations: BMI: Body-mass index; SBP: Systolic blood pressure; DBP: Diastolic blood pressure; SpO2: pulse oximetry; COPD: Chronic Obstructive Pulmonary Disease; HIV: Human Immunodeficiency Virus; CVD: Cardiovascular disease; CKD: Chronic Kidney Disease

Comparison of Nutri-CoV with other severity measures

We assessed the performance of the Nutri-CoV score in both the training and validation cohorts using a simple imputed dataset with mice and rms R packages, with correction for overoptimism being estimated using bias corrected accelerated (BCa) bootstrapping. We also estimated additional performance metrics including time-dependent sensitivity, specificity, positive and negative predictive values were obtained using the timeROC R packages using Inverse Probability of Censoring Weighting (IPCW) estimation of Cumulative/Dynamic time-dependent ROC curves for 7, 10, 15, 20 and 30 days after symptom onset with weighting performed using Cox proportional risk regression. Finally, we compared performance of MSL-COVID-19, ABC-GOALS, qSOFA, NEWS, NEWS2, and the CCI to predict disease severity using decision curve analyses estimated with the rmda R package. A p-value <0.05 was considered as statistical significance threshold. All analyses were performed using R software version 3.6.2. To facilitate the use of the score, we implemented both the MSL-COVID-19 and the Nutri-CoV score in a web-based tool using the ShinyApps R package hosted at: https://uiem.shinyapps.io/nutri_cov/, and also available in Spanish at https://uiem.shinyapps.io/nutri_cov_es/.

Sensitivity analyses

We performed recalculation of validation metrics using the rms package under the following scenarios: 1) Comparing participants with and without comorbidities, 2) excluding participants who were not yet discharged and were censored at the end of follow-up, 3) excluding participants who had severe COVID-19 upon admission, 4) comparing cases above and below 60 years of age, and 5) considering complete case analysis vs. multiply imputed analysis.

Results

Study population

We included 3,007 consecutive patients evaluated during the study period. Amongst them, 1317 patients were hospitalized at INCMNSZ (43.8%), 15 were transferred after initial evaluation (0.5%) and the rest (1,675, 55.7%) were sent for ambulatory management. The median time from symptom onset to initial clinical evaluation was 6.0 days (IQR 3.0–8.0) and for hospitalization was 7.0 days (IQR 5.0–10.0). Amongst those treated at our institution, 546 required ICU admission (18.2%) and 261 (8.7%) required mechanical invasive ventilation; median time from symptom onset to ICU admission was 8.0 days (IQR 3.5–14.0). We recorded 317 deaths within this population (10.5%), and the average time from symptom onset to death was 14.0 days (9.0–19.0). When accounting for severe COVID-19, we observed 575 cases (19.1%), with an average of 15.0 days (IQR 10.0–26.0) from symptom onset to severe outcome; overall, severe cases were older, most likely male and had higher BMI, RR, HR, body temperature and lower BP and SpO2 levels. Rates of obesity, overweight and diabetes were higher amongst severe COVID-19 cases (Table 1).

Predictors of COVID-19 disease severity

In univariate analyses using Cox proportional risk regression, we observed a higher risk of severe COVID-19 with increasing age (HR 1.024, 95%CI 1.018–1.031), less days from symptom onset to evaluation (HR 0.931, 95%CI 0.913–0.948), male sex (HR 1.258, 95%CI 1.051–1.506), higher BMI (HR 1.019, 95%CI 1.004–1.034), first (HR 1.231, 95%CI 1.157–1.309) and second restricted cubic splines of RR (HR 0.808, 95%CI 0.747–0.874), first (HR 0.992, 95%CI 0.985–0.999) and second restricted cubic splines of SpO2 (HR 0.872, 95%CI 0.845–0.901), type 2 diabetes (HR 1.365, 95%CI 1.147–1.625), and hypertension history (HR 1.408, 95%CI 1.188–1.669). When analyzing symptoms, dyspnea was a predictor of severe COVID-19 (HR 2.54, 95%CI 1.940–3.333), whilst headache (HR 0.873, 95%CI 0.688–0.996) was a protective factor. In the fully adjusted models, the only independent predictors for severe COVID-19 were increasing age, and higher RR, whilst higher SpO2 was a protective factor (Table 2). Inclusion of use of supplementary oxygen in the model did not attenuate the observed associations.

Table 2. Cox proportional risk regression models for prediction of severe COVID-19 using clinical characteristics at triage evaluation.

Model Parameter β HR (95%CI) p-value
c-statistic = 0.694 (95% CI 0.668–0.721) MSL-COVID-19 0.108 1.114 (1.088–1.140) <0.001
c-statistic = 0.790 (95%CI 0.767–0.813) Age 0.024 1.024 (1.018–1.031) <0.001
SpO2 (%) -0.028 0.972 (0.9670.977) <0.001
Respiratory rate 0.022 1.022 (1.012–1.033) <0.001
c-statistic = 0.791 (95%CI 0.768–0.813) MSL-COVID-19 0.081 1.085 (1.059–1.111) <0.001
SpO2 (%) -0.026 0.975 (0.969–0.980) <0.001
Respiratory rate 0.018 1.019 (1.008–1.029) <0.001

Validation of the MSL-COVID-19 score in the INCMNSZ cohort

For prediction of mortality, we identified a good discriminative capacity by the continuous values of the MSL-COVID-19 score as assessed by the calibration slope, c-statistic and Sommer’s Dxy of MSL-COVID-19 (Slope = 1.010, Dxy = 443; the c-statistic was 0.722 (95%CI 0.690–0.753). For prediction of ICU admission, the MSL-COVID-19 score also had good discriminative capacity (Slope = 0.992, Dxy = 0.663) with a slightly higher c-statistic (0.832, 95%CI 0.812–0.851); for intubation, the performance was not as good compared to the other outcomes (Slope = 0.781, Dxy = 0.885, c-statistic = 0.456, 95%CI 0.369–0.542). Overall, for prediction of severe COVID-19, we observed that MSL-COVID-19, MSL-COVID-19 had a good performance Slope = 0.781, Dxy = 0.885, c-statistic = 0.456, 95%CI 0.369–0.542). When the MSL-COVID-19 score was combined with RR and SpO2, prediction of severe/critical disease improved significantly, with slight overfitting, indicating the possibility of improved performance with inclusion of these metrics to improve risk assessment (Slope = 0.979, Dxy = 0.583, c-statistic = 0.793, 95%CI 0.771–0.815, Table 2).

Derivation of the Nutri-CoV score

Next, we sought to develop a score to predict disease severity including previous assessment with MSL-COVID-19. Using this Elastic Net Cox Regression, we confirmed that a combination of RR, SpO2 and the MSL-COVID-19 score improved prediction of severe COVID-19 (Fig 1). We fitted the model in the training cohort (n = 1831, 373 outcomes) and identified an adequate discriminative capacity (slope = 1.0013, Dxy = 0.523, c-statistic = 0.772, 95%CI 0.745–0.800). Using time-dependent ROC curves with IPCW obtained with the Kaplan-Meier estimator, we identified that a cut-off of 8.0 had 94.4% time-dependent sensitivity (95%CI 90.8–98.6%), 34.7% time-dependent specificity (95%CI 30.2–39.1%) to identify subjects with 15-day risk of severe COVID-19. The PPV and NPV were 22.4% (95%CI 18.9–25.9) and 96.9% (95%CI 94.7–99.1%). Using this cut-off, we observed 358 false positives and 15 false negatives; lowering the cut-off to 5 for low risk only yielded 3 false negatives and as such was chosen to represent the low-risk category in the training cohort.

Fig 1. Proposal for calculation of the Nutri-CoV score with initial estimation of the MSL-COVID-19 score for stratification of risk for severe COVID-19.

Fig 1

Abbreviations: RR, Respiratory Rate; SpO2: Oxygen saturation, CKD = Chronic Kidney Disease, COPD: Chronic Obstructive Pulmonary Disease.

Validation of the Nutri-CoV score in the validation cohort

Next, we applied the model to the validation cohort (n = 1,176, 202 outcomes), where the score also presented a good predictive capacity (slope = 0.999, Dxy = 0.592, c-statistic = 0.797, 95%CI 0.765–0.826). With time dependent AUROC, we observed that the cut-off of 8 had had 95.0% time-dependent sensitivity (95%CI 90.4–99.5%), 37.0% time-dependent specificity (95%CI 31.0–42.9%) to identify subjects with 15-day risk of severe COVID-19. The PPV and NPV were 22.4% (95%CI 18.1–27.1) and 96.9% (95%CI 95.0–99.8%). Using the low-risk category cut-off (Nutri-CoV<5) we identified only 1 false negative (Table 3). The proposed risk cut-offs in the training cohort significantly discriminate 15-day risk of severe COVID-19 in the validation cohort using Kaplan-Meier curves (log-rank p<0.001, Fig 2).

Table 3. Comparison of time-dependent sensitivity, specificity, positive and negative predictive values for Nutri-CoV in the validation cohort for prediction of severe COVID-19 at 7, 15 and 30 days using a Nutri-CoV cut-off of 8.

Time point Sensititivy (95%CI) Specificity (95%CI) PPV (95%CI) NPV (95%CI)
7 days 94.44 (83.86–100.0) 68.53 (65.69–71.37) 4.47 (2.39–6.55) 99.87 (99.61–100.0)
15 days 94.97 (90.40–99.54) 36.95 (30.95–42.95) 22.58 (18.05–27.11) 97.43 (95.04–99.82)
30 days 92.89 (88.09–97.69) 11.36 (1.97–20.75) 52.57 (44.26–60.88) 60.16 (33.62–86.70)

Fig 2. Kaplan-Meier analysis of risk categories for the Nutri-CoV score according to proposed cut-offs presented in Fig 1 for the training (A) and validation cohorts (B).

Fig 2

Comparison of Nutri-CoV with severity scores

Finally, we compared performance of Nutri-CoV compared to MSL-COVID-19, ABC-GOALS, the ROX index, qSOFA, and NEWS to predict COVID-19 severity overall and in the validation cohort (Table 4). We observed a significant and improved performance, as measures by time-dependent AUROC, for Nutri-CoV to predict 15-day risk of severe COVID-19 in comparison to all scores in the validation cohort. When assessing its performance compared to other indexes using decision curve analysis, significant clinical benefit was established for the Nutri-CoV score compared to other indexes (Fig 3).

Table 4. Comparison of severity scores with the developed Nutri-CoV score in the validation cohort using time-dependent area under the ROC curves (t-AUROC) using an age and sex adjusted Cox estimator.

Score c-statistic (95%CI) 7-day t-AUROC 15-day t-AUROC 30-day t-AUROC
Nutri-CoV 0.797 (0.765–0.829) 0.864 0.914 0.877
MSL-COVID-19 0.713 (0.669–0.757) 0.769 0.780 0.998
ABC-GOALS 0.759 (0.718–0.800) 0.844 0.836 0.856
Charlson comorbidity index 0.654(0.604–0.703) 0.710 0.641 0.536
qSOFA 0.657 (0.629–0.685) 0.679 0.752 0.720
NEWS 0.733 (0.696–0.770) 0.820 0.855 0.750
NEWS2 0.727 (0.690–0.765) 0.817 0.848 0.972

Fig 3. Density histogram of Nutri-CoV scores with the proposed cut-off for the low risk category within the validation cohort (A) an decision curve analysis to assess the clinical benefit of Nutri-CoV compared to other severity scores including MSL-COVID-19, NEWS, NEWS-2, qSOFA, Charlson comorbidity index and ABC-GOALS.

Fig 3

Sensitivity analyses

We conducted sensitivity analyses for relevant risk categories and to assess methodological decisions during score derivation (Table 5). Notably, Nutri-CoV observed improved performance in cases who were younger than age 60 and had no comorbidities, with decreased performance for use of only hospitalized cases. Notably, we observed a minimal decrease in performance when conducting complete-case analysis compared to multiply imputed data or exclusing cases who had not been discharged.

Table 5. Sensitivity analyses of Nurtri-CoV performance on selected groups in the validation cohort, comparing performances and calibration measures corrected for optimism using bias corrected accelerated bootstrapping (B = 1,000).

Subgroup N/Outcomes c-statistic (95%CI) Calibration slope Sommer’s Dxy
Imputed analysis 1,176/202 0.797 (0.765–0.829) 0.998 0.594
Complete-case analysis 1,080/196 0.789 (0.756–0.822) 0.995 0.577
>60 years 262/94 0.716 (0.663–0.769) 1.005 0.430
≤60 years 914/108 0.831 (0.792–0.871) 1.002 0.663
No comorbidities 552/36 0.895 (0.854–0.936) 0.985 0.791
≥1 comorbidity 624/166 0.735 (0.694–0.776) 1.005 0.470
Outcome-1 494/201 0.683 (0.643–0.724) 1.010 0.366
Outcome-2 1,079/196 0.789 (0.756–0.822) 1.002 0.579

Outcome-1 = Excluding cases who were not hospitalized; Outcome-2 = Excluding cases who were not yet discharged.

Discussion

Here, we validated the MSL-COVID-19 score for prediction of inpatient mortality at a COVID-19 reference center in Mexico City. We also demonstrated the role of MSL-COVID-19 in predicting severe and critical COVID-19 and how this estimation can be improved by considering additional clinical data obtained at triage. The repurposed score, which we named Nutri-CoV, includes demographics and comorbidity assessment as well as physical examination in the form of SpO2 and RR assessment, achieved significant discriminative capacity to detect potential cases of severe COVID-19. We deployed both the MSL-COVID-19 and the Nutri-CoV scores onto a web-based tool which could be readily used within a triage setting to identify individuals at the highest risk of developing COVID-19 complications. Our model could be used to make prompt decisions regarding timely admission and treatment initiation in patients at risk of severe and critical COVID-19, as well as resource allocation in the setting of increased healthcare stress during pandemic peaks. Given the remarkable increase in the availability of COVID-19 predictive models, we expect the derivation of our score will be helpful in triage settings with similarities to our institution; however, extensive external validation and calibration studies are required to evaluate its performance in a triage prior to clinical utilization [17]. Given the conducted sensitivity analyses, we do not currently recommend the use of our score for in-hospital clinical deterioration, which could be assessed with other available tools after due external validation [18].

Underlying factors which explain the pathogenesis of severe COVID-19 complications have been extensively studied [19, 20]. A role for metabolic comorbidities including type 2 diabetes and its associated glycemic control, obesity, hypertension, older age and male sex has been reported as risk factors for severe COVID-19. In Mexicans, our group developed the MSL-COVID-19 score, which attempts to capture the risk of most of these factors with a particular focus on predicting mortality [12]. Despite the relevance of estimating mortality risk in COVID-19, the elevated healthcare stress which has been steadily observed during the pandemic makes the consideration of ICU admission and requirement for invasive intubation relevant, especially for resource allocation and prompt treatment initiation of high-risk patients [21, 22]. However, controversy surrounding early intubation for severe COVID-19 could make identification of these cases relevant for close clinical follow-up and evaluation. Previous severity scores developed for COVID-19 consider radiological and laboratory findings in relation to clinical status and comorbidities [10]. Nevertheless, a main concern in low-resource settings is the availability of healthcare resources and a highly sensitive clinical score might be relevant to detect severe disease; whilst the MSL-COVID-19 score is useful for quick clinical examination, updating these estimations with current clinical status improves its predictive capacity. Nutri-CoV shows excellent discriminative capacity for low-risk cases, with a low number of eventual false positives which makes it a good screening tool to predict disease severity and resource allocation.

Onset of acute respiratory distress syndrome (ARDS) in COVID-19 patients can be attributable to underlying endothelial vascular injury which disrupts the regulation of pulmonary blood flow, leading to a ventilation-perfusion mismatch and eventually reducing oxygenation despite increased lung compliance and thrombogenesis [23]. Low oxygenation can be promptly and objectively assessed using pulse oximetry, allowing identification of COVID-19 related early hypoxia, particularly when no overt dyspnea is present [24]. This pattern of lung injury, related to decreased elastance in the setting of increased lung compliance has been described as Type L COVID pneumonia, which has different implications compared to type H COVID pneumonia, more traditionally linked to ARDS [2426]. Hypoxemia usually leads to increase in minute ventilation by increasing tidal volume, which can be assessed as an increased respiratory rate despite absence of dyspnea in the setting or normal lung compliance. Notably, lung function is impaired in comorbidities linked to increased risk of severe COVID-19 including obesity and type 2 diabetes [2729]; furthermore, most cardio-metabolic comorbidities have been linked to increased expression of the angiotensin converting enzyme (ACE) [30, 31]. These factors are jointly evaluated by the Nutri-CoV and MSL-COVID-19 scores, thus indicating its potential to capture pathophysiological components that eventually lead to severe COVID-19 disease and increased mortality. Notably, although dyspnea was associated with increased disease severity, it was not an independent predictor severe disease, probably because of collinearity with variables already included such as pulse oximetry and respiratory rate. The use of our algorithm should accompany clinical judgement as the evaluation of any individual component of the score does not predict the outcome with absolute certainty.

Our study had some strengths and limitations. By assessing patients who were considered to have mild, moderate and severe COVID-19, we were able to capture a wide array of clinical characteristics at triage evaluation. Nevertheless, since all patients were examined from a single center, the possibility of referral or representability bias might reduce performance of Nutri-CoV in other populations. MSL-COVID-19 was developed using nationally representative data, the good performance shown within our institution, validates its implementation in low resource settings. A potential limitation of our approach is that pulse oximetry evaluation was performed in patients who are already on supplementary oxygen; whilst this was controlled in the statistical analysis and considered within the Elastic Net regression framework, the possibility of residual confounding is present. Considering the use of a training and a validation dataset as well as cross-validation for selection of penalization parameters, the model is likely to have adequate performance; however, our model is better suited for adequate performance within our institution or similar institutions within of Mexico and thus requires additional external validation to confirm its applicability in other settings. The web application should be useful to physicians within our institution and for external settings once it is externally validated and could be complementary to decision making in a triage setting.

In conclusion, we validated the MSL-COVID-19 score using data from a COVID-19 reference center in Mexico City. We demonstrated this score could be useful to predict other outcomes related to COVID-19 including ICU admission and requirement for invasive ventilation. Including clinical parameters in the MSL-COVID-19 score, related to COVID-19 ventilatory pathophysiology, increases its performance for prediction of severe COVID-19 and makes it a useful tool for clinical decision making and resource allocation during the healthcare stress of a pandemic scenario. Both the MSL-COVID-19 and Nutri-CoV scores were deployed as interactive webtools for clinical use in a triage setting.

Supporting information

S1 File

(DOCX)

Acknowledgments

AVV, NEAV are enrolled at the PECEM program of the Faculty of Medicine at UNAM. NEAV and AVV are supported by CONACyT. The authors would like to acknowledge the invaluable work of all healthcare workers at the Instituto Nacional de Ciencias Médicas y Nutrición Salvador Zubirán for its community in managing the COVID-19 epidemic. Its participation in the COVID-19 surveillance program has made this work a reality, we are thankful for your effort.

Data Availability

Code for verification of all analytical procedures is available at https://github.com/oyaxbell/nutri_cov. Data cannot be made publicly available due to ethical restrictions, however they are available on request from the Institutional Data Access Ethics Committee with contact via the corresponding author or the president of the ethics committee at arturo.galindof@incmnsz.mx, for researchers who meet the criteria for access to confidential data. For more information please consult: https://www.incmnsz.mx/opencms/contenido/investigacion/comiteEtica/miembros_comite.html.

Funding Statement

The author(s) received no specific funding for this work.

References

  • 1.Clark A, Jit M, Warren-Gash C, Guthrie B, Wang HHX, Mercer SW, et al. Global, regional, and national estimates of the population at increased risk of severe COVID-19 due to underlying health conditions in 2020: a modelling study. The Lancet Global Health. 2020; 10.1016/S2214-109X(20)30264-3 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 2.Datos Abiertos—Dirección General de Epidemiología | Secretaría de Salud | Gobierno | gob.mx [Internet]. [cited 18 Apr 2020]. https://www.gob.mx/salud/documentos/datos-abiertos-152127?idiom=es
  • 3.Verity R, Okell LC, Dorigatti I, Winskill P, Whittaker C, Imai N, et al. Estimates of the severity of coronavirus disease 2019: a model-based analysis. Lancet Infect Dis. 2020;20: 669–677. 10.1016/S1473-3099(20)30243-7 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4.Guan W-J, Liang W-H, Zhao Y, Liang H-R, Chen Z-S, Li Y-M, et al. Comorbidity and its impact on 1590 patients with COVID-19 in China: a nationwide analysis. Eur Respir J. 2020;55 10.1183/13993003.00547-2020 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5.Abrams EM, Szefler SJ. COVID-19 and the impact of social determinants of health. Lancet Respir Med. 2020; 10.1016/S2213-2600(20)30234-4 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6.Bello-Chavolla OY, González-Díaz A, Antonio-Villa NE, Fermín-Martínez CA, Márquez-Salinas A, Vargas-Vázquez A, et al. Unequal impact of structural health determinants and comorbidity on COVID-19 severity and lethality in older Mexican adults: Considerations beyond chronological aging. J Gerontol A, Biol Sci Med Sci. 2020; 10.1093/gerona/glaa163 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7.Brown MJ, Goodwin J. Allocating Medical Resources in the Time of Covid-19. N Engl J Med. 2020;382: e79 10.1056/NEJMc2009666 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.Farrell TW, Francis L, Brown T, Ferrante LE, Widera E, Rhodes R, et al. Rationing Limited Health Care Resources in the COVID-19 Era and Beyond: Ethical Considerations Regarding Older Adults. J Am Geriatr Soc. 2020; 10.1111/jgs.16539 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9.Mejia-Vilet JM, Cordova-Sanchez BM, Fernandez-Camargo D, Mendez-Perez RA, Morales-Buenrostro LE, Hernandez-Gilsoul T. A Risk Score to Predict Admission to Intensive Care Unit in Patients With COVID-19: The ABC-GOALS Score. medRxiv. 2020; 10.1101/2020.05.12.20099416 [DOI] [PubMed] [Google Scholar]
  • 10.Liang W, Liang H, Ou L, Chen B, Chen A, Li C, et al. Development and Validation of a Clinical Risk Score to Predict the Occurrence of Critical Illness in Hospitalized Patients With COVID-19. JAMA Intern Med. 2020; 10.1001/jamainternmed.2020.2033 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11.Roca O, Caralt B, Messika J, Samper M, Sztrymf B, Hernández G, et al. An Index Combining Respiratory Rate and Oxygenation to Predict Outcome of Nasal High-Flow Therapy. Am J Respir Crit Care Med. 2019;199: 1368–1376. 10.1164/rccm.201803-0589OC [DOI] [PubMed] [Google Scholar]
  • 12.Bello-Chavolla OY, Bahena-López JP, Antonio-Villa NE, Vargas-Vázquez A, González-Díaz A, Márquez-Salinas A, et al. Predicting mortality due to SARS-CoV-2: A mechanistic score relating obesity and diabetes to COVID-19 outcomes in Mexico. J Clin Endocrinol Metab. 2020; 10.1210/clinem/dgaa346 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.Ortiz-Brizuela E, Villanueva-Reza M, González-Lara MF, Tamez-Torres KM, Román-Montes CM, Díaz-Mejía BA, et al. Clinical and epidemiological characteristics of patients diagnosed with covid-19 in a tertiary care center in mexico city: a prospective cohort study. Rev Invest Clin. 2020;72: 165–177. 10.24875/RIC.20000211 [DOI] [PubMed] [Google Scholar]
  • 14.Seymour CW, Liu VX, Iwashyna TJ, Brunkhorst FM, Rea TD, Scherag A, et al. Assessment of Clinical Criteria for Sepsis: For the Third International Consensus Definitions for Sepsis and Septic Shock (Sepsis-3). JAMA. 2016;315: 762–774. 10.1001/jama.2016.0288 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15.National Early Warning Score (NEWS) 2 | RCP London [Internet]. [cited 15 Jul 2020]. https://www.rcplondon.ac.uk/projects/outputs/national-early-warning-score-news-2
  • 16.Riley RD, Ensor J, Snell KIE, Harrell FE, Martin GP, Reitsma JB, et al. Calculating the sample size required for developing a clinical prediction model. BMJ. 2020;368: m441 10.1136/bmj.m441 [DOI] [PubMed] [Google Scholar]
  • 17.Update to living systematic review on prediction models for diagnosis and prognosis of covid-19. BMJ. 2020;370: m2810 10.1136/bmj.m2810 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18.Gupta RK, Harrison EM, Ho A, Docherty AB, Knight SR, van Smeden M, et al. Development and validation of the 4C Deterioration model for adults hospitalised with COVID-19. medRxiv. 2020; 10.1101/2020.10.09.20209957 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19.Zhou F, Yu T, Du R, Fan G, Liu Y, Liu Z, et al. Clinical course and risk factors for mortality of adult inpatients with COVID-19 in Wuhan, China: a retrospective cohort study. Lancet. 2020;395: 1054–1062. 10.1016/S0140-6736(20)30566-3 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 20.Grasselli G, Zangrillo A, Zanella A, Antonelli M, Cabrini L, Castelli A, et al. Baseline Characteristics and Outcomes of 1591 Patients Infected With SARS-CoV-2 Admitted to ICUs of the Lombardy Region, Italy. JAMA. 2020; 10.1001/jama.2020.5394 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21.Shen C, Wang Z, Zhao F, Yang Y, Li J, Yuan J, et al. Treatment of 5 Critically Ill Patients With COVID-19 With Convalescent Plasma. JAMA. 2020; 10.1001/jama.2020.4783 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22.Grein J, Ohmagari N, Shin D, Diaz G, Asperges E, Castagna A, et al. Compassionate Use of Remdesivir for Patients with Severe Covid-19. N Engl J Med. 2020;382: 2327–2336. 10.1056/NEJMoa2007016 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 23.Marini JJ, Gattinoni L. Management of COVID-19 Respiratory Distress. JAMA. 2020; 10.1001/jama.2020.6825 [DOI] [PubMed] [Google Scholar]
  • 24.Gattinoni L, Chiumello D, Caironi P, Busana M, Romitti F, Brazzi L, et al. COVID-19 pneumonia: different respiratory treatments for different phenotypes? Intensive Care Med. 2020; 10.1007/s00134-020-06033-2 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25.Ziehr DR, Alladina J, Petri CR, Maley JH, Moskowitz A, Medoff BD, et al. Respiratory Pathophysiology of Mechanically Ventilated Patients with COVID-19: A Cohort Study. Am J Respir Crit Care Med. 2020;201: 1560–1564. 10.1164/rccm.202004-1163LE [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 26.Wilkerson RG, Adler JD, Shah NG, Brown R. Silent hypoxia: A harbinger of clinical deterioration in patients with COVID-19. Am J Emerg Med. 2020; 10.1016/j.ajem.2020.05.044 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27.Simonnet A, Chetboun M, Poissy J, Raverdy V, Noulette J, Duhamel A, et al. High prevalence of obesity in severe acute respiratory syndrome coronavirus-2 (SARS-CoV-2) requiring invasive mechanical ventilation. Obesity (Silver Spring). 2020; 10.1002/oby.22831 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 28.Dixon AE, Peters U. The effect of obesity on lung function. Expert Rev Respir Med. 2018;12: 755–767. 10.1080/17476348.2018.1506331 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 29.Cariou B, Hadjadj S, Wargny M, Pichelin M, Al-Salameh A, Allix I, et al. Phenotypic characteristics and prognosis of inpatients with COVID-19 and diabetes: the CORONADO study. Diabetologia. 2020; 10.1007/s00125-020-05180-x [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 30.Rao S, Lau A, So H-C. Exploring diseases/traits and blood proteins causally related to expression of ACE2, the putative receptor of 2019-nCov: A Mendelian Randomization analysis. medRxiv. 2020; 10.2337/dc20-0643 [DOI] [PubMed] [Google Scholar]
  • 31.Li M-Y, Li L, Zhang Y, Wang X-S. Expression of the SARS-CoV-2 cell receptor gene ACE2 in a wide variety of human tissues. Infect Dis Poverty. 2020;9: 45 10.1186/s40249-020-00662-x [DOI] [PMC free article] [PubMed] [Google Scholar]

Decision Letter 0

Itamar Ashkenazi

Transfer Alert

This paper was transferred from another journal. As a result, its full editorial history (including decision letters, peer reviews and author responses) may not be present.

28 Sep 2020

PONE-D-20-25115

Validation and repurposing of the MSL-COVID-19 score for prediction of severe COVID-19 using simple clinical predictors in a triage setting: The Nutri-CoV score

PLOS ONE

Dear Dr. Aguilar-Salinas,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

Please submit your revised manuscript by Nov 12 2020 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: http://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols

We look forward to receiving your revised manuscript.

Kind regards,

Itamar Ashkenazi

Academic Editor

PLOS ONE

Journal Requirements:

When submitting your revision, we need you to address these additional requirements.

1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at

https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and

https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf

2. We note that you state in your ethics statement in the online submission form "proceedings were approved by the INCMNSZ Research and Ethics Committee". Please clarify whether your ethics committee specifically approved this study and provide the full name of this committee.

3. Please include the date(s) on which you accessed the databases or records to obtain the data used in your study.

4. We note that you have indicated that data from this study are available upon request. PLOS only allows data to be available upon request if there are legal or ethical restrictions on sharing data publicly. For information on unacceptable data access restrictions, please see http://journals.plos.org/plosone/s/data-availability#loc-unacceptable-data-access-restrictions.

In your revised cover letter, please address the following prompts:

a) If there are ethical or legal restrictions on sharing a de-identified data set, please explain them in detail (e.g., data contain potentially identifying or sensitive patient information) and who has imposed them (e.g., an ethics committee). Please also provide contact information for a data access committee, ethics committee, or other institutional body to which data requests may be sent.

b) If there are no restrictions, please upload the minimal anonymized data set necessary to replicate your study findings as either Supporting Information files or to a stable, public repository and provide us with the relevant URLs, DOIs, or accession numbers. Please see http://www.bmj.com/content/340/bmj.c181.long for guidelines on how to de-identify and prepare clinical data for publication. For a list of acceptable repositories, please see http://journals.plos.org/plosone/s/data-availability#loc-recommended-repositories.

We will update your Data Availability statement on your behalf to reflect the information you provide.

5. PLOS requires an ORCID iD for the corresponding author in Editorial Manager on papers submitted after December 6th, 2016. Please ensure that you have an ORCID iD and that it is validated in Editorial Manager. To do this, go to ‘Update my Information’ (in the upper left-hand corner of the main menu), and click on the Fetch/Validate link next to the ORCID field. This will take you to the ORCID site and allow you to create a new iD or authenticate a pre-existing iD in Editorial Manager. Please see the following video for instructions on linking an ORCID iD to your Editorial Manager account: https://www.youtube.com/watch?v=_xcclfuvtxQ

6. Please ensure that you refer to Figure 2 in your text as, if accepted, production will need this reference to link the reader to the figure.

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Partly

Reviewer #2: Yes

**********

2. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: No

Reviewer #2: Yes

**********

3. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: No

**********

4. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

**********

5. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: I have a number of concerns regarding statistical design, analysis, and representation. I will try to go in order of line number.

Line 112. It is unclear to me whether or not the univariate analyses influenced which variables were used in the multivariable model which used BIC for model selection. For example, were only variables that were considered significantly associated with the outcome (p<0.05) used in the multivariable analysis?

Line 116-119. Was the data used to make/decide the 4 risk group stratification the same data that was used to validate it? Or was the validation set used to validate it?

Line 120. For bootstrapping, I assume the ordinary bootstrap method was used. Which confidence interval was used for the provided metrics? Percentile? BCA? etc. It is worth mentioning that the validation functions in the rms package are not intended to be used in complicated modeling designs wherein decisions such as cutoffs for continuous variables, risk stratification, etc. with the same data that is used to validate it. The result, optimistic estimates for performance. Even if you use the bootstrap later because you are not including data-driven decisions in the bootstrap process. You will likely need to code the loop yourself to include these elements in the bootstrap process.

Line 129 indicates that lambda is the tuning parameter of the Elastic-Net model, but this is only half true. Lambda is the penalty parameter in a LASSO model -- Elastic-Net adds in a mixture parameter, typically represented as alpha, that determines how much LASSO penalty and how much ridge penalty is used. Can the authors please tell me how much mix, if any between the two were used? This does not have to appear in the main paper, but some more details in an appendix or similar would be helpful. This includes how the hyperparameter optimization was done -- exhaustive grid search, randomly sampled xxx points from the hypergrid, etc.

Line 131. glmnet instead of gmlnet.

Line 132. For design, it appears that there may be a typo with regards to the number of folds, 20.

If k is indeed 20, then within the CV routine 1/20th of the sample used as the test set at any given point. This amounts to about 58 patients, approximately 6 of which will be events. In my view, this is too few. I strongly recommend that the authors instead use a 10 fold or perhaps a 5 fold CV.

Line 136. likelihood ratios instead of likelihood rations

Line 174, 178, 187. why are the c-statistics and AUC different? did you have censoring? if so, the AUC might have been calculated using a method that ignores this.

Line 190. Was the cutoff created in the training cohort only?

Line 190 - 197. Are these estimates all from the validation cohort?

Line 259 "thus increasing the external validity of our model" how? including more variables does not necessarily increase the external validity. would remove this.

Line 268. "the model is likely to have adequate performance when applied in external datasets" you have no external data to suggest this.

Missing limitation elements from discussion. For example, the study is skewed in the direction of your model because the data used to validate your model is very similar to the data used to fit/train your model. The other models under evaluation here were, if my understanding is correct, were derived using data from other institutions. If you compare how well their respective papers said they would perform and compare that to how well they performed on your data, you will likely see that their models did not perform as well.

The limitations here as I see it is three-fold: 1) repeated misuse of the bootstrap method, specifically not including cutoff and risk stratification inside the validation; 2) too many folds. i suspect that you will be better off using 5-fold or some such CV. Repeating the CV routine many times (repeated k-fold CV) would be one step better than that.; 3) recognize in the discussion that your model will inherently do better than models that were created outside of your institution.

Please also remake the plots a higher quality file as they are not legible in their current state.

Reviewer #2: In their paper " Validation and repurposing of the MSL-COVID-19 score for prediction of severe COVID-19 using simple clinical predictors in a triage setting: The Nutri-CoV score", Bello-Chavolla et al. develop prediction scores for several aspects of the COVID-19 disease using time-to-event models and present applications. In principle, the analysis is sound. However, I have one major point the authors have to address and a couple of minor comments.

Major point:

- In time-to-event models it is very important to define the time-scale exactly. Time zero is not defined in the paper. The authors write "We included patients aged >18 years with complete clinical data from March 16th to June 4th, 2020 who were evaluated at triage at the Instituto Nacional de Ciencias Médicas y Nutrición Salvador Zubirán (INCMNSZ), a COVID-19 reference center in Mexico City...". I would therefore conclude that the timepoint of presentation is taken to be timepoint zero. This would be problematic as it creates an immortal time bias for the presenting population. The interpretation of the model would then only be applicable in a setting where the same pattern of presenting at the triag unit would be present. A more logical timepoint zero would be the onset of symptoms, a population that would be well-defined. At the very least the authors need to perform a sensitivy analysis for such model, using time-dependent covariates to account for the immortal time bias

Minor points:

- Wording l. 81: "all proceedings" is somewhat unclear (all analyses?)

- l. 107: It is only meaningful to to perform the "outcome assessment" at several time horizons which need to be specified. It is not meaningful to use the data set "as is", i.e. only the presence of the outcome at any timepoint.

- l. 112: As the assumptions are checked visually, the plot need to be supplied as a supplement.

- l. 136 (and other places): The curves need to be evaluated at time horizons. These need to be clearly specified.

- Some of the predictors are known to act non-linearly such as age. As penalized regression is used, the authors are advised to include non-linear termas, such as quadratic terms.

- The authors should stress that the validation is internal and not external. The authors should also use test-data instead of validation data to avoid confusion when describing cross-validation (the cross-validation uses the validation data, evaluation is performed in the test data).

- The model should be evaluated in a stratified way using the known risk groups, as this is highly relevant for practical application.

- The authors need to provide a URL for the data set.

**********

6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

PLoS One. 2020 Dec 16;15(12):e0244051. doi: 10.1371/journal.pone.0244051.r002

Author response to Decision Letter 0


10 Nov 2020

November 2nd, 2020

Dear Prof. Itamar Ashkenazi,

Academic Editor, PLOS ONE.

Thank you for giving us the opportunity to revise our work. We have addressed the reviewers’ invaluable comments and modified the manuscript accordingly. We believe the manuscript to be better after these revisions. Considering recent evidence on the quality of predictive modeling in COVID-19, we have rewritten several sections of our manuscript to adhere to TRIPOD guidelines and expanded our dataset to account for most recent cases, with an overall larger sample size (n=3,007). Below, we include a detailed response to each of the reviewer’s inquiries. We appreciate the opportunity to submit our work for consideration by PLoS One.

Sincerely,

Carlos A. Aguilar-Salinas, MD PhD

Corresponding author

Reviewer #1:

I have a number of concerns regarding statistical design, analysis, and representation. I will try to go in order of line number.

R= Thank you for your revision, the points you raised helped us re-evaluate our statistical design and provide more transparent reporting, we are thankful for your suggestions.

Line 112. It is unclear to me whether or not the univariate analyses influenced which variables were used in the multivariable model which used BIC for model selection. For example, were only variables that were considered significantly associated with the outcome (p<0.05) used in the multivariable analysis?

R= We evaluated all variables presented in Table 1 for inclusion in the model, regardless of their univariate association with Severe COVID-19 to rule out potential exclusion of variables which were associated but failed to reach a significance threshold prior to adjustment. We have now adhered our reporting to the TRIPOD guidelines and have included this specification into the revised version of our manuscript.

Line 116-119. Was the data used to make/decide the 4 risk group stratification the same data that was used to validate it? Or was the validation set used to validate it?

R= We used validation data to validate these cut-offs which were obtained in the original derivation paper for the MSL-COVID-19 score (https://doi.org/10.1210/clinem/dgaa346). For the Nutri-CoV score, we estimated risk categories within the training cohort and then evaluated in the validation cohort.

Line 120. For bootstrapping, I assume the ordinary bootstrap method was used. Which confidence interval was used for the provided metrics? Percentile? BCA? etc. It is worth mentioning that the validation functions in the rms package are not intended to be used in complicated modeling designs wherein decisions such as cutoffs for continuous variables, risk stratification, etc. with the same data that is used to validate it. The result, optimistic estimates for performance. Even if you use the bootstrap later because you are not including data-driven decisions in the bootstrap process. You will likely need to code the loop yourself to include these elements in the bootstrap process.

R= We agree with your comments. The development and internal validation of the MSL-COVID-19 score was carried out in a separate dataset and we now validated it using data from our institution. We used bootstrapping (B=1,000) to correct for overoptimism with bias corrected accelerated boostrapping when validating all scores using the rms package; for this purpose, we only used the continuous scores and not the risk categories. Since this is an external validation of the MSL-COVID-19 model, we expect the score to have a less optimistic performance compared to the derivation dataset. We have included these methodological details into the revised version of our manuscript.

Line 129 indicates that lambda is the tuning parameter of the Elastic-Net model, but this is only half true. Lambda is the penalty parameter in a LASSO model – Elastic-Net adds in a mixture parameter, typically represented as alpha, that determines how much LASSO penalty and how much ridge penalty is used. Can the authors please tell me how much mix, if any between the two were used? This does not have to appear in the main paper, but some more details in an appendix or similar would be helpful. This includes how the hyperparameter optimization was done – exhaustive grid search, randomly sampled xxx points from the hypergrid, etc.

R= Thank you for your comment, we selected the optimum alpha mixture parameter using simultaneous cross-validation for consecutive alpha mixture values ranging from 0 to 1 using 0.1 increments in each multiply imputed dataset and obtaining an average from all alpha values; to implement this, we used the cva.glmnet function of the glmnetUtils R package. We then used the identified alpha value to fit the Elastic Net model using both MAMI and glmnet R packages to consider multiply imputed data and to obtain the Elastic Net estimator as previously described:

1. http://mami.r-forge.r-project.org/MAMI_manual.pdf

2. Schomaker, M. (2012). Shrinkage averaging estimation. Statistical Papers 53, 1015–1034.

Line 131. Glmnet instead of gmlnet.

R= Thank you, we have corrected this typo.

Line 132. For design, it appears that there may be a typo with regards to the number of folds, 20. If k is indeed 20, then within the CV routine 1/20th of the sample used as the test set at any given point. This amounts to about 58 patients, approximately 6 of which will be events. In my view, this is too few. I strongly recommend that the authors instead use a 10-fold or perhaps a 5 fold CV.

R= We have updated the sample size for model development and re-considered the data splitting method to include model development from March 16th to June 4th and validation From June 5th to August 17th in accordance with TRIPOD guidelines. Using this technique, we ended up with a training sample of 1,831 (373 events) and a validation sample of 1176 (202 events). We performed cross-validation using k=5 and k=10, we present the results of coefficient comparison in Supplementary Material.

Line 136. likelihood ratios instead of likelihood rations

R= Thank you, this has now been corrected.

Line 174, 178, 187. Why are the c-statistics and AUC different? Did you have censoring? If so, the AUC might have been calculated using a method that ignores this.

R= Thank you for your observation. We re-estimated c-statistics and AUC to account for censoring, we are now only presenting c-statictics and its corresponding confidence intervals to assess model performance. We also assessed time-dependent ROC curves and time-dependent sensitivity, specificity, positive and negative predictive values, which were estimated using the timeROC R package.

Line 190. Was the cutoff created in the training cohort only?

R= Yes, the cut-off was created in the training cohort and subsequently validated in the validation cohort. We have included this specification in the revised version of the manuscript.

Line 190 - 197. Are these estimates all from the validation cohort?

R= We have estimated metrics for both training and validation cohorts and included them within the revised version of the manuscript. For model comparison and sensitivity analyses, we only used data from the validation cohort.

Line 259 "thus increasing the external validity of our model" how? including more variables does not necessarily increase the external validity. would remove this.

R= We agree with your observation, we have removed this comment.

Line 268. "the model is likely to have adequate performance when applied in external datasets" you have no external data to suggest this.

R= Thank you, we agree with your observation. We have removed this comment from the revised version of our manuscript.

Missing limitation elements from discussion. For example, the study is skewed in the direction of your model because the data used to validate your model is very similar to the data used to fit/train your model. The other models under evaluation here were, if my understanding is correct, were derived using data from other institutions. If you compare how well their respective papers said they would perform and compare that to how well they performed on your data, you will likely see that their models did not perform as well.

R= We agree with your assertion, we have included this within the limitations of our work in the revised version of the manuscript. The revised section reads: “however, our model is likely to perform better within our institutions or similar institutions outside of Mexico and require additional external validation to confirm its applicability in other settings”.

The limitations here as I see it is three-fold: 1) repeated misuse of the bootstrap method, specifically not including cutoff and risk stratification inside the validation.

R= To account for this, we have only validated the continuous score under the assumption that risk categories could perform differently across different populations.

2) too many folds. i suspect that you will be better off using 5-fold or some such CV. Repeating the CV routine many times (repeated k-fold CV) would be one step better than that.

R= All models have been fit with both k-fold cross-validation, with k=5 and k=10; we present comparison of these results in Supplementary Material.

3) recognize in the discussion that your model will inherently do better than models that were created outside of your institution.

Please also remake the plots a higher quality file as they are not legible in their current state.

R= Thank you, we have now appended high quality files for all our images.

Reviewer #2: In their paper " Validation and repurposing of the MSL-COVID-19 score for prediction of severe COVID-19 using simple clinical predictors in a triage setting: The Nutri-CoV score", Bello-Chavolla et al. develop prediction scores for several aspects of the COVID-19 disease using time-to-event models and present applications. In principle, the analysis is sound. However, I have one major point the authors have to address and a couple of minor comments.

Major point:

- In time-to-event models it is very important to define the time-scale exactly. Time zero is not defined in the paper. The authors write "We included patients aged >18 years with complete clinical data from March 16th to June 4th, 2020 who were evaluated at triage at the Instituto Nacional de Ciencias Médicas y Nutrición Salvador Zubirán (INCMNSZ), a COVID-19 reference center in Mexico City...". I would therefore conclude that the timepoint of presentation is taken to be timepoint zero. This would be problematic as it creates an immortal time bias for the presenting population. The interpretation of the model would then only be applicable in a setting where the same pattern of presenting at the triag unit would be present. A more logical timepoint zero would be the onset of symptoms, a population that would be well-defined. At the very least the authors need to perform a sensitivy analysis for such model, using time-dependent covariates to account for the immortal time bias.

R= Thank you for your observation, we have included a detailed description of time zero and time estimation to reduce the likelihood of immortal time bias. We have modified the manuscript as follows: “Briefly, we included patients aged >18 years with complete clinical data from March 16th to August 17th, 2020 who were evaluated at triage at INCMNSZ, a COVID-19 reference center in Mexico City, and had confirmed SARS-CoV-2 infection by RT-PCR test in respiratory samples. All patients were followed up until September 15th, 2020 […] Follow-up time was estimated from date of symptom onset, considered as time zero of follow-up, up to last follow-up (censoring) or the composite event of severe COVID-19, which included death, requirement for invasive ventilation or ICU admission.”

Minor points:

- Wording l. 81: "all proceedings" is somewhat unclear (all analyses?)

R= We referred to all clinical procedures and measurements. We have now corrected this.

- l. 107: It is only meaningful to to perform the "outcome assessment" at several time horizons which need to be specified. It is not meaningful to use the data set "as is", i.e. only the presence of the outcome at any timepoint. - l. 136 (and other places): The curves need to be evaluated at time horizons. These need to be clearly specified.

R= We agree with your comment, we have now estimated performance of all risk metrics to estimate risk at days 7, 15, and 30-days using time dependent ROC curves.

- l. 112: As the assumptions are checked visually, the plot need to be supplied as a supplement.

R= We verified assumptions for all models using hypothesis testing over model residuals in addition to visual inspection. We have included those graphs in Supplementary Material.

- Some of the predictors are known to act non-linearly such as age. As penalized regression is used, the authors are advised to include non-linear terms, such as quadratic terms.

R= We evaluated non-linear associations using restricted cubic splines and compared the performance of the model using continuous models with splines compared to categorized variables. Since we wanted the score to become widely applicable within our institution, we opted to use categorized variables; however, we have included the restricted cubic spline modeling and non-linear associations in Supplementary Materials.

- The authors should stress that the validation is internal and not external. The authors should also use test-data instead of validation data to avoid confusion when describing cross-validation (the cross-validation uses the validation data, evaluation is performed in the test data).

R= We agree with your observation, we have made that precision within the revised version of our manuscript.

- The model should be evaluated in a stratified way using the known risk groups, as this is highly relevant for practical application.

R= We performed relevant sensitivity analyses stratified by age, gender and number of comorbidities to evaluate the impact of each on the model; we have specified all conducted sensitivity analyses in Table 5 of the revised version of our manuscript.

- The authors need to provide a URL for the data set.

R= This data belongs to an internal registry for the hospital and cannot be uploaded or shared publicly. Data can be provided upon reasonable request to the corresponding author; we have uploaded R code for all performed analyses into a GitHub repository which can be verified at https://github.com/oyaxbell/nutri_cov.

Attachment

Submitted filename: Reviewer comments - PLoS One.docx

Decision Letter 1

Itamar Ashkenazi

3 Dec 2020

Validation and repurposing of the MSL-COVID-19 score for prediction of severe COVID-19 using simple clinical predictors in a triage setting: The Nutri-CoV score

PONE-D-20-25115R1

Dear Dr. Aguilar-Salinas,

We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements.

Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication.

An invoice for payment will follow shortly after the formal acceptance. To ensure an efficient process, please log into Editorial Manager at http://www.editorialmanager.com/pone/, click the 'Update My Information' link at the top of the page, and double check that your user information is up-to-date. If you have any billing related questions, please contact our Author Billing department directly at authorbilling@plos.org.

If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

Kind regards,

Itamar Ashkenazi

Academic Editor

PLOS ONE

Additional Editor Comments (optional):

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation.

Reviewer #1: All comments have been addressed

**********

2. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

**********

3. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

**********

4. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

**********

5. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

**********

6. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: Thank you for addressing the statistical concerns. I would just like to reiterate a concept that I including in my review of the paper. Which is that I suspect that even if this model were used at a similar institution, the efficacy observed will likely not agree with the efficacy reported in the paper.

**********

7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Acceptance letter

Itamar Ashkenazi

7 Dec 2020

PONE-D-20-25115R1

Validation and repurposing of the MSL-COVID-19 score for prediction of severe COVID-19 using simple clinical predictors in a triage setting: The Nutri-CoV score

Dear Dr. Aguilar-Salinas:

I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now with our production department.

If your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information please contact onepress@plos.org.

If we can help with anything else, please email us at plosone@plos.org.

Thank you for submitting your work to PLOS ONE and supporting open access.

Kind regards,

PLOS ONE Editorial Office Staff

on behalf of

Dr. Itamar Ashkenazi

Academic Editor

PLOS ONE

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    S1 File

    (DOCX)

    Attachment

    Submitted filename: Reviewer comments - PLoS One.docx

    Data Availability Statement

    Code for verification of all analytical procedures is available at https://github.com/oyaxbell/nutri_cov. Data cannot be made publicly available due to ethical restrictions, however they are available on request from the Institutional Data Access Ethics Committee with contact via the corresponding author or the president of the ethics committee at arturo.galindof@incmnsz.mx, for researchers who meet the criteria for access to confidential data. For more information please consult: https://www.incmnsz.mx/opencms/contenido/investigacion/comiteEtica/miembros_comite.html.


    Articles from PLoS ONE are provided here courtesy of PLOS

    RESOURCES