Skip to main content
PLOS Digital Health logoLink to PLOS Digital Health
. 2022 Jan 18;1(1):e0000005. doi: 10.1371/journal.pdig.0000005

Applied machine learning for the risk-stratification and clinical decision support of hospitalised patients with dengue in Vietnam

Damien K Ming 1,*, Bernard Hernandez 2,3, Sorawat Sangkaew 2, Nguyen Lam Vuong 4,5, Phung Khanh Lam 4,5, Nguyen Minh Nguyet 4, Dong Thi Hoai Tam 4, Dinh The Trung 4, Nguyen Thi Hanh Tien 4, Nguyen Minh Tuan 4,6, Nguyen Van Vinh Chau 4,7, Cao Thi Tam 7, Ho Quang Chanh 4,7, Huynh Trung Trieu 4,7, Cameron P Simmons 8, Bridget Wills 4,9, Pantelis Georgiou 2,3, Alison H Holmes 2, Sophie Yacoub 4,9; on behalf of the Vietnam ICU Translational Applications Laboratory (VITAL) investigators
Editor: Ryan S McGinnis10
PMCID: PMC9931311  PMID: 36812518

Abstract

Background

Identifying patients at risk of dengue shock syndrome (DSS) is vital for effective healthcare delivery. This can be challenging in endemic settings because of high caseloads and limited resources. Machine learning models trained using clinical data could support decision-making in this context.

Methods

We developed supervised machine learning prediction models using pooled data from adult and paediatric patients hospitalised with dengue. Individuals from 5 prospective clinical studies in Ho Chi Minh City, Vietnam conducted between 12th April 2001 and 30th January 2018 were included. The outcome was onset of dengue shock syndrome during hospitalisation. Data underwent random stratified splitting at 80:20 ratio with the former used only for model development. Ten-fold cross-validation was used for hyperparameter optimisation and confidence intervals derived from percentile bootstrapping. Optimised models were evaluated against the hold-out set.

Findings

The final dataset included 4,131 patients (477 adults and 3,654 children). DSS was experienced by 222 (5.4%) of individuals. Predictors were age, sex, weight, day of illness at hospitalisation, indices of haematocrit and platelets over first 48 hours of admission and before the onset of DSS. An artificial neural network model (ANN) model had best performance with an area under receiver operator curve (AUROC) of 0.83 (95% confidence interval [CI], 0.76–0.85) in predicting DSS. When evaluated against the independent hold-out set this calibrated model exhibited an AUROC of 0.82, specificity of 0.84, sensitivity of 0.66, positive predictive value of 0.18 and negative predictive value of 0.98.

Interpretation

The study demonstrates additional insights can be obtained from basic healthcare data, when applied through a machine learning framework. The high negative predictive value could support interventions such as early discharge or ambulatory patient management in this population. Work is underway to incorporate these findings into an electronic clinical decision support system to guide individual patient management.

Introduction

Dengue is a systemic viral disease which exerts a significant health and economic burden worldwide. Up to 5% of hospitalised patients develop severe dengue, a life-threatening complication manifesting as shock, bleeding and/or organ dysfunction [1]. With an estimated 51 million symptomatic cases each year, seasonal epidemics and high caseloads impose a huge strain on local healthcare services [2]. The wide spectrum, and non-specific nature of clinical presentations pose further challenges to effective healthcare planning [3].

Strategies to identify patients who are at increased risk of complications such as dengue shock syndrome (DSS) during the early febrile phase of illness have been a subject of considerable research [4,5]. A widely-adopted approach particularly in low- and middle-income countries (LMICs) is the use of clinical warning signs outlined in the World Health Organisation (WHO) 2009 dengue guidelines [6]. The absence of these signs provides a high negative predictive value for severe dengue [7] and furthermore has relatively few requirements for implementation–needing only clinical examination findings and results from basic haematological tests. However, heterogeneity between care settings [8] and a suboptimal specificity have led to concerns that use of this system might not alleviate the excess, and potentially unnecessary hospitalisations when this system is adopted for admission triage [9]. In response, clinical prognostic models [1012] and dengue-specific biomarkers [13,14] have been researched, but their limited performances or requirement for specialised laboratory testing represent barriers to widespread adoption. Tools for risk-stratification in dengue should: maximise the utility of existing patient data, be calibrated to local healthcare setting, and fit into the clinical decision-making workflow.

The rapid scale up of digital health developments, particularly within LMICs have enabled a data-driven approach previously not possible. In particular integration of electronic health record systems can strengthen healthcare systems [15] and provide the platforms to capitalise on novel data science methodologies such as machine learning. Machine learning is particularly suited for analysis of large scale, multidimensional data [16], as well as diverse data types such as physiological signals and radiological images utilised within healthcare. They provide empirical approaches to utilising data with reduced reliance on a priori assumptions. Within dengue research, machine learning methods have been applied in outbreak prediction and disease forecasting [17] but their role in clinical decision-making and risk stratification remains limited [1820].

As part of a multi-disciplinary group for development of technological interventions in critical illness within LMIC settings, the Vietnam ICU Translational Applications Laboratory (VITAL), we adopted a data-driven approach to utilise supervised machine learning in a cohort of patients hospitalised with acute dengue to predict the development of DSS. We hypothesise that adopting such approaches on large, diverse datasets would offer a robust approach to apply findings from research data to real-world settings. The aim of this work is the development of a clinical decision-support system (CDSS) to serve as an adjunct in the clinical management of patients hospitalised with dengue.

Methods

Results are reported following the TRIPOD Statement [21].

Ethical approval

The study was approved by the scientific and ethical committee of the Hospital for Tropical Diseases (HTD), Ho Chi Minh City and by the Oxford Tropical Research Ethics Committee (OxTREC) with datasets pseudonymised prior to analyses (references 145–0420 and 146–0420).

Source of data

The data used for this study comprises of an aggregation of prospective clinical studies conducted by Oxford University Clinical Research Unit (OUCRU) between 12th April 2001 and 30th January 2018 in healthcare facilities including the HTD within Ho Chi Minh City, Vietnam [10,12,2224]. These included 4 prospective observational studies and 1 randomised control trial (ISRCTN39575233). Broadly, patients were eligible for inclusion if they presented with an acute febrile illness compatible with dengue on clinical assessment. Subsequent confirmation of dengue was done through one, or more of the following: i) a positive NS1 point of care assay or NS1 ELISA, ii) positive reverse transcriptase polymerase chain reaction (RT-PCR), iii) positive dengue IgM through acute serology, iv) or seroconversion of paired IgM samples. These criteria for dengue diagnosis are in line with standard study definitions employed in this setting. Further information on individual studies including recruitment criteria can be found in the S1 Appendix.

For patients enrolled, the decision for hospital admission was made according to the study inclusion criteria, and/or clinician assessment in line with national guidelines. Individuals managed in hospital include patients at higher risk of clinical deterioration but also those unsuitable for ambulatory management because of clinical and non-clinical factors. We defined adults as patients who were 18 years or above at time of enrolment.

Information common across studies including demographics, presenting features, investigations results and outcomes were extracted. In order to develop the prediction model, we excluded any observations obtained at the same time, or after the onset of DSS. For patients who did not develop DSS, observations obtained after 120 hours of illness onset were also excluded, where the onset of fever is taken as the start of illness. This timepoint represented the typical period of the critical phase and onset of shock in dengue. We included only patients who were randomised to the placebo arm for any interventional studies. In the final models patient predictive variables from all studies were measured within the first 48 hours of hospital admission–corresponding to a median of 4 and 5 days of illness and provides a useful timeframe for management in the early phase of dengue in hospital.

A diagram of patient inclusion is shown in Fig 1. Patient records were excluded from the final dataset given the following criteria–patients: i) with a final diagnosis of an illness other than dengue (n = 6,329), ii) enrolled in the intervention arm of the randomised control trial (n = 150), iii) not hospitalised during illness (n = 1,226), iv) without data prior to day 5 of illness onset and where the development of DSS occurred at the same time, or before dates of predictor variables (n = 283). The final dataset consisted of 4,131 adult and paediatric patients.

Fig 1. CONSORT diagram for patients recruited in original clinical studies and processing to derive a development and hold out set for model evaluation.

Fig 1

Prediction outcomes

The primary outcome was development of dengue shock syndrome (DSS) during hospital admission period as a binary classifier. DSS is defined as a pulse pressure equal to or less than 20 mmHg, or low blood pressure (BP) for age, with clinical signs of reduced peripheral perfusion [6]. A secondary analysis was also carried out with complicated dengue as the outcome of interest. This was defined as the presence of shock, significant bleeding and plasma leakage at any point during hospital stay (S1 Appendix).

Predictors and missing data

Predictors used in the models include routine clinical and laboratory parameters, selected on the basis of expert consensus, data completeness and pragmatic utility suited to a LMIC healthcare setting. The final dataset for all models included predictors with less than 1.5% missing data. For each model, categorical data was fitted on the training set and imputed using the mode, and missing numerical data was imputed using both the mean and median and the method which resulted in better cross-validation performance was chosen for that particular model. Numerical features (age, day of illness, weight, haematocrit, platelet count) were either transformed by standardisation to achieve a mean value of 0 and standard deviation of 1, or left untransformed again depending on performance attained in cross-validation through a grid search process.

Development and validation

The data underwent an initial random split using an 80/20 ratio to create a development and hold-out set using a stratified process to ensure proportional distribution of outcomes in both sets: the proportion of DSS in the development and hold-out sets were 5.4% and 5.3% respectively. The hold-out set was not used in any stage of the model development process and used only for testing of the final models.

Model development process

Four diverse machine learning algorithms namely: extreme gradient boosting (XGBoost), random forest classifier, artificial neural network and support vector machines were used for model development. Logistic regression using lasso or ridge regression was also used as a baseline comparator.

Each algorithm was trained using the development set only through stratified k-fold cross-validation in order to establish optimal hyperparameters whilst minimising overfitting. The development set was split into 10 equal-sized folds and models with specific hyperparameter configurations trained on the 9 folds, and validated against the remaining fold. This process was then repeated 10 times with a different fold in each iteration. A grid-search process was used to generate different combinations of hyperparameters for each classification algorithm. The area under the receiver operating curve (AUROC) was used for scoring, and other reported metrics included sensitivity, specificity, positive predictive value (PPV), negative predictive value (NPV) and calibration using the Brier score. Confidence intervals for model performance metrics were estimated through bootstrap resampling of the development set and tested against the out of bag samples, repeated 1,000 times. We addressed the class imbalance in the dataset through Synthetic Minority Over-sampling Technique (SMOTE) within our pipelines. A probabilistic classifier method was used to compare predicted and actual outcomes. The optimal cut-off value was determined according to Youden’s J-statistic for each algorithm in order to provide best all-around performance. Models were calibrated through isotonic calibration through a 10-fold cross validation process.

Final models for each of the algorithms were then tested with the hold-out set, in order to describe performance against an independent, previously unseen dataset. A Shapley Additive exPlanations (SHAP) analysis was performed to analyse models for interpretability. This utilises an independent method to evaluate the impact of individual predictors within the model with regards to predicted outcomes: a greater SHAP value denotes a higher contribution towards a prediction of complicated dengue, taking into account interaction effects. References and the model development process are detailed in S1 Appendix. All analyses were performed in Python 3.7.

Results

Description

The final dataset included 4,131 patients hospitalised between 12th April 2001 and 30th January 2018 from 5 studies. Baseline patient characteristics are shown in Table 1. The number of female patients was 2,008 (49%) and the median age was 12 years old (interquartile range, IQR 9–14 years). The median day of illness on presentation to hospital was 3 days (IQR 2–4 days). In total, 222 (5.4%) patients experienced DSS during hospitalisation of which 209 (94.1%) were children and 13 (5.9%) were adults.

Table 1. Baseline characteristics of individual studies included in the final dataset.

Study Recruitment dates Patients enrolled (n) Patients with confirmed diagnosis of dengue (n) Patients included in analyses (n) Patients included in analyses
Age (median, IQR) Day of illness onset on hospitalisation (median, IQR) Female sex (n) Secondary infection (n) Shock (n) Outcome measurement
S1 12/4/2001–24/7/2009 3,042 3,042 2,615 12 (10–13) 4 (3–4) 1,079 (41%) 1,536/1,903 (81%) 170 (7%) WHO 1997 dengue outcomes
S2 3/8/2009–8/12/2010 225 225 73 15 (13–20) 1 (1–2) 18 (25%) 42/64 (66%) 3 (4%) WHO 2009 dengue outcomes
S3 8/12/2010–16/6/2011 88 88 47 23 (21–25) 3 (3–3) 28 (40%) 11/38 (29%) 8 (17%) WHO 2009 dengue outcomes
S4 19/10/2010–4/12/2014 8,100 2,245 940 12 (10–16) 2 (1–3) 427 (45%) 687/915 (75%) 31 (3%) WHO 2009 dengue outcomes
S5 20/10/2016–30/1/2018 664 542 456 30 (27–45) 4 (3–5) 456 (100%) 494/539 (92%) 10 (2%) WHO 2009 dengue outcomes

Model predictors

Final predictors included in the model are patient age, sex, weight, day of illness onset on hospital admission, and the minimum, median and maximum haematocrit and platelet count obtained over the first 48 hours of hospital admission. There were significant differences in distributions of sex, day of illness on hospital admission and indices of haematocrit and platelets values between patients with, or without DSS. Characteristics of the two groups and the comparisons of predictors are shown in Table 2.

Table 2. Characteristics of cohort (n = 4,131) divided by complication outcome during hospital. Haematological values and ranges refer to results taken over the initial 48 hours of hospitalisation only.

Data is presented as median and brackets denote the interquartile range. Univariate analyses were done using the Mann-Whitney test or Chi Squared test as appropriate.

Shock n = 222 No shock n = 3,909 p-value Missing data (%)
Median (IQR) Median (IQR)
Median age (years) 11 (9–13) 12 (9–14) 0.07 1.5
Median day of illness at hospital admission (days) 3 (2–4) 3 (2–4) 0.05 0
Median weight (kg) 34 (27–43) 36 (27–45) 0.26 0.9
Female sex (%) 90/222 (41%) 1,918/3,909 (49%) 0.02 0
Median haematocrit (%) 41 (39–43) 40 (37–42) <0.001 0.2
Maximum haematocrit (%) 48 (44–51) 42 (39–45) <0.001 0.2
Minimum haematocrit (%) 37 (35–39) 38 (35–40) 0.05 0.2
Maximum platelet count (x 109/L) 140 (102–192) 159 (118–207) <0.001 0.2
Median platelet count (x 109/L) 70 (45–99) 116 (83–159) <0.001 0.2
Minimum platelet count (x 109/L) 29 (18–58) 86 (51–135) <0.001 0.2

Model evaluation

Univariate analysis between the development (n = 3,304) and hold-out (n = 837) set showed no significant statistical differences across included predictors. In cross-validation of the development dataset, all machine learning algorithms used in this study resulted in similar mean AUROC (between 0.77 and 0.83). The performance of logistic regression using linear terms resulted in an AUROC of 0.79 (95% CI 0.74–0.83).

The artificial neural network (ANN) classifier model provided the best discrimination performance in predicting DSS, with an AUROC of 0.83 (95% CI, 0.76–0.85). When optimal cut-off thresholds were chosen using the J-statistic, this model provided a specificity of 0.88 (0.68–0.93), sensitivity of 0.66 (0.52–0.81) PPV of 0.24 (0.12–0.33), NPV of 0.98 (0.97–0.99) and a Brier score of 0.07. The performance results of all optimised final models from cross-validation using development data are displayed in Table 3.

Table 3. Performance of final models for each algorithm with respect to internal 10-fold cross validation on the development set (n = 3,304).

The 95% confidence intervals shown in brackets were derived from bootstrap resampling of the development and testing against out of bag samples, repeated 1,000 times.

Model Mean AUROC Specificity Sensitivity Positive Predictive value Negative Predictive value Brier score
XGBoost 0.81 (0.74–0.84) 0.90 (0.70–0.93) 0.59 (0.52–0.78) 0.26 (0.12–0.31) 0.97 (0.97–0.98) 0.068
Random forest 0.80 (0.75–0.84) 0.82 (0.75–0.93) 0.71 (0.54–0.76) 0.18 (0.14–0.31) 0.98 (0.97–0.98) 0.068
Logistic regression 0.79 (0.74–0.83) 0.89 (0.71–0.91) 0.62 (0.56–0.78) 0.24 (0.13–0.28) 0.98 (0.97–0.98) 0.075
Artificial neural network 0.83 (0.76–0.85) 0.88 (0.68–0.93) 0.66 (0.52–0.81) 0.24 (0.12–0.33) 0.98 (0.97–0.99) 0.068
Support vector machines* 0.82 (0.75–0.84) 0.86 (0.68–0.92) 0.66 (0.53–0.82) 0.21(0.11–0.29) 0.98 (0.97–0.99 0.13

*implementations of SVM were done with SMOTE.

Sensitivity analyses exploring performance of the model to classify complicated dengue as an endpoint, contribution of batch effects, WHO classifications and performance of the model stratified by age (paediatric or adult cohort) are shown in the S1 Appendix.

Model evaluation against hold-out test set

Final models were evaluated against the independent hold-out set of 827 patients not involved in development. The XGBoost model provided the highest discrimination performance (AUROC 0.85, specificity 0.91, sensitivity 0.64, PPV of 0.29 and NPV of 0.98), although discrimination across other machine learning classifiers were similar (AUROC ranging from 0.79 to 0.85). Logistic regression model performance in this setting provided an AUROC of 0.79 (Table 4).

Table 4. Performance metrics of final calibrated models when evaluated against the hold-out set (n = 827).

Model Mean AUROC Specificity Sensitivity Positive Predictive value Negative Predictive value Brier score
XGBoost 0.85 0.91 0.64 0.29 0.98 0.04
Random forest 0.84 0.87 0.68 0.23 0.98 0.04
Logistic regression 0.79 0.90 0.55 0.24 0.97 0.04
Artificial neural network (ANN) 0.82 0.84 0.66 0.18 0.98 0.04
Support vector machines 0.83 0.84 0.66 0.19 0.98 0.04

Model interpretability

To provide interpretability of the XGBoost model we performed a SHAP analysis. A summary plot (Fig 2) shows that within the model a higher maximum haematocrit, a lower minimum platelet count and female sex and lower age contributed towards complicated dengue prediction with a non-linear relationship between SHAP values and individual predictor variables (Fig 3). SHAP values for individual features for the other models are shown in the S1 Appendix. All models ranked maximum haematocrit and minimum platelet count highest in terms of the importance of features as characterised by SHAP.

Fig 2. Summary plot of SHAP values for the XGBoost model.

Fig 2

The plot shows the contribution of individual predictors and their range of values towards final model output prediction, where shock and no shock are represented by 1 and 0 on the x-axis respectively. The main predictors are arranged in descending importance for the model. The colours of the individual features represent whether the values are high or low. For example, a higher maximum haematocrit (red) is associated with a positive impact on model output and thus associated with dengue shock. Female sex is represented by blue (0) and male sex represented by red (1).

Fig 3.

Fig 3

Scatter plot of predictor values (x-axis) against SHAP values (y-axis) in the XGBoost model for haematocrit, platelet count and age. The plot shows a non-linear relationship between predictor values and model output.

Discussion

We trained and evaluated machine learning models to predict onset of dengue shock syndrome in hospitalised patients in Vietnam, using only basic demographic and laboratory predictors measured during the first 48 hours of hospital admission. These models were able to robustly model the non-linear relationships between predictors and outcome, with the ANN algorithm demonstrating optimal discrimination in terms of AUROC in cross-validation and when evaluated against the independent hold-out set.

Our models take into account changes in haematocrit and platelet over the first 48 hours of admission as predictors and classifies subsequent risk of DSS. The negative predictive value of 0.98 provided by the ANN model may be of clinical utility by allowing an automatic identification of admitted patients who are at lower risk of developing complications. The performance metrics of our machine learning model is commensurate with those of other prognostic models [18] although direct comparisons can be difficult given differences in practice and care pathways.

Patients with dengue in our setting are admitted to hospital for a variety of indications: when clinicians consider them at higher risk, have certain comorbidities, or in presence of non-medical factors such as logistical challenges for outpatient follow up. We plan to implement these findings through a clinical decision support system (CDSS) interface developed by our group, and previously validated in real-world clinical settings [25]. This web-based system allows the model to be run on most computers or tablets without any specialised requirements. Usability and uptake are important for implementation within clinical settings. Studies characterising workflow, end-user requirements and optimal interface designs are currently taking place at our hospital to optimise local roll-out and utility in the first instance. The feasibility of applying these models to other regions with endemic dengue has been shown to be possible [26], but factors such as differences in local healthcare processes, treatment guidelines and access to healthcare have to be taken into account.

The inclusion of only basic predictors for our model, namely age, sex, weight, haematocrit and platelet concentrations were prioritised. These measurements are readily-accessible in many LMIC care settings and are known significant factors in dengue risk stratification [12,27]. Although overall performance of our optimal model may not supersede that of clinical assessment, its implementation within a CDSS environment in hospitals could provide an additional layer of patient safety and consistency in care. It could also support interventions such as early hospital discharge and/or evaluate the suitability for ambulatory management within this group. The role of such a CDSS is not to take away from clinicians’ assessment but one which supports and complements their workflow [28]. Ensuring that the model is updated with available new data will be essential such as use of an iterative system of training and evaluation over time. A prospective validation study is currently underway at our institution to examine the real-world performance of these models. In their development we used the J-statistic to optimise all-round performance but specific metrics of the model, such as its negative predictive value, will need to maintained by threshold adjustment in order for this to be clinically relevant. This threshold is likely dynamic and will depend on factors including seasonality and hospital caseload.

This study provides a proof-of-principle and baseline that basic healthcare data can offer added-value when processed through novel data-science methods such as machine learning. These methods are particularly suited for data without a clear underlying linear relationship or that which are prone to noise or artifact, including continuous physiological data. For example, photoplethysmography signals used in pulse oximetry is predictive for dengue shock analysed through machine learning [29]. Capture of such data through low-cost (<$100 USD) devices or wearable technologies [30], and their incorporation into predictive algorithms is an area of active area interest.

The strengths of our study include the use of a prospectively-collected set of clinical data for machine learning model development and validation. This is the largest sample size used for this purpose to our knowledge and increases the prospective generalisability of our findings. Minimal missing data (<1.5%) reduced requirements for imputation also increases validity of the results. For model development, the use of cross-validation and an independent hold-out set are robust methods of reducing overfitting, impact of noise and issues regarding heterogeneity within the data.

There were limitations to our study. We used data aggregated from five different clinical studies conducted at OUCRU over 19 years to build a large training cohort. Inherent differences in study design between datasets, such as the threshold for hospital admission or pregnancy status exist, leading to underlying heterogeneity. Although the overall inclusion criteria for patients into the analysis were comparable i.e., hospitalised patients with dengue during the febrile phase–underlying biases might exist which affect future performance. There are intrinsic differences between each clinical study as they address specific research questions and span many years not amenable to standard normalisation methods (S1 Appendix). We have therefore not explicitly corrected for these differences prior to dataset aggregation, in part to also retain variance needed for prospective generalisability and real-world performance. Whether this approach is optimal will be ultimately dependent on performance of the model on an independent, prospective cohort of patients.

We adopted DSS as the primary outcome as this was captured most accurately within the datasets. This is a relatively rare outcome, leading to an imbalanced dataset and affects model calibration. It is also acknowledged that other clinical features of moderate and severe dengue (such as bleeding and organ impairment) have not been considered in the primary model—we therefore developed a secondary model examining the association of complicated dengue with feature variables (S1 Appendix) and show similar performances suitable for implementation. Dengue shock and severe disease can also present earlier in illness prior to hospitalisation and therefore this model might have limited utility for this subset of patients.

Only a limited number of features collected as part of routine clinical care were included in our final models and as a result feature selection/ engineering was not possible. It is possible that relevant clinical features such as dengue serotype or indices of viral load could result in better model performance although this has to be balanced with feasibility of model implementation across LMIC clinical settings. Finally, this was a predominantly paediatric patient training cohort, and model performance was less robust when tested against older patients. Future iterations of the model would need to take into account optimising performance to ensure fairness across different age groups.

In conclusion, we present results from a machine learning approach to predict the risk of dengue shock in hospitalised patients with dengue. We demonstrate performance metrics suitable for clinical evaluation and plan prospective studies to understand if these methods, when delivered through a decision support tool, can translate into improvements in clinical care.

Supporting information

S1 Appendix. Description of data source and individual studies, description of data source and individual studies and secondary and sensitivity analyses.

(DOCX)

Acknowledgments

The authors would like to acknowledge the patients who participated, the doctors and nurses who cared for the patients, and the laboratory staff at the Hospital for Tropical Diseases.

Data Availability

The dataset used for analysis is available at the Oxford University Research Archive: https://doi.org/10.5287/bodleian:gAzqvApA4.

Funding Statement

This work was supported by the Wellcome Trust grant [215010/Z/18/Z]. Authors DM and BH receive their salaries from and are supported by the grant. The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.

References

  • 1.Wilder-Smith A, Ooi E-E, Horstick O, Wills B. Dengue. The Lancet. 2019;393: 350–363. doi: 10.1016/S0140-6736(18)32560-1 [DOI] [PubMed] [Google Scholar]
  • 2.Cattarino L, Rodriguez-Barraquer I, Imai N, Cummings DAT, Ferguson NM. Mapping global variation in dengue transmission intensity. Science Translational Medicine. 2020;12. doi: 10.1126/scitranslmed.aax4144 [DOI] [PubMed] [Google Scholar]
  • 3.Muller DA, Depelsenaire ACI, Young PR. Clinical and Laboratory Diagnosis of Dengue Virus Infection. The Journal of Infectious Diseases. 2017;215: S89–S95. doi: 10.1093/infdis/jiw649 [DOI] [PubMed] [Google Scholar]
  • 4.Jaenisch T, Tam DTH, Kieu NTT, Van Ngoc T, Nam NT, Van Kinh N, et al. Clinical evaluation of dengue and identification of risk factors for severe disease: protocol for a multicentre study in 8 countries. BMC infectious diseases. 2016;16: 120. doi: 10.1186/s12879-016-1440-3 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5.Yacoub S, Wills B. Predicting outcome from dengue. doi: 10.1186/s12916-014-0147-9 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6.Special Programme for Research and Training in Tropical Diseases, World Health Organization, editors. Dengue: guidelines for diagnosis, treatment, prevention, and control. New ed. Geneva: TDR: World Health Organization; 2009. [Google Scholar]
  • 7.Barniol J, Gaczkowski R, Barbato EV, da Cunha RV, Salgado D, Martínez E, et al. Usefulness and applicability of the revised dengue case classification by disease: multi-centre study in 18 countries. BMC Infectious Diseases. 2011;11: 106. doi: 10.1186/1471-2334-11-106 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.Morra ME, Altibi AMA, Iqtadar S, Minh LHN, Elawady SS, Hallab A, et al. Definitions for warning signs and signs of severe dengue according to the WHO 2009 classification: Systematic review of literature. Reviews in Medical Virology. 2018;28: e1979. doi: 10.1002/rmv.1979 [DOI] [PubMed] [Google Scholar]
  • 9.Srikiatkhachorn A, Rothman AL, Gibbons RV, Sittisombut N, Malasit P, Ennis FA, et al. Dengue—How Best to Classify It. Clinical Infectious Diseases. 2011;53: 563–567. doi: 10.1093/cid/cir451 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10.Nguyen MT, Ho TN, Nguyen VVC, Nguyen TH, Ha MT, Ta VT, et al. An Evidence-Based Algorithm for Early Prognosis of Severe Dengue in the Outpatient Setting. Clin Infect Dis. 2017;64: 656–663. doi: 10.1093/cid/ciw863 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11.Lee I-K, Liu J-W, Chen Y-H, Chen Y-C, Tsai C-Y, Huang S-Y, et al. Development of a Simple Clinical Risk Score for Early Prediction of Severe Dengue in Adult Patients. Huy NT, editor. PLoS ONE. 2016;11: e0154772. doi: 10.1371/journal.pone.0154772 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12.Lam PK, Ngoc TV, Thu Thuy TT, Hong Van NT, Nhu Thuy TT, Hoai Tam DT, et al. The value of daily platelet counts for predicting dengue shock syndrome: Results from a prospective observational study of 2301 Vietnamese children with dengue. PLoS Negl Trop Dis. 2017;11. doi: 10.1371/journal.pntd.0005498 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.Yacoub S, Lam PK, Vu LHM, Le TL, Ha NT, Toan TT, et al. Association of Microvascular Function and Endothelial Biomarkers With Clinical Outcome in Dengue: An Observational Study. J Infect Dis. 2016;214: 697–706. doi: 10.1093/infdis/jiw220 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14.Vuong NL, Le Duyen HT, Lam PK, Tam DTH, Vinh Chau NV, Van Kinh N, et al. C-reactive protein as a potential biomarker for disease progression in dengue: a multi-country observational study. BMC Medicine. 2020;18: 35. doi: 10.1186/s12916-020-1496-1 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15.Silvestre Eva. How Electronic Health Records Strengthen the Health Systems of Low- and Middle-Income Countries: Learning from Eswatini and Mexico. Chapel Hill, NC: MEASURE Evaluation, University of North Carolina; 2018. Available: https://www.measureevaluation.org/resources/publications/tr-18-285 [Google Scholar]
  • 16.Panch T, Szolovits P, Atun R. Artificial intelligence, machine learning and health systems. J Glob Health. 8. doi: 10.7189/jogh.08.020303 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17.Shi Y, Liu X, Kok S-Y, Rajarethinam J, Liang S, Yap G, et al. Three-Month Real-Time Dengue Forecast Models: An Early Warning System for Outbreak Alerts and Policy Decision Support in Singapore. Environ Health Perspect. 2016;124: 1369–1375. doi: 10.1289/ehp.1509981 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18.Huang S-W, Tsai H-P, Hung S-J, Ko W-C, Wang J-R. Assessing the risk of dengue severity using demographic information and laboratory test results with machine learning. PLOS Neglected Tropical Diseases. 2020;14: e0008960. doi: 10.1371/journal.pntd.0008960 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19.Ho T-S, Weng T-C, Wang J-D, Han H-C, Cheng H-C, Yang C-C, et al. Comparing machine learning with case-control models to identify confirmed dengue cases. PLOS Neglected Tropical Diseases. 2020;14: e0008843. doi: 10.1371/journal.pntd.0008843 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 20.Tan KW, Tan B, Thein TL, Leo Y-S, Lye DC, Dickens BL, et al. Dynamic dengue haemorrhagic fever calculators as clinical decision support tools in adult dengue. Trans R Soc Trop Med Hyg. 2020;114: 7–15. doi: 10.1093/trstmh/trz099 [DOI] [PubMed] [Google Scholar]
  • 21.Moons KGM, Altman DG, Reitsma JB, Ioannidis JPA, Macaskill P, Steyerberg EW, et al. Transparent Reporting of a multivariable prediction model for Individual Prognosis or Diagnosis (TRIPOD): explanation and elaboration. Ann Intern Med. 2015;162: W1–73. doi: 10.7326/M14-0698 [DOI] [PubMed] [Google Scholar]
  • 22.Tien NTH, Lam PK, Duyen HTL, Ngoc TV, Ha PTT, Kieu NTT, et al. Assessment of Microalbuminuria for Early Diagnosis and Risk Prediction in Dengue Infections. PLOS ONE. 2013;8: e54538. doi: 10.1371/journal.pone.0054538 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 23.Tam DTH, Ngoc TV, Tien NTH, Kieu NTT, Thuy TTT, Thanh LTC, et al. Effects of Short-Course Oral Corticosteroid Therapy in Early Dengue Infection in Vietnamese Patients: A Randomized, Placebo-Controlled Trial. Clin Infect Dis. 2012;55: 1216–1224. doi: 10.1093/cid/cis655 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 24.Tuan NM, Nhan HT, Chau NVV, Hung NT, Tuan HM, Tram TV, et al. Sensitivity and Specificity of a Novel Classifier for the Early Diagnosis of Dengue. PLoS Negl Trop Dis. 2015;9. doi: 10.1371/journal.pntd.0003638 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25.Rawson TM, Hernandez B, Moore LSP, Herrero P, Charani E, Ming D, et al. A Real-world Evaluation of a Case-based Reasoning Algorithm to Support Antimicrobial Prescribing Decisions in Acute Care. Clin Infect Dis. [cited 22 Sep 2020]. doi: 10.1093/cid/ciaa383 [DOI] [PubMed] [Google Scholar]
  • 26.Sangkaew S. Enhancing risk prediction of progression to severe disease during the febrile phase of dengue: A systematic review and meta-analysis. 238. 2020. [cited 22 Oct 2021]. doi: 10.1016/j.ijid.2020.11.056 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27.Sangkaew S, Ming D, Boonyasiri A, Honeyford K, Kalayanarooj S, Yacoub S, et al. Risk predictors of progression to severe disease during the febrile phase of dengue: a systematic review and meta-analysis. The Lancet Infectious Diseases. 2021;0. doi: 10.1016/S1473-3099(20)30601-0 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 28.Byrne C, Sherry D, Mercincavage L, Johnston D, Pan E, Schiff G. Advancing Clinical Decision Support—Key Lessons In Clinical Decision Support Implementation. Office of the National Coordinator for Health IT, Office of Policy and Planning. 2012; 101. [Google Scholar]
  • 29.Moulton SL, Mulligan J, Srikiatkhachorn A, Kalayanarooj S, Grudic GZ, Green S, et al. State-of-the-art monitoring in treatment of dengue shock syndrome: a case series. J Med Case Rep. 2016;10: 233. doi: 10.1186/s13256-016-1019-z [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 30.Ming DK, Sangkaew S, Chanh HQ, Nhat PTH, Yacoub S, Georgiou P, et al. Continuous physiological monitoring using wearable technology to inform individual management of infectious diseases, public health and outbreak responses. Int J Infect Dis. 2020;96: 648–654. doi: 10.1016/j.ijid.2020.05.086 [DOI] [PMC free article] [PubMed] [Google Scholar]
PLOS Digit Health. doi: 10.1371/journal.pdig.0000005.r001

Decision Letter 0

Martin G Frasch, Ryan S McGinnis

7 Sep 2021

PDIG-D-21-00002

Applied machine learning for the risk-stratification and clinical decision support of hospitalised patients with dengue in Vietnam

PLOS Digital Health

Dear Dr. Ming,

Thank you for submitting your manuscript to PLOS Digital Health. After careful consideration, we feel that it has merit but does not fully meet PLOS Digital Health’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

==============================

EDITOR: Please insert comments here and delete this placeholder text when finished. Be sure to:

  • Indicate which changes you require for acceptance versus which changes you recommend

  • Address any conflicts between the reviews so that it's clear which advice the authors should follow

  • Provide specific feedback from your evaluation of the manuscript

Please ensure that your decision is justified on PLOS Digital Health’s publication criteria and not, for example, on novelty or perceived impact.

==============================

Please submit your revised manuscript by Nov 06 2021 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at digitalhealth@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pdig/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

We look forward to receiving your revised manuscript.

Kind regards,

Ryan S McGinnis, Ph.D.

Academic Editor

PLOS Digital Health

Journal Requirements:

1. Please provide separate figure files in .tif or .eps format only, and remove any figures embedded in your manuscript file.  If you are using LaTeX, you do not need to remove embedded figures.

For more information about figure files please see our guidelines: https://journals.plos.org/digitalhealth/s/figures

2. Please amend your detailed Financial Disclosure statement. This is published with the article, therefore should be completed in full sentences and contain the exact wording you wish to be published.

i). Please include all sources of funding (financial or material support) for your study. List the grants (with grant number) or organizations (with url) that supported your study, including funding received from your institution. 

ii). State the initials, alongside each funding source, of each author to receive each grant.

iii). State what role the funders took in the study. If the funders had no role in your study, please state: “The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.”

iv). If any authors received a salary from any of your funders, please state which authors and which funders.

3. Please update the completed 'Competing Interests' statement, including any COIs declared by your co-authors. If you have no competing interests to declare, please state "The authors have declared that no competing interests exist". Otherwise please declare all competing interests beginning with the statement "I have read the journal's policy and the authors of this manuscript have the following competing interests:

4. In the online submission form, you indicated that "The original clinical datasets used contains identifying or sensitive patient information relating to past and ongoing clinical studies and access is subject to research ethics committee approval from HTD ethical committee and OxTREC approval. Please contact the authors for further information and data requests.". All PLOS journals now require all data underlying the findings described in their manuscript to be freely available to other researchers, either 1. In a public repository, 2. Within the manuscript itself, or 3. Uploaded as supplementary information.

This policy applies to all data except where public deposition would breach compliance with the protocol approved by your research ethics board. If your data cannot be made publicly available for ethical or legal reasons (e.g., public availability would compromise patient privacy), please explain your reasons by return email and your exemption request will be escalated to the editor for approval. Your exemption request will be handled independently and will not hold up the peer review process, but will need to be resolved should your manuscript be accepted for publication. One of the Editorial team will then be in touch if there are any issues.

Additional Editor Comments (if provided):

The three reviewers have seen merit in this work, but also identify several areas that could be improved to strengthen the manuscript. In particular, and in keeping with the aims of the journal, the authors are encouraged to make their code and data available.

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Does this manuscript meet PLOS Digital Health’s publication criteria? Is the manuscript technically sound, and do the data support the conclusions? The manuscript must describe methodologically and ethically rigorous research with conclusions that are appropriately drawn based on the data presented.

Reviewer #1: Partly

Reviewer #2: Yes

Reviewer #3: Yes

**********

2. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

Reviewer #2: Yes

Reviewer #3: Yes

**********

3. Have the authors made all data underlying the findings in their manuscript fully available (please refer to the Data Availability Statement at the start of the manuscript PDF file)?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception. The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: No

Reviewer #2: No

Reviewer #3: No

**********

4. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS Digital Health does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

Reviewer #3: Yes

**********

5. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: The paper presents the use of machine learning algorithms to predict dengue shock syndrome based on an integrated dataset originating from 5 different studies. Overall, the research methodology and the methods are well presented. The issues with the work are as follows: (i) The selection of the predictors is opportunistic and hence it is unclear whether there are other clinical features that should have been included. The list of predictors is quite simplistic from a clinical perspective, though the positive is that these predictors still are given high prediction accuracy; (ii) Given the limited number of predictors, there is no opportunity to perform feature selection which is a limitation; (iii) Correlation between the features and the outcome should be checked, and confounders should be established; (iv) The prediction models perform binary classification, where there is a significant class imbalance such that only 5.4% of patients had a positive dengue shock. The study does not use any methods to address the class imbalance, and it is interesting to note such a high sensitivity reported in the results despite the class imbalance. Furthermore, which such a low number of the dengue class the authors should point out the ratio of the two classes in both the training and evaluation sets. the study can be improved by performing data augmentation using GAN or simply using SMOTE and then compare results between the original and the augmented dataset; (v) Table 1: is there an explanation for why the dengue shock numbers are so high in cohort S1 despite cohort S4 has a higher number of patients. It seems that there may be some issues with the inclusion criteria that need proper investigation; (vi) The aggregation of data from multiple temporal and institutional cohorts should take into account the clinical interpretation of dengue shock across these institutions to have a standardized inclusion criteria. It would be useful to develop independent models for each cohort to investigate the influence of the different predictors across cohorts; (vii) The data spans 19 years which is a rather long period given improvements in diagnostic methods and treatments--as the authors suggest that this study points to the utility of the model in LMIC given the need for a limited number of clinical attributes, it will be useful to stratify the data into temporal periods and then develop prediction models for each period to confirm that the predictors hold there importance across time; (viii) The data does not distinguish between paediatric patients and adults/elderly patients despite both groups have different clinical characteristics, especially elderly having co-morbidities. Therefore, having a universal model for all age groups is susceptible to bias towards a specific age group. (ix) The authors should provide details of how the trained models were calibrated; (x) The discussion should include insights into the clinical pragmatics of this work, the actual design and implementation of the digital health based solution in a clinical workflow, alluding to data access interfaces, etc. Also, the clinical setting in which it will be used and how will its use be determined on a per patient basis; (xi) The authors should comment on the scalability of this model to other regions; (xii) What plans are in place to keep the model current given the availability of new data and changes in clinical processes;

The paper does not present any novel work with regards to methods--it reports the application of existing machine learning methods to a curated dataset. So the original contribution of the paper basically is the preparation of the dataset which is insufficient. The authors should highlight any novelty of their work in terms of methods and outcomes.

Reviewer #2: The authors have ~4000 patients from 5 different data sources, and are trying to make different models (XGBoost, Logistic Regression, ANN, Random forest and SVM) to classify the patients according to if they have DSS (Dengue shock syndrome) they also do a secondary analysis making models for complications in DSS they have also done SHAP interpretability analysis for xgboost and logistic regression. Overall, the paper was well written we have some points that we found could be beneficial to improve the quality of the work

Major comments:

Authors should mention the effect of different data sources, i.e. data batches. Specifically, if the sampling strategy employed addresses potential biases due to discontinous data collection in batches over the years. technologies/approaches must have changed over the long period of collection. It is important to account for these batch effects as they may dramatically change the outcome.

As a suggestion, to reduce the bias in the model caused by data coming from 5 sources, and subsequently increase its generalization performance, the authors could have normalized the data from each source separately and then concatenated these 5 normalized datasets into one, instead of first concatenating and then normalizing. The authors haven’t mentioned when they performed the normalizing step but since they

mentioned it at the end in line 177, it is reasonable to assume that it was the last step. Having clarity on how this normalization was performed and at which step during the process is important.

Also, the entire analysis and the deidentified data should be hosted on a public platform such as github to allow researchers to reproduce the work.

Minor comments:

1) In Abstract Findings : The authors should give the distribution of patients who experienced DSS (how many of them were adults and how many were children)

2) Line 146: The authors should mention what symptoms they have taken to be ‘onset of illness’ in patients who did not develop DSS.

3) Line 177: Predictors used in the model which were transformed should be mentioned.

4) The authors could specify the Brier scores for evaluation against the hold out set for the primary analysis too, as they have done for the secondary analysis (Supplementary Appendix 3)

5) Supplementary Appendix 2: In the hyperparameters of the XGBoost model, the authors mention 2 optimal models. The authors should specify if these have equal AUROC, or if one was chosen because of less overfitting and other for better performance, or whatever else the case might be. Also, the ‘min_child_weight’ in both of these is different from the set of values they have taken initially (0.005 and 0.001 are not present in [0.05, 0.1, 0.2])

6) Supplementary Appendix 2: As a side note, the authors have taken very few parameters into account for the XGBoost model, and it could be possible to achieve a better AUROC using hyperparameters like ‘subsample’, ‘colsample_by...’ and ‘alpha’ and using a more extensive set of values for parameters like ‘max_depth’ and especially ‘min_child_weight’, for which the values used are quite small in magnitude.

7) Supplementary Appendix 2 (Interpretability and SHAP analysis): SHAP is only applied for XGBoost and logistic regression. The authors can use SHAP’s KernelExplainer to explain any model. However, making the MLP(ANN) using PyTorch/Tensorflow/keras and then using SHAP’s DeepExplainer or GradientExplainer classes will lead to better explanations and interpretability.

Reviewer #3: It was a delight to review this paper, which in its present form I could readily imagine appearing in the journal. I only have a few comments for the authors to address.

L174: Provide more information on how imputation was done. I can’t work out from this description how the mean or median was selected.

L176: Please provide a better explanation of which data were standardised and how the decision was made to standardise. Don’t some of the algorithms (such as lasso) automatically standardise covariates?

L202: I was surprised to see the optimal cut off being determined through a statistical algorithm rather than through the clinical requirements. I was expecting the decision to be based on, for instance, a maximum tolerable number of patients with DSS being sent home (presumably to die) or the hospital capacity, rather than Youden’s J statistic.

L237: Was stochastic domination assured for these p-values?

L260: What is linear logistic regression?

L269 (figure 2): This is very pretty but probably requires a bit more information to guide readers’ interpretation. For instance, with sex, how do we interpret the different shades of blue and red and white? Also I wonder whether the plot should be taller so that the densities are easier to perceive.

L316: Please check and confirm that the Tan Tock Seng Hospital (Singapore) dengue clinical cohort is not larger. I had the impression it was similar in size, and they’ve used it for similar applications, though obviously their patients are quite different.

Data availability: It would be preferable if a sanitised version of the dataset could be provided, with just the variables that were included in the final analysis so no need for information that could identify the patients.

**********

6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

Do you want your identity to be public for this peer review? If you choose “no”, your identity will remain anonymous but your review may still be made public.

For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

Reviewer #3: No

**********

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

PLOS Digit Health. doi: 10.1371/journal.pdig.0000005.r003

Decision Letter 1

Martin G Frasch, Ryan S McGinnis

15 Nov 2021

Applied machine learning for the risk-stratification and clinical decision support of hospitalised patients with dengue in Vietnam

PDIG-D-21-00002R1

Dear Dr. Ming,

We're pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements.

Within one week, you'll receive an e-mail detailing the required amendments. When these have been addressed, you'll receive a formal acceptance letter and your manuscript will be scheduled for publication.

An invoice for payment will follow shortly after the formal acceptance. To ensure an efficient process, please log into Editorial Manager at https://www.editorialmanager.com/pdig/ click the 'Update My Information' link at the top of the page, and double check that your user information is up-to-date. If you have any billing related questions, please contact our Author Billing department directly at authorbilling@plos.org.

If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they'll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact digitalhealth@plos.org.

Kind regards,

Ryan S McGinnis, Ph.D.

Academic Editor

PLOS Digital Health

Additional Editor Comments (optional):

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation.

Reviewer #3: All comments have been addressed

**********

2. Does this manuscript meet PLOS Digital Health’s publication criteria? Is the manuscript technically sound, and do the data support the conclusions? The manuscript must describe methodologically and ethically rigorous research with conclusions that are appropriately drawn based on the data presented.

Reviewer #3: Yes

**********

3. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #3: Yes

**********

4. Have the authors made all data underlying the findings in their manuscript fully available (please refer to the Data Availability Statement at the start of the manuscript PDF file)?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception. The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #3: Yes

**********

5. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS Digital Health does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #3: Yes

**********

6. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #3: Thanks for addressing my previous points. It would be good at the copy editing stage to give the doi of the data uploaded to ORA to make it easier to find.

**********

7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

Do you want your identity to be public for this peer review? If you choose “no”, your identity will remain anonymous but your review may still be made public.

For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #3: No

**********

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    S1 Appendix. Description of data source and individual studies, description of data source and individual studies and secondary and sensitivity analyses.

    (DOCX)

    Attachment

    Submitted filename: Response to Reviewers.docx

    Data Availability Statement

    The dataset used for analysis is available at the Oxford University Research Archive: https://doi.org/10.5287/bodleian:gAzqvApA4.


    Articles from PLOS Digital Health are provided here courtesy of PLOS

    RESOURCES