Skip to main content
PLOS Medicine logoLink to PLOS Medicine
. 2021 Mar 4;18(3):e1003542. doi: 10.1371/journal.pmed.1003542

Probabilistic seasonal dengue forecasting in Vietnam: A modelling study using superensembles

Felipe J Colón-González 1,2,3,4,*, Leonardo Soares Bastos 1,3,5, Barbara Hofmann 6, Alison Hopkin 6, Quillon Harpham 6, Tom Crocker 7, Rosanna Amato 7, Iacopo Ferrario 6, Francesca Moschini 6, Samuel James 6, Sajni Malde 6, Eleanor Ainscoe 6, Vu Sinh Nam 8, Dang Quang Tan 9, Nguyen Duc Khoa 9, Mark Harrison 7, Gina Tsarouchi 6, Darren Lumbroso 6, Oliver J Brady 1,2,#, Rachel Lowe 1,2,3,#
Editor: Alex Richard Cook10
PMCID: PMC7971894  PMID: 33661904

Abstract

Background

With enough advanced notice, dengue outbreaks can be mitigated. As a climate-sensitive disease, environmental conditions and past patterns of dengue can be used to make predictions about future outbreak risk. These predictions improve public health planning and decision-making to ultimately reduce the burden of disease. Past approaches to dengue forecasting have used seasonal climate forecasts, but the predictive ability of a system using different lead times in a year-round prediction system has been seldom explored. Moreover, the transition from theoretical to operational systems integrated with disease control activities is rare.

Methods and findings

We introduce an operational seasonal dengue forecasting system for Vietnam where Earth observations, seasonal climate forecasts, and lagged dengue cases are used to drive a superensemble of probabilistic dengue models to predict dengue risk up to 6 months ahead. Bayesian spatiotemporal models were fit to 19 years (2002–2020) of dengue data at the province level across Vietnam. A superensemble of these models then makes probabilistic predictions of dengue incidence at various future time points aligned with key Vietnamese decision and planning deadlines. We demonstrate that the superensemble generates more accurate predictions of dengue incidence than the individual models it incorporates across a suite of time horizons and transmission settings. Using historical data, the superensemble made slightly more accurate predictions (continuous rank probability score [CRPS] = 66.8, 95% CI 60.6–148.0) than a baseline model which forecasts the same incidence rate every month (CRPS = 79.4, 95% CI 78.5–80.5) at lead times of 1 to 3 months, albeit with larger uncertainty. The outbreak detection capability of the superensemble was considerably larger (69%) than that of the baseline model (54.5%). Predictions were most accurate in southern Vietnam, an area that experiences semi-regular seasonal dengue transmission. The system also demonstrated added value across multiple areas compared to previous practice of not using a forecast. We use the system to make a prospective prediction for dengue incidence in Vietnam for the period May to October 2020. Prospective predictions made with the superensemble were slightly more accurate (CRPS = 110, 95% CI 102–575) than those made with the baseline model (CRPS = 125, 95% CI 120–168) but had larger uncertainty. Finally, we propose a framework for the evaluation of probabilistic predictions. Despite the demonstrated value of our forecasting system, the approach is limited by the consistency of the dengue case data, as well as the lack of publicly available, continuous, and long-term data sets on mosquito control efforts and serotype-specific case data.

Conclusions

This study shows that by combining detailed Earth observation data, seasonal climate forecasts, and state-of-the-art models, dengue outbreaks can be predicted across a broad range of settings, with enough lead time to meaningfully inform dengue control. While our system omits some important variables not currently available at a subnational scale, the majority of past outbreaks could be predicted up to 3 months ahead. Over the next 2 years, the system will be prospectively evaluated and, if successful, potentially extended to other areas and other climate-sensitive disease systems.

Author summary

Why was this study done?

  • A climate-driven dengue early warning system would allow public health decision-makers to design, implement, and target timely interventions to the most at-risk places.

  • Dengue is sensitive to changes in temperature, rainfall, humidity, and wind speed. Therefore, a dengue prediction model driven by seasonal climate forecasts offers a valuable tool for predicting dengue risk in advance.

What did the researchers do and find?

  • We developed a superensemble of probabilistic models to predict dengue incidence across Vietnam up to 6 months ahead. The predictive ability of the superensemble was assessed using multiple verification metrics and compared to a baseline model at multiple lead times, seasons, and locations.

  • We found that the model superensemble generated more accurate predictions than the baseline model 1 to 3 months ahead but not 4 to 6 months ahead.

  • The model superensemble, however, was better able to predict outbreaks than the baseline model across all lead times.

  • The predictive ability of the model varied with geographic location, forecast horizon, and time of the year and performed best in the peak season in areas experiencing a high level of transmission.

What do these findings mean?

  • While outbreaks of infectious diseases are difficult to predict, particularly several months ahead, model superensembles, which combine multiple climatic drivers with dengue surveillance data at the time of forecast issue date, provide a useful decision-support tool.

  • Early warning systems driven by seasonal climate forecasts could shift dengue control towards a more preventative approach, guiding bespoke and targeted public health interventions and a more efficient allocation of scarce resources.

Introduction

Dengue is a mosquito-transmitted viral infection spread by Aedes mosquitoes in urban and peri-urban environments in tropical and subtropical countries [13]. About half of the global population is at risk of dengue transmission [4,5]. Dengue infection is characterised by flu-like symptoms that may include sudden intermittent high fever, retro-orbital pain, muscle and joint pain, severe headache, and widespread red skin rash. Its treatment typically includes the supportive care of symptoms. There is no specific antiviral treatment for dengue, and efforts to control transmission focus on controlling vector populations [6]. A live, attenuated, tetravalent dengue vaccine has demonstrated efficacy in 2 large-scale Phase III trials [7] and has now been licensed. However, while the vaccine shows high levels of protection against disease in people with previous exposure to dengue viruses, it may increase the risk of severe dengue if given to seronegative individuals [7]. A lack of reliable and scalable tests for seropositivity currently hinders wider rollout of the Dengvaxia vaccine and so most countries still primarily rely on insecticide-based mosquito control interventions to limit dengue virus transmission [8]. The increasing resistance to insecticides highlights the need for targeted and effective interventions [9].

Vietnam is particularly affected by dengue with an estimated burden of about 2 million yearly infections [5,10], although, on average, only 95,000 cases have been reported annually to the Ministry of Health over the period 2002 to 2020. Underreporting may be due to 2 main factors. First, dengue surveillance in Vietnam is mostly passive, relying on clinical cases reported by patients seeking healthcare [11]. Second, up to 80% of the cases may be asymptomatic or minimally symptomatic and will likely not seek healthcare [12]. The economic impact of dengue in Vietnam is estimated to be US$30 to US$95 million per annum [10,13,14]. Dengue in Vietnam is primarily spread by Aedes aegypti and, to a lesser extent, by Aedes albopictus [15,16].

In Vietnam, dengue is characterised by strong seasonality and substantial interannual and spatial variability (S1 Fig). Dengue exhibits different behaviour in different parts of the country. In the north, where temperatures are lower than in the rest of the country, most provinces have few or no cases. An exception is Hanoi, which has reported, on average, about 8,700 dengue cases per year over the past 10 years. In central and southern provinces where temperatures are warmer, many provinces report thousands of cases annually, albeit with large interannual variation.

Dengue control in the country primarily involves community engagement and mobilisation to reduce breeding sites and outdoor low-volume insecticide spraying in the vicinity of reported dengue cases to kill adult mosquitoes [11]. One limitation of dengue control measures is that they are essentially reactive, meaning they take place after cases have occurred. Thus, vector control, dengue diagnosis, and reporting all suffer delays which vary across time and space [17]. This situation hinders the timely generation and communication of information on when and where transmission occurs, limiting the ability of health professionals to plan and execute control measures.

If accurate predictions are available, public health decision-makers and planners could design, implement, and target interventions to the most at-risk places in a timely fashion. Disease models driven by Earth observations have been valuable for predicting dengue risk ahead of time, supporting decision-making in multiple settings [1820]. Climate variation is one of the main drivers of dengue ecology. It is known that temperature regulates the development, biting, reproduction rates, and the spatial distribution of Aedes mosquitoes [2123]. The titre and replication of dengue viruses within mosquitoes is temperature dependent [21,24]. For example, rising temperatures increase dengue transmission to an optimum range of 26 to 29°C [25], and large diurnal temperature ranges (DTRs; above 20°C) reduce transmission and increase mosquito mortality [26]. Mosquito survival is also affected by humidity. Research indicated that A. aegypti eggs survive across a wide range of humidity and temperature combinations. However, Ae. albopictus eggs experience high mortality when relative humidity is lower than 95% if temperatures are above 22°C [27]. Under high humidity conditions (i.e., above 80%), adult mosquitoes survive and remain active by replenishing transient body fluid depletion with plant sugars and host blood [28]. This behaviour is also observed in lower humidity environments, although to a lesser extent [28]. The effect of precipitation on dengue is related to the creation or flush away of mosquito breeding sites [29]. The effects of wind speed have been seldom explored in the literature. However, research suggests that wind speed reduces the biting activity of mosquitoes, reducing dengue risk [30]. Most of the effects of climate on dengue incidence are delayed by 1 to 2 months due to their effect on the life cycle of both mosquitoes and the dengue virus [31]. Thus, the climatic conditions just before the transmission season starts may be indicative of dengue incidence in the following dengue transmission season [32,33].

Multiple studies have highlighted the potential usefulness of seasonal climate-driven epidemiological surveillance for decision-making and planning [32,3436]. These studies have used subseasonal (i.e., between 2 weeks and 2 months ahead) forecasts to inform disease models and compute predictions of dengue risk. There has been limited progress in using subseasonal to seasonal climate forecasting to compute prospective forecasts on a routine basis. There are several challenges for implementing operational and sustainable subseasonal (henceforth seasonal) early warning systems [37]. Some of these challenges include the lack of multi-decadal health data sets with which to train and validate seasonal climate-driven early warning systems, the common mismatch of scales between climate data outputs and data used for decision-making, and a general lack of consensus as to how to communicate uncertainties to users [36].

Dengue early warning systems driven by Earth observations and seasonal climate forecasts have been proposed using a range of modelling approaches [38], including autoregressive integrative moving average (ARIMA) [39], point forecasts [32,40], spatiotemporal Bayesian hierarchical models [18,19], least absolute shrinkage and selection operator (LASSO) regression [41,42], and machine learning [43]. Often, models are validated using block cross-validation to select the model specification with the lowest out-of-sample predictive error [18,32]. This approach takes advantage of all available data to make repeated out-of-sample model predictions, which increases the robustness of model adequacy and verification statistics. One drawback of this method is that it does not preserve the time ordering of the data. Also, predictions are computed for some time periods using a model trained on data from a later time period [44].

Previous dengue risk prediction studies have relied on outputs from 1 or 2 competing models [18,32]. However, combining forecasts from multiple competing models into a superensemble may result in more accurate predictions than those from any individual model [38,45]. The use of model superensembles for the development of dengue early warning systems has been seldom explored. Moreover, predictions are typically made for a selected year or month [18,32,34] rather than for a series of lead times into the future, or a whole season [19,41]. In some cases, systems are designed exclusively for research purposes in isolation from relevant stakeholders who may become potential users. This situation is partly due to the difficulties related to the integration of these systems into existing public health procedures and occasional lack of technical expertise.

Typically, dengue early warning systems are based on deterministic models [32,39,40] which may underrepresent heterogeneity and stochastic cessation of transmission. However, decision-makers are increasingly interested in understanding the uncertainties related to the models used to develop decision-support tools and in the probabilities that an event of public health concern may or may not take place [34,46,47]. Spatiotemporal probabilistic models have the advantage of being able to quantify the probability that an event (e.g., an outbreak) may occur at specific times and for specific locations. Public health officials may be more inclined to take action if the probability of observing an outbreak exceeds a certain value [18]. Both modellers and decision-makers should pay attention to and agree on the definition of outbreaks thresholds so that model predictions are a useful guide for planning and decision-making.

In several countries, including Vietnam, outbreaks are defined using a so-called endemic channel [48,49], which corresponds to the mean number of cases per month or season over a long-term period [50]. In Vietnam, endemic channels are defined for each province using the last 5 years of dengue surveillance data. Outbreak years are removed from the computation of the endemic threshold. When dengue cases exceed the mean plus 2 standard deviations, an outbreak is declared. One limitation of this approach is that often, outbreak years are removed arbitrarily or quasi-quantitatively to increase the sensitivity of the outbreak threshold [50]. In areas where dengue incidence is typically low (e.g., less than 10 cases per month), the endemic channel may be frequently exceeded, generating statistical alarms of little public health importance [51]. Despite these limitations, endemic channels are widely used for dengue control decision-making in a variety of countries and provide a practical decision point around which forecasts can be targeted [52,53].

Here, we introduce a superensemble of probabilistic spatiotemporal hierarchical dengue models driven by Earth observations and seasonal climate forecasts. The model framework was codesigned with stakeholders from the World Health Organization (WHO), the United Nations Development Programme, the Vietnamese Ministry of Health, the Pasteur Institute Ho Chi Minh City, the Pasteur Institute Nha Trang, the Institute of Hygiene and Epidemiology Tay Nguyen (TIHE), and the National Institute of Hygiene and Epidemiology (NIHE). The system is designed to generate monthly estimates of dengue risk across Vietnam (331,210 km2) at the province level (n = 63) in near-real time (i.e., current time minus processing time). The superensemble is used in an expanding window time series cross-validation (TSCV) framework [54] to generate probabilistic dengue forecasts, which allow us to calculate the probability of exceeding predefined dengue outbreak thresholds for a forecast horizon (i.e., lead time) of 1 to 6 months.

The superensemble constitutes the dengue fever component of a forecasting system called D-MOSS (i.e., Dengue forecasting MOdel Satellite-based System). The system operates using a suite of Earth observation data sources from satellites and has its first implementation in Vietnam. Vietnam is divided into 63 provinces which are subdivided into 713 districts and has an estimated population of 95.5 million people. The D-MOSS system produces results at the province level and is accompanied by an assessment of its predictive ability that is applied consistently across the whole of Vietnam. The intention is to give an overall evaluation of the performance of our method, rather than an assessment that pertains to the characteristics of any particular province.

Materials and methods

The analysis plan was described in a UK Space Agency grant proposal for the International Partnership Programme in September 2017. The aims, objectives, and proposed methodology were developed prior to data analysis and are described in S1 Text. Changes to the analytical plan in response to suggestions from reviewers took place between August and October 2020. These changes include (1) calculating the bias and sharpness of each of the competing models included in the superensemble to decompose the overall prediction error; (2) formulating a climate naïve baseline model that captures the spatial dependency and seasonality for each province and estimates the same seasonal average of dengue incidence each month; (3) modifying the model superensemble specification from Bayesian moving averaging to model stacking based on the continuous rank probability score (CRPS); (4) using weights based on the variance of errors rather than maximum likelihood; and (5) modifying the colour palette and the methodology used for communicating and visualising the outputs from the probabilistic superensemble. Our study did not require a separate ethical approval.

Dengue surveillance data

Monthly dengue cases were obtained from the Vietnamese Ministry of Health. Data were retrieved for the period August 2002 to December 2019 at the province level (n = 63). Data comprised suspected and confirmed dengue cases, although there was no indication as to how many cases fell within each category. The data set did not contain serotype-specific information.

According to the national guidelines on dengue surveillance, suspected cases are defined as people living in or coming from endemic areas or from areas that have had dengue outbreak foci over the previous 14 days with manifestation of sudden high fever lasting for 2 to 7 days and with at least 2 of the following signs: hemorrhagic manifestation (positive tourniquet test, hypodermic petechiae or ecchymosis, and stump hemorrhage or nasal hemorrhage), headache, anorexia, nausea, vomiting, skin flush, rash, muscular, joint and orbital pains, writhe, unconsciousness, pain in liver area or pain when pressing on the liver area, hepatomegaly, thrombocytopenia, or increased hematocrit. Confirmed cases correspond to those confirmed by laboratory tests using MAC-ELISA, PCR, and NS1, or virus isolation. Over the period 2002 to 2019, there have been no changes in the classification of suspected and confirmed dengue cases in Vietnam.

Historical demographic and land cover data

Total population per province and per year were retrieved from the Socioeconomic Data and Applications Center (SEDAC) Gridded Population of the World project version 4.11 [55] at a 1 km2 resolution, 5 yearly for the period 2000 to 2020. Intervening years were generated using linear interpolation. The percentage of urban, peri-urban, and rural land cover per province and per year for the period 2002 to 2019 was derived from the ESA CCI Land Cover project [56,57], which describes the land surface into 22 classes at a spatial resolution of 0.00277 degrees.

Historical Earth observation data

Minimum, maximum, and mean air temperature at 2 metres above ground (°C) were derived from MODIS daily L3 global land surface temperature products [58] with a spatial resolution of 1 km2. Precipitation amount per day (mm day−1) was initially retrieved from from the Tropical Rainfall Measurement Mission [59] at a spatial resolution of 25 km2 up to April 2014. After April 2014, precipitation data were obtained from the Global Precipitation Mission [60] at a spatial resolution of 10 km2. Daytime-specific surface humidity (dimensionless) was calculated using the daytime MODIS L2 water vapour near infrared MOD 5 products [61] with a spatial resolution of 1 km2. Average daily wind speed at 10 metres above ground (m s−1) was retrieved from the European Centre for Medium-Range Weather Forecasts (ECMWF) ERA-5 reanalysis [57] for the period 2002 to 2011, at a spatial resolution of 31 km2. After 2011, wind speed data were obtained from the NOAA Climate Forecast System [62] at a 20 km2 resolution. Monthly sea surface temperature anomalies for the Niño region 3.4 (5°S−−5°N and 170°−120°W) were obtained from the NOAA Center for Weather and Climate Prediction Climate Prediction Center [63] for the period 2002 to 2020.

Earth observation data were obtained as a gridded data set in netcdf format. Sociodemographic data were averaged at the province level using the rasterstats module in Python 3.6 [64]. When a polygon covered only part of a grid cell, averages were calculated for the approximate fraction of the cell covered by the polygon (rounded 1/100). The administrative boundaries of each province were defined using a shapefile provided by the Vietnamese General Department of Preventive Medicine, Hanoi.

Climate data were aggregated at the province level using population weighted averages for each month. SEDAC population data, although gridded, had identical values at each pixel falling within a given administrative district (i.e., admin 2 level). SEDAC data could not be used for the computation of population weighted averages as each grid cell within an administrative district would carry the same weight. As a compromise, we used WorldPop [65] estimates for which population estimates vary per pixel. Population data from the WorldPop project [65] at a 100 m2 spatial resolution was used to calculate annual gridded weights for each province using the rasterstats module in Python 3.6 [64]. At the time of processing, WorldPop data were only available for the years 2009, 2010, 2015, and 2020. Intervening years were calculated using linear interpolation. Province-specific population-weighted averages were calculated as follows:

x^=i=1nwixii=1nwi, (1)

where n is the total number of pixels i falling within a given province polygon, x is the value of each climate predictor at pixel i, and w denotes the pixel-specific weights calculated as

wi=piN, (2)

where pi is the pixel-specific population count, and N is the total population count falling into a given province.

There were multiple changes in the administrative boundaries of the provinces over the study period. The last change took place in 2008. Consequently, we froze the administrative boundaries of the country at their 2008 level. Dengue data were allocated to each of the 2008 level administrative boundaries using district-level records. The allocation of dengue cases to each province was conducted by the General Department of Preventive Medicine, Hanoi.

Seasonal climate forecasts

Seasonal climate forecasts of minimum temperature, maximum temperature, daily precipitation, specific humidity, wind speed, and sea surface temperature anomalies for the Niño region 3.4 were obtained from the UK Met Office Global Seasonal Forecasting System version 5 (GloSea5) [66,67]. GloSea5 comprises 42 ensemble members built around a high-resolution climate prediction model (HadGEM3). Ensemble members differ due to small stochastic physics perturbations provided by the Stochastic Kinetic Energy Backscatter v2 [68]. GloSea5 has a resolution of 0.83 degrees in latitude and 0.56 degrees in longitude for the atmosphere and 0.25 × 0.25 degrees for the ocean. GloSea5 has 2 major components, the forecast itself and the associated hindcasts or historical re-forecasts, which are used for calibration and assessment. An evaluation of the predictive ability of the GloSea5 system can be found elsewhere [66]. S2 Fig shows that compared to the climatology (i.e., the month-specific mean for each variable), the out-of-sample errors of the GloSea5 system are considerably low.

Seasonal climate hindcasts

Hindcast data (i.e., historical forecasts) from the GloSea5 system [66,67] were retrieved using the Copernicus Climate Data Store [57] at monthly time steps for each of the 28 ensemble members, for lead times of 1 to 6 months ahead, and for the period January 2007 to December 2016. At the time of the computations, data for the period May to October 2016 were unavailable from the Copernicus Climate Data Store.

Model specification

Let Yi,t be the number of dengue cases for province i = 1, ⋯, I and time t = 1, ⋯, T where I is the total number of provinces in the data set, and T is the total number of time steps for which the model is fitted using Bayesian generalised linear mixed models (GLMM). Models were fitted with a conditional negative binomial distribution given the number of dengue cases in the previous time step. The general algebraic definition of the models at the linear predictor scale is given by

log(μi,t)=α+log(Pi,a[t])+ρlog(Yi,t1+1)+kβkXk,i,t+jϵjLj,a[t]+γi,a[t]+δi,m[t]+ui+νi, (3)

where α corresponds to the intercept; log (Pi,a[t]) denotes the logarithm of the population at risk for province i and year a[t], included as an offset to adjust case counts by population; log (Yi,t−1+1) denotes the logarithm of the observed number of dengue cases in the previous month plus one with an autoregressive parameter ρ; X is a matrix of k seasonal meteorological explanatory variables with regression coefficients β; and Lj,a[t] is a matrix of j land cover (peri-urban and urban) variables with regression coefficients ϵ. Long-term trends are modelled using province-specific unstructured random effects for each year (γi,a[t]). Seasonality is accounted for by using province-specific structured random effects for each calendar month (δi,m[t]) with first order autoregressive prior to allow each month to depend on the previous one. Unknown confounding factors, such as interventions and spatial dependency structures representing, for example, human mobility, were incorporated using structured (νi) and unstructured (ui) random effects for each province i.

Spatial random effects were specified using a Besag–York–Mollie model [69], which incorporates a spatial effect with a Gaussian exchangeable prior to account for unstructured variation and a spatial effect with an intrinsic conditional autoregressive prior to account for spatially structured variability. Delayed effects of meteorological factors on dengue were accounted for by incorporating a 3-month moving average of temperature, precipitation, specific humidity, and DTR centred at lag of 1 month (i.e., each value is given the average of lag 0 to a lag of 2 months). Sea surface temperature anomalies in the Niño region 3.4 were lagged 0 to 3 months based on previous research [20,33,70] and exploratory analyses. No delayed effects were considered for wind speed based on previous studies, indicating a highest effect of wind at a lag of 0 months [71].

Flat priors were set to regression coefficients (α,ρ,β), and penalising complexity priors were assumed for the precision for all random effects [72]. Priors were specified using a Gaussian distribution. We did not explicitly specify a prior distribution for the dispersion parameter (ϕ). We used the default pc.gamma specification in the INLA R package [73], which assumes a Gamma distribution with values of ϕ within the interval 1–5668.2. Models were fitted in R version 3.6.1 using the INLA package [73]. The relevant R code is available at https://github.com/FelipeJColon/paper_dengue_superensemble.

Model selection

The best subset of seasonal climate predictors leading to the lowest observed prediction discrepancies for a given model was obtained using an expanding window TSCV algorithm [54]. Land cover variables were included in all competing models as they varied annually and are unlikely to change significantly at monthly time steps. We iteratively fitted all possible models containing 1 seasonal climate predictor at the time, then 2 seasonal climate predictors, and so on, until all seasonal climate variables were included in a full model [20]. Thus, we tested 128 unique model specifications across 114 forecast months (i.e., January 2007 to December 2016). The predictive ability of each model was evaluated using the CRPS which generalises the mean absolute error for the case of probabilistic forecasts. The advantage of the CRPS compared to the mean absolute error (MAE) or root mean squared error (RMSE) is that the CPRS does not focus on a specific point of the probability distribution of the forecasts, but on their distribution as a whole. CRPS values were computed in R using the SpecsVerification package [74]. In response to peer review comments, we also evaluated the bias of the forecasts with the lowest CRPS to investigate whether models systematically over- or underpredicted dengue counts [75]. The bias for a continuous forecast at time t was defined as [76]

Bt(Pt,xt)=12×(Pt(xt)), (4)

where Pt is the empirical cumulative distribution function of the prediction for the true value xt [75]. Bias values range from −1 to 1, with Bt = 0 as an ideal value. The sharpness of the best performing models was assessed to investigate their ability to generate predictions within a narrow range [75]. The sharpness of forecasts at time t was defined using the normalised median absolute deviation about the median (MADN) [75,76]. We defined sharpness as

St(Pt)=10.675median(|ymedian(y)|), (5)

where y is a forecast with cumulative distribution function Pt; the division by 0.675 ensures a value equivalent to the standard deviation if the predictive distribution is normal [75]. Sharpest models have St = 0, while blurred forecasts have St→∞ [75]. Bias and sharpness were calculated using the scoringutils [76] R package.

TSCV was implemented using an expanding window approach dividing the data set into multiple training and testing sets. The initial training set comprised data from August 2002 to December 2006. Each time step (k), a further month of data was added to the training set until the training set contained n-6 observations, where n is the total number of observations in the set. The testing set comprised the climate hindcast data for the 6 months immediately after the last observation in the training set for each geographical area. Seasonal climate hindcast data were used to simulate the behaviour of an operational system over the period 2007 to 2016. We calculated the mean CRPS for each model specification across all time steps and ensemble members. The best 5 performing models were selected to create a superensemble.

Prospective lagged dengue cases

The number of dengue cases occurring at time t is directly dependent on the number of cases that occurred in the recent past. Previous research suggests that including the logarithm of the number of cases in the previous month log(Yt−1) helps accounting for such temporal correlation in disease transmission [77]. Additionally, incorporating log(Yt−1) as a covariate improves the predictive ability of seasonal climate-informed disease models by reducing residual dispersion [77]. One complication of accounting for the number of cases in the previous month in an operational system is that log(Yt−1) in the temporal window of the forecast can only be known up to time t+1. More specifically, to generate dengue forecasts 1 month ahead, log(Yt−1) for time t+1 corresponds to the logarithm of the number of dengue cases at time t. We used a GLMM with a negative binomial specification to estimate the log(Yi,t−1) for time t+1 to time t+6 to ensure that the number of cases in the previous month were exactly the same for all competing models. Thus, reconstructed lagged dengue time series was exactly the same for all models. The algebraic definition of the model is

log(μit)=α+log(Pi,a[t])+ρlog(Yi,t1+1)+γi,a[t]+δi,m[t]+ui+νi, (6)

with α as the intercept; log (Pi,a[t]) as the population at risk in province i at time a[t], included as an offset; log (Yi,t−1+1) is the logarithm of the observed dengue cases plus one in the previous month with regression coefficient ρ; γi,a[t] as province-specific unstructured yearly random effect; δi,m[t] as province-specific structured monthly random effect with an AR1 auto-correlation term; and ui and νi as province-specific structured and unstructured random effects.

We then predicted the number of dengue cases for time t+1. The logarithm of the predicted number of cases at time t+1 was then used as log(Yi,t−1) for time t+2. We repeated these steps for each lead time in the forecast until time t+6. Note we assumed a conditional negative binomial distribution to generate one-step-ahead point predictions for the reconstructed lagged dengue time series. Once the reconstructed lagged dengue time series was complete, we fitted all competing models as indicated in the Model selection section.

Baseline model

We developed a historical baseline model, which forecasts the same seasonal average of dengue incidence every month at each model run following [18,34,78] to compare all models to a common reference. The baseline model is specified as follows:

log(μit)=α+log(Pi,a[t])+δi,m[t]+ui+νi, (7)

where α is the intercept; log (Pi,a[t]) is the population at risk in province i at time a[t], included as an offset; δi,m[t] are province-specific monthly structured random effects with an AR1 auto-correlation term; and ui and νi as province-specific structured and unstructured random effects. We compared the predictive ability of the superensemble to that of the baseline model using the continuous rank probability skill score (CRPSS). The CRPSS is defined as follows:

CRPSS=1CRPSfCRPSb, (8)

where CRPSf is the CRPS value of the forecast, and CRPSb is the corresponding CRPS value of the baseline model.

Model superensemble

Given a number of competing models, we generated a model superensemble [38] stacking models according to their CRPS following suggestions from peer reviewers. Stacking involves averaging predictions from multiple models using weighted averages. We found model stacking outperformed Bayesian moving average. Therefore, this technique was adopted for the final version of the system. Stacking was performed in 3 stages [79]. On the first stage, we simulated 1,000 samples from the posterior distribution of dengue cases for each of the best performing models and the baseline model. Then, we calculated their CRPS for the historical period comparing the samples against observed dengue counts. In the second stage, optimal weights [80,81] were calculated for each model as follows:

wi=(Vi)1j=1n(Vj)1, (9)

where V is the variance represented by the square of the CRPS of the ith forecast errors. A new set of weights were computed each time a forecast was issued using training data from all previous years. On the last stage, the 1,000 samples were combined using a weighted average. We then calculated the 2.5th, 50th, and 9.75th percentiles of their distribution.

Outbreak detection evaluation

The predictive ability of the model for outbreak detection was evaluated using the Brier score [82]. For the Brier score, smaller values indicate better predictions. Four moving outbreak thresholds were defined to evaluate the predictive ability of the superensemble for outbreak prediction: (i) the endemic channel plus 1 standard deviation; (ii) the endemic channel plus 2 standard deviations; (iii) the 75th percentile of the distribution of dengue cases per month; and (iv) the 95th percentile of the distribution of dengue cases per month. The endemic channel was calculated as the number of dengue cases per month and per province over the previous 5 years in agreement with current practice at the Vietnamese Ministry of Health. The 75th and 95th percentiles were calculated over the whole observational period at each time step. We then calculated the probability of exceeding the moving outbreak threshold generating 1,000 samples from the posterior distribution of the point forecasts to reflect forecast uncertainty. The Brier score was calculated as follows:

BS=1Nt=1N(ftOt)2, (10)

where N is the number of predictions; ft is the forecast probability that an outbreak may happen; and Ot takes the value of 1 if there was an outbreak or 0 if there was no outbreak. Brier scores were computed in R using the scoring package [83].

Public health officials may be more likely to take preventive action if the probability of observing an outbreak exceeds a certain value [18]. We undertook a signal detection analysis to determine the ability of the forecast probabilities to classify the predicted number of dengue cases as outbreaks. There are 4 possible outcomes in this analysis: hit (true positive), correct rejection (true negative), false alarm (false positive), and missed (false negative). A count for each of the 4 possible outcomes was produced.

Relative value of using a forecasting system

Unlike other measures, the relative value (V) of the forecasts generated by the superensemble depends on requirements set by the user [84]. Typically, V is evaluated in monetary terms and is particularly useful when the probability of occurrence of an adverse event (e.g., a major outbreak) is known. If the probability of occurrence of an outbreak is greater than the ratio of the cost of taking preventive action divided by the loss incurred by not taking action (C/L ratio) and an outbreak occurs, then it will pay off to take action. If the probability of occurrence of an outbreak is lower than the C/L ratio, then it does not pay off to take preventive action. If the probability is equal to the C/L ratio, it does not matter if action is taken or not. Our analysis provides an initial screening process to identify areas where a forecast is most likely to be cost-effective and thus, target of comprehensive economic evaluation studies in the future.

Given that data on the cost of taking preventive action and on the losses incurred by not taking action were unknown to us at the time of writing this manuscript, V was estimated for a range of theoretical C/L ratios following [84] by comparing the mean cost of using the forecasting system for outbreak detection compared to the mean expense incurred by either never preventing outbreaks or, on the contrary, taking preventive action every month of the year. V takes a value of 1 if the forecast is perfect and a value of 0 if it is no better than the default action plan. If V is negative, it indicates that the forecast is so poor that it would be more cost effective not to use it. The algebraic definition of V is

V=E(S)E(A)E(S)E(P), (11)

where E(S) is the theoretical expense incurred by taking preventive action each month or the losses incurred by no taking action at all even when an outbreak occurred, whichever is the least expensive method when not using the forecast; E(A) represents the total cost of the forecast calculated as the cost of type 1 (false positive) and type 2 (false negative) errors plus the cost of acting when an outbreak was predicted and it occurred (true positive); and E(P) indicates the cost incurred with a perfect forecast where outputs are exclusively true positives and true negatives.

Results

Model selection

We fitted a total of 14,592 different models (128 unique model specifications across 114 forecast months, i.e., January 2007 to December 2016) by using all climate variable in isolation, as well as all their possible combinations. One major issue with regression models is potential overfitting which may arise from data redundancy due to high correlation between predictors. The Spearman rank correlation coefficients calculated for all pairs of candidate predictors was less than 0.7, except for between minimum and maximum temperature (ρ = 0.94). However, these 2 variables were not present in any of the best performing models.

Model superensemble evaluation

The 5 models with the lowest CRPS values (i.e., best performing) and the baseline model were used to generate a superensemble. It is noted that the 5 best performing models were the only models with a CRPS below 90 when computing prospective predictions. For CRPS, lower values indicate a smaller difference between the forecasts and the observations. There was strong seasonal variation in the weights assigned to each model (S3 Fig). All competing models had similar weights across the whole evaluation period (January 2007 to December 2016), whereas the baseline model had the lowest weights. We note that the weights assigned to the baseline model gradually decreased with time.

Averaged across all lead times, the superensemble led to slightly lower CRPS values than each of the competing models, including the baseline (Table 1). However, when stratified by lead time, the superensemble outperformed the competing models at all time leads but only outperformed the baseline at leads of 1 to 3 months (Fig 1A). The uncertainty of the forecasts increased with lead time. This is also evident when using the CRPSS for which values above 0 (indicating better predictions) were observed for the superensemble only for leads 1 to 3 months (Fig 1B). The CRPSS can be interpreted as the relative improvement observed from using a model compared to a reference baseline. Thus, a value of 0.4 would suggest that the values from a given model have an absolute error that is 40% smaller than the absolute error of the baseline model. The CRPSS indicates that at a lead time of 1 month, the superensemble and the competing models outperform the baseline by about 42%, whereas at a time lead of 6 months, their absolute error is 12.5% larger than that of the baseline model.

Table 1. Verification metrics and seasonal climate predictors of the model superensemble and the best performing models.

Model CRPS Bias Sharpness Seasonal climate predictors
Baseline 79.9 0.37 0.01 None
Superensemble 73.4 0.44 1.90 All included in Models 1–5
1 79.8 0.49 3.12 SH02, WS0, DTR02, SST03
2 80.2 0.49 1.67 SH02, DTR02
3 80.9 0.48 1.69 SH02, WS0
4 81.4 0.50 2.79 SH02, DTR02, SST03
5 81.5 0.50 2.67 SH02, WS0, SST03

DTR02, diurnal temperature range averaged over a 3-month period lagged 0 to 2 months (°C); SH02, specific humidity averaged over a 3-month period lagged 0 to 2 months (dimensionless); WS0, wind speed in the current month (m s−1); SST03, sea surface temperature anomalies in the Niño region 3.4 averaged over a 4-month period lagged 0 to 3 months (°C). CRPS, continuous rank probability score. CRPS and sharpness assume values between 0 and infinity with an ideal value of 0. Bias assumes values between −1 and 1, with 0 as ideal.

Fig 1. Verification metrics by lead time.

Fig 1

Variation across all lead times averaged across the whole of Vietnam for (A) the CRPS, (B) the CRPSS, (C) the bias of the forecasts, and (D) the sharpness of the forecasts. Red lines indicate the performance of the model superensemble; blue lines depict the metrics for each of the 5 competing models; and grey lines indicate the behaviour of the baseline model. CRPS and sharpness assume values between 0 and infinity, with 0 representing a perfect forecast. Bias assumes values between −1 and 1, with 0 representing unbiased forecasts. CRPS, continuous rank probability score; CRPSS, continuous rank probability skill score.

All models showed a tendency to overestimate the predicted number of dengue cases indicated by bias values above 0 (Table 1). Bias also increased as the forecast horizon increased, suggesting that the tendency to overestimate increases as the forecast horizon increases (Fig 1C). The model with the smallest bias was the baseline followed by the superensemble. The sharpness of the forecasts deteriorates with lead time (Fig 1D) as smaller values indicate a better forecast. The superensemble showed slightly worse sharpness than 2 of the competing models for lead times of 4 to 6 months. It is noted that the sharpness of the baseline model was always 0 as it generates the same seasonal values for all years in the data set (Table 1).

Across all metrics, the predictive ability of the superensemble deteriorated as the forecast horizon increased from 1 to 6 months. This situation is also evident where the forecast ensemble mean and its 95% credible interval (i.e., the interval in the domain of the posterior probability distribution) are plotted against the observed number of dengue cases (S4 Fig). Notice that the accuracy of the predictions worsens as the forecast horizon expands. We noted that the credible intervals of the predictions gradually became narrower as the number of months used to train the models increased (S4 Fig).

The predictive ability of the superensemble also varied with the month of the year. Overall, better performance was observed over the period July to December for CRPS, CRPSS, and bias (Fig 2A–2C). The sharpness of the forecasts, however, deteriorated over the same period (Fig 2D).

Fig 2. Verification metrics by month of the year.

Fig 2

Variation across the months of the year averaged across the whole of Vietnam for (A) the CRPS, (B) the CRPSS, (C) the bias of the forecasts, and (D) the sharpness of the forecasts. Red lines indicate the performance of the model superensemble; blue lines depict the metrics for each of the 5 competing models; and grey lines indicate the behaviour of the baseline model. CRPS and sharpness assume values between 0 and infinity with an ideal value of 0. Bias assumes values between −1 and 1, with 0 as ideal. CRPS, continuous rank probability score; CRPSS, continuous rank probability skill score.

Compared to the baseline model, the predictive ability of the forecast evaluated using the CRPSS showed significant spatiotemporal variation. Fig 3 shows that the CRPSS was consistently better than the baseline (red shaded areas) across most of the country for the period July to January. From February to June, however, the predictive ability of the baseline model is better than that of the superensemble for multiple provinces identified by CRPSS values below 0 particularly during March and April.

Fig 3. Predictive ability across time and space.

Fig 3

Spatiotemporal variation of the CRPSS of the model superensemble. Orange shaded areas indicate a better performance of the superensemble compared to a baseline model. Blue areas indicate a lower performance of the superensemble compared to a baseline model. The shapefile used to create this figure was obtained from DIVA-gis (https://www.diva-gis.org). CRPSS, continuous rank probability skill score.

Outbreak detection

Predictive ability was also evaluated by comparing the predicted probability of predicting outbreaks using the Brier score and 4 moving outbreak thresholds. Observed outbreak months were defined as months where the number of predicted dengue cases exceeded outbreak thresholds. As expected, highest predictive ability was achieved at a lead time of 1 month, after which the predictive ability of the superensemble gradually declined (Fig 4A). Across the forecast horizon, the highest predictive ability was observed when using an outbreak threshold based on the endemic channel plus 2 standard deviations, closely followed by the 95th percentile. Stratified by month of the year, the predictive ability of the superensemble was generally greater between April and September for all thresholds except for the 75th percentile. When stratified by month, predictive ability was also larger using the endemic channel plus 2 standard deviations closely followed by the 95th percentile of the distribution of dengue cases.

Fig 4. Temporal variation of the Brier score.

Fig 4

Temporal variation of the Brier score averaged across all 63 Vietnamese provinces stratified by (A) lead time and (B) month of the year. The red lines indicate variation in the Brier score using percentile-related moving outbreak thresholds. The blue lines indicate variation in the Brier score using the endemic channel–related thresholds. The Brier scores assume values between 0 and 1, with 0 as ideal.

There was significant spatial variation in the outbreak detection ability of the superensemble. Fig 5 shows that across all moving outbreak thresholds, outbreak detection ability was greater in the northern provinces compared to the central and southern provinces. It is noted that the 95th percentile threshold results in a larger number of provinces with low Brier scores (i.e., lower than 0.1) than the endemic channel plus 2 standard deviations.

Fig 5. Spatial variation in the Brier score.

Fig 5

Spatial variation of the mean Brier score averaged across all lead times calculated for 4 different moving outbreak thresholds over the period January 2007 to December 2016 and for each of the 63 Vietnamese provinces. Lower Brier scores (in orange) indicate a greater accuracy for detecting outbreaks. The Brier scores assume values between 0 and 1, with 0 as ideal. The shapefile used to create this figure was obtained from DIVA-gis (https://www.diva-gis.org).

A signal detection analysis indicates that using the superensemble with a moving outbreak threshold based on the endemic channel plus 2 standard deviations resulted in 73% correct predictions. Similarly, using the superensemble with an outbreak threshold based on the 95th percentile led to 72.5% correct predictions. The baseline model had the lowest predictive ability, with 61.3% correct predictions (Fig 6). However, using the 95th percentile outbreak threshold resulted in a slightly higher probability of detection (70% of all observed outbreaks) than using the endemic channel plus 2 standard deviations (68%). The baseline model had the lowest probability of detecting outbreaks (54.5%).

Fig 6. Signal detection analysis.

Fig 6

Proportion of hits (dark orange), correct rejections (light orange), false alarms (light blue), and missed outbreaks (dark blue) for the baseline model and the model superensemble. Similar results were obtained at all lead times.

Decision-support tools

Portraying prospective forecasts

The model superensemble was driven by seasonal climate forecasts to generate dengue forecasts for the period May to October 2020 using near-real time seasonal climate forecast data from the UK Met Office Global Seasonal forecasting system version 5 (GloSea 5) [66,67]. For the historical period (i.e., August 2002 to April 2020), the superensemble made slightly more accurate predictions (CRPS = 66.8, 95% CI 60.6 to 148.0) than a baseline model which forecasts the same incidence rate every month (CRPS = 79.4, 95% CI 78.5 to 80.5) at lead times of 1 to 3 months, albeit with larger uncertainty. For the period May to October 2020, the posterior median of the predictions made with the superensemble was also slightly more accurate (CRPS = 110, 95% CI 102 to 575) than the posterior median of the predictions made with the baseline model (CRPS = 125, 95% CI 120 to 168), albeit with larger uncertainty. Fig 7 shows an example of how this information could be portrayed to users. In the figure, solid black lines indicate the posterior mean of the predicted values for each ensemble member. The red lines depict the observed dengue cases which were not known to the model at the time of the computation and are included here as a reference. Dotted black lines indicate the 95% credible interval of the predictions. Shaded areas indicate the low risk, moderate risk, high risk, and very high risk areas based on the endemic channel calculated using historical data from the previous 5 years (Table 2).

Fig 7. Portraying prospective predictions.

Fig 7

Predicted dengue incidence rate for the period May to October 2020 for the Vietnamese province of Dong Nai using a model superensemble. The forecast was issued on May 10, 2020. The x axis indicates the time lead of the predictions. The y axis indicates the predicted dengue incidence rate. Solid black lines indicate the posterior mean estimate for each of the 42 forecast ensemble members. Red lines indicate the observed dengue incidence rate which was not known by the model at the time of the computation. The black dashed lines indicate the upper and lower bounds of the 95% credible intervals for each of the 42 ensemble members. The upper bound of the shaded areas indicate the month-specific risk based on the endemic channel plus 2 standard deviations calculated with historical data from the previous 5 years.

Table 2. Thresholds used to define low, moderate, high, and very high risk levels based on an endemic channel.
Risk level Endemic channel
Low 0 to mean
Moderate Mean to mean + 1 SD
High Mean + 1 SD to mean + 2 SD
Very high Larger than mean + 2 SD

Mean, the mean number of dengue cases in the previous 5 years; SD, standard deviation of the observed number of dengue cases in the previous 5 years.

S5 Fig shows the posterior mean of the predicted values for each ensemble member and for 4 Vietnamese provinces where the system is being piloted along with their corresponding 95% credible interval (dashed lines). It can be observed that there is little spread between the 42 ensemble members indicating little between-member variability. As expected, the credible intervals increase as the forecast horizon increases, reflecting the uncertainties associated with the seasonal climate models used to generate the forecasts. It is noted that for none of the 4 provinces, the predicted mean number of dengue cases is above the high-risk region, suggesting that the September to February transmission season may be at normal conditions.

The spatial distribution of the posterior mean of the Model predictions could also be portrayed using risk maps (Fig 8) so that users can be made aware of the high-risk areas. Maps indicating the probability of exceeding a predefined outbreak threshold could also be easily created using the system outputs.

Fig 8. Risk maps.

Fig 8

(A) Spatial distribution of the posterior mean of the predicted number of dengue cases 1 month ahead for a forecast initialised on May 10, 2020. (B) Spatial distribution of the probability of exceeding an outbreak threshold based on the 95th percentile. The shapefile used to create this figure was obtained from DIVA-gis (https://www.diva-gis.org).

Relative value

An analysis of the relative value of using dengue forecasts generated by the superensemble can be undertaken using the model outputs to guide decisions-making processes. For such analysis, a value index ranging between 0 and 1 is used, with 1 indicating a perfect forecast [85,84]. The relative value of the forecasts is interpreted in comparison to either never taking preventative action or always taking action.

We defined a range of theoretical epidemic thresholds ranging between the 51st and the 99th percentiles of the distribution of dengue cases for the whole time series. Outbreak thresholds were province and month specific. Overall, the forecasts generated by the superensemble showed relative value in 76% of the provinces 1 month ahead, 73% of the provinces 2 months ahead, 68% of them 3 months ahead, and 65% of the provinces 4 to 6 months ahead (Fig 9). We note that the superensemble has relative value in areas where dengue is typically endemic such as the central and southern provinces. The superensemble had no relative value in the northern provinces where dengue is typically absent.

Fig 9. Spatial variation of the relative value of the forecasts.

Fig 9

Spatial variation of the relative value of the forecasts generated by the model superensemble 1 to 6 months ahead. Orange shaded areas indicate provinces where there is relative value (based on a range of theoretical cost–loss ratios and outbreak thresholds). Blue shaded areas indicate provinces where the superensemble had no relative value. The shapefile used to create this figure was obtained from DIVA-gis (https://www.diva-gis.org).

Discussion

This study details a probabilistic dengue early warning system based on a model superensemble, formulated using Earth observations and driven by seasonal climate forecasts. The system is able to generate accurate probabilistic forecasts of dengue metrics that have the potential for guiding policy- and decision-making processes in Vietnam. We make predictions for each of the 28 ensemble members of the seasonal climate forecast and generate 1,000 samples from the posterior distribution of the predicted number of cases.

Deciding which predictive model is the best from a suite of competing models is not a straightforward task. Each model carries somewhat different information of the modelled processes. Here, we present a method for reconciling between-model disagreements while improving forecast accuracy. The combination of models into a superensemble helps offset individual model biases across time and space [86]. Superensembles were initially developed for climate modelling [87] and have gradually gained popularity in disease modelling (see, for example, [38,45,88]).

Our novel dengue early warning system relies on probabilistic models to reflect forecast uncertainty and to explicitly assign probabilities to outcomes [89]. The system has been developed to aid policy- and decision-making processes in Vietnam with the guidance of key stakeholders in the Ministry of Health of Vietnam, WHO regional office, the Pasteur Institutes in Nha Trang and Ho Chi Minh, Province-level Ministries of Health, the Vietnamese NIHE, and the Tay Nguyen Institute of Hygiene and Epidemiology. A range of stakeholder engagement workshops, face-to-face meetings with users, and surveys were conducted to tailor the system to the users’ needs. Our results demonstrate that the forecasts generated by our spatiotemporal superensemble have the potential to guide changes to the current practice in dengue control towards a more preventative approach allowing bespoke and targeted public health interventions and a more efficient allocation of scarce resources. Research on use of the system by stakeholders and its effect on changing dengue control practices is currently ongoing.

We demonstrate that the superensemble slightly outperforms the predictive ability of the individual probabilistic models selected from a suite of top performing models at multiple lead times and months of the year in line with previous research [38,45,87,88]. Compared to a baseline model that predicts the same dengue incidence rate for each season and province, the superensemble generated more accurate predictions at lead times of 1 to 3 months but not beyond. Model performance was assessed using a range of verification metrics for probabilistic forecasts across time and space. To our knowledge, this is one of the first early warning systems informed by Earth observations to demonstrate predictive ability for prospective year-round dengue prediction in a robust out-of-sample framework. It is also one of the first prototypes for routine dengue early warning at multiple time leads.

We found that the performance of the superensemble varied with geographic location, forecast horizon, and time of the year. The system showed ability in predicting spatiotemporal variations in dengue cases and outbreak occurrence at forecast horizons of up to 6 months ahead. Predictions deteriorated, and uncertainty increased with lead time, as previously observed in other settings and diseases [36,42,75,90]. Forecasts improved, and the credible intervals decreased as time progressed and dengue data increased, likely due to an improvement in the associations learned by the superensemble. Forecast errors increased for the onset and peak of the transmission season, possibly due to substantial interannual variation (S6 Fig).

Relative to a baseline model, the superensemble made slightly more accurate predictions of the estimated number of dengue cases across most provinces, at lead times of 1 to 3 months in agreement with previous studies [18,42,45]. At lead times beyond 3 months, however, the baseline model made more accurate predictions possibly because, as suggested by [45], incorporating seasonal climate data into predictive models may increase model complexity at the expense of lower out-of-sample predictive ability. Previous research has found that climate naïve models may outperform climate-informed models [32,45]. Those studies, however, used different sources of climate and environmental data than our study and were conducted in different settings (i.e., Thailand, Peru, and Puerto Rico) where the effects of climate might be significantly different to those observed in Vietnam. In addition, those studies initialised their prospective forecasts at fixed times of the year, whereas our study initialised forecasts at each calendar month, providing the model with more opportunities to learn the relationships between dengue incidence and each of the climatic and sociodemographic predictors.

Outbreaks are difficult to predict, even more so at forecast horizons of several months ahead. Nevertheless, our superensemble demonstrated predictive ability for outbreak detection up to a lead time of 6 months, evaluated using proper scores over a suite of moving outbreak thresholds. One of the best performing outbreak thresholds was the endemic channel plus 2 standard deviations, computed using data from the previous 5 years. This method is currently in use across Vietnam [48,49], although with the difference of excluding outbreak years from the computation of the threshold. Following our results, however, we recommend the use of the 95th percentile as a threshold for outbreak detection as it gave slightly better results for detecting outbreaks. Recognising the limitations of the province-level data, it is encouraging that our predictions are accurate in most provinces up to 6 months in advance.

In disease forecasting, each decision (e.g., to prevent or not to prevent an outbreak) has an associated cost that will lead to a benefit or a loss depending on the outcome. Decision-makers have the task of selecting the action that minimises potential losses. We used a simple analysis [84] to assess the relative value of the forecasts generated by the superensemble. Our figures are only illustrative. Still, they highlight that using a dengue early warning system has relative value compared to not using a forecast. The superensemble had considerable value across most provinces. However, in northern provinces, where dengue is essentially absent, the forecast is predicted to have limited relative value compared to always preparing for an outbreak or never preparing for it. The assessment of the relative value of the dengue early warning system may help stakeholders justify public investment in the development and generation of seasonal forecasts or to help raise awareness of their potential value.

Although our proposed early warning system provides useful information for public health preparedness and response, it has some limitations worth mentioning. First, while our modelling framework incorporates important determinants of dengue occurrence such as climate and urbanisation, it does not explicitly incorporate, at this stage, the potential effects of other important determinants of disease such as the deployment of mosquito control interventions, vector indices, serotype-specific circulation, herd immunity, and the mobility of people and goods, all of which may lead to significant changes in the level of risk experienced locally [91]. Stakeholders in Vietnam have recently started collecting supporting data on vector indices and dengue virus serotypes. However, there is a lack of publicly available, continuous, and long-term data sets that could be used to inform modelling efforts. In this study, we account for some of the variation that might be attributed to these factors by using spatiotemporal random effects in each of the models included in the superensemble. Future developments of the system may incorporate some of these factors if data are made available. Second, the quality and consistency of the dengue data are affected by the limited confirmation of suspected cases through laboratory diagnostic test, leading to large uncertainties that are difficult to quantify. Third, our computations of dengue risk do not take into consideration uncertainties due to the potential under- or misreporting of dengue cases. Consequently, our model superensemble forecasts may underestimate the real number of cases occurring at any given time. Finally, our forecasts do not carry forward the uncertainty of the one-step-ahead forecasts used to simulate prospective lagged cases. Future developments may represent this uncertainty more accurately by using simulation-based path forecasts in a Bayesian framework [92].

Conclusions

We have demonstrated that a superensemble of probabilistic dengue models, formulated using Earth observations and driven by seasonal climate forecasts and a reconstructed/prospective estimate of lagged dengue cases, is useful to generate probabilistic predictions of dengue risk across Vietnam at multiple lead times and months of the year. We acknowledge that there are alternative state-of-the-art approaches to dengue forecasting that may be equally as effective as the one presented here. However, our aim was to develop a system that outperformed previous practice in Vietnam where no forecasting system was used to guide decision-making processes. A theoretical analysis showed that compared to not using a forecasting system, the superensemble has relative value across most of the country, suggesting that it has potential for guiding decision-making processes. The dengue forecasting system presented here has been rolled out across Vietnam and could be tailored for other dengue-endemic countries. Further work may include investigating the feasibility of producing probabilistic forecasts with sufficient predictive ability at the district or commune levels and a comparison with other statistical and mechanistic approaches.

Supporting information

S1 Text. Prospective analysis plan.

(DOCX)

S1 Fig. Observed dengue cases across Vietnam.

Time series of monthly dengue cases from the 63 provinces in Vietnam (August 2002 to March 2020). Provinces are ordered from north (top) to south (bottom) according to the latitude coordinates of their centroid. White boxes indicate missing data.

(TIF)

S2 Fig. Verification metrics of the seasonal climate forecasts evaluated across Vietnam.

The x axis indicates the month of the year. The y axis indicates value of the CRPS for each variable. The lines indicate the lead time for the forecast. CRPS, continuous rank probability score.

(TIF)

S3 Fig. Optimal weights.

Seasonal variation of the optimal weights assigned to each of the models included in the superensemble over the period January 2007 to December 2016.

(TIF)

S4 Fig. Observed vs. predicted dengue cases.

Observed (dashed lines) and predicted (solid lines) dengue cases across Vietnam aggregated at the national level. Shaded areas represent the 95% credible interval. Predictions are shown for the forecast horizons of 1 (top), 3 (middle), and 6 (bottom) months ahead. Data correspond to the period January 2007 to December 2016.

(TIF)

S5 Fig. Prospective dengue predictions.

Predicted dengue cases for the period May to October 2020 for 4 pilot Vietnamese provinces using a model superensemble. The forecast was issued on May 10, 2020. The x axis (top) indicates the month of the predictions. The x axis (bottom) indicates the time lead of the predictions. The y axis indicates the predicted incidence rate. Black solid lines indicate the posterior mean estimate for each of 42 forecast ensemble members. Red lines indicate the observed dengue incidence rate which was not known by the model at the time of the computation and are included here as a reference. The black dashed lines indicate the upper and lower bounds of the 95% credible intervals for the 42 ensemble members. The upper bound of the shaded areas indicates the month- and province-specific percentiles based on dengue data for previous 5 years.

(TIF)

S6 Fig. Month-specific variability in dengue cases across Vietnam.

The x axis indicates the month of the year. The y axis indicates increases in the number of dengue cases (square root transformed). The upper and lower limits of each box represent the interquartile range of the distribution of dengue cases for each month. The middle solid line indicates the median value. The upper and lower whiskers indicate the maximum and minimum values of the dengue case distribution (excluding outliers which are indicated with dark purple circles). Outliers are values beyond 1.5 times the interquartile range.

(TIF)

Acknowledgments

We would like to thank Dr Satoko Otsu and Dr Trang Cong Dai from the World Health Organization Office in Vietnam and Mr Dao Khanh Tung from the United Nations Development Programme Office in Vietnam for the helpful discussions and support for the development of this project. We would also like to acknowledge the support and insights of the Pasteur Institute Ho Chi Minh City, the Pasteur Institute Nha Trang, the Institute of Hygiene and Epidemiology Tay Nguyen (TIHE), and the National Institute of Hygiene and Epidemiology (NIHE).

Abbreviations

ARIMA

autoregressive integrative moving average

CRPS

continuous rank probability score

CRPSS

continuous rank probability skill score

D-MOSS

Dengue forecasting MOdel Satellite-based System

DTR

diurnal temperature range

ECMWF

European Centre for Medium-Range Weather Forecasts

GLMM

generalised linear mixed model

LASSO

least absolute shrinkage and selection operator

MADN

median absolute deviation about the median

MAE

mean absolute error

NIHE

National Institute of Hygiene and Epidemiology

RMSE

root mean squared error

SEDAC

Socioeconomic Data and Applications Center

TIHE

Institute of Hygiene and Epidemiology Tay Nguyen

TSCV

time series cross-validation

WHO

World Health Organization

Data Availability

The data underlying the results presented in the study are available from: Dengue data: Dr Pham Hung, Head of Disease Control Department, Vietnam. Email: hungvncdc@gmail.com Precipitation The precipitation dataset used is a derivative work from NASA GPM and NASA TRRM, https://www.usa.gov/government-works. Wind Speed The wind speed dataset used is a derivative work from NOAA Climate Forecast System Version 2 (CFSv2) Operational Analysis, https://www.usa.gov/government-works. Land Cover The land cover dataset used is a derivative work from Land Cover, © ESA Climate Change Initiative - Land Cover led by UCLouvain (2017). https://www.esa-landcover-cci.org/?q=node/164 Population Population data was used from: WorldPop (www.worldpop.org - School of Geography and Environmental Science, University of Southampton; Department of Geography and Geosciences, University of Louisville; Departement de Geographie, Universite de Namur) and Center for International Earth Science Information Network (CIESIN), Columbia University (2018). Global High Resolution Population Denominators Project - Funded by The Bill and Melinda Gates Foundation (OPP1134076). https://dx.doi.org/10.5258/SOTON/WP00645. Licenced under Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0). GPW Population Count, v4.10 (2000, 2005, 2010, 2015, 2020). Licenced under Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0). UN WPP UN WPP-Adjusted Population Count, v4.10 (2000, 2005, 2010, 2015, 2020). Center for International Earth Science Information Network (CIESIN), Columbia University. 2017. Gridded Population of the World, Version 4 (GPWv4): Population Count Adjusted to Match 2015 Revision of UN WPP Country Totals, Revision 10. Palisades, NY: NASA Socioeconomic Data and Applications Center (SEDAC). https://doi.org/10.7927/H4JQ0XZW. Accessed 16/09/2019. Humidity The humidity dataset used is a derivative work from NASA MODIS/Terra Total Precipitable Water Vapor 5-Min L2 Swath 1km and 5km – NRT, https://www.usa.gov/government-works. Temperature The temperature dataset used is a derivative work from NASA MODIS/Terra Land Surface Temperature/Emissivity Daily L3 Global 1km SIN Grid V006 and MODIS/Aqua Land Surface Temperature/Emissivity Daily L3 Global 1km SIN Grid V006, https://www.usa.gov/government-works.

Funding Statement

This work was funded by the UK Space Agency (https://www.gov.uk/government/organisations/uk-space-agency) Dengue forecasting MOdel Satellite-based System (D-MOSS) awarded to GT, DL, MH, OJB and RL. FJCG, LSB, BH, QH, AH, TC, RA, IF, FM, SJ, SM, and EA were supported by the D-MOSS grant. OJB was supported by a Wellcome Trust Sir Henry Wellcome Fellowship (206471/Z/17/Z). RL was supported by a Royal Society Dorothy Hodgkin Fellowship. The funders had no role in the study design, data collection and analysis, decision to publish, or preparation of the manuscript.

References

  • 1.Powell J, Tabachnick W. History of domestication and spread of Aedes aegypti—a review. Mem Inst Oswaldo Cruz. 2013;108:11–7. 10.1590/0074-0276130395 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 2.Li Y, Kamara F, Zhou G, Puthiyakunnon S, Li C, Zhoi Y, et al. Urbanization increases Aedes albopictus larval habitats and accelerates mosquito development and survivorship. PLoS Negl Trop Dis. 2014;8:e3301. 10.1371/journal.pntd.0003301 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3.Kraemer MU, Sinka ME, Duda KA, Mylne AQ, Shearer FM, Barker CM, et al. The global distribution of the arbovirus vectors Aedes aegypti and Ae. albopictus. Elife. 2015;4:e08347. 10.7554/eLife.08347 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4.Brady O, Gething P, Bhatt S, Messina J, Brownstein J, Hoen A, et al. Refining the Global Spatial Limits of Dengue Virus Transmission by Evidence-Based Consensus. PLoS Negl Trop Dis. 2012;6(8):e1760. 10.1371/journal.pntd.0001760 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5.Bhatt S, Gething P, Brady O, Messina J, Farlow A, Moyes C, et al. The global distribution and burden of dengue. Nature. 2013;496(7446):504–7. 10.1038/nature12060 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6.World Health Organization. Background Paper on dengue vaccines. Geneva, Switzerland: World Health Organization; 2018. Available from: https://www.who.int/immunization/sage/meetings/2018/april/2_DengueBackgrPaper_SAGE_Apr2018.pdf [Google Scholar]
  • 7.Coudeville L, Baurin N, Shepard DS. The potential impact of dengue vaccination with, and without, pre-vaccination screening. Vaccine. 2020;38:1363–9. 10.1016/j.vaccine.2019.12.012 [DOI] [PubMed] [Google Scholar]
  • 8.Dusfour I, Vontas J, David J, Weetman D, Fonseca D, Corbel V, et al. Management of insecticide resistance in the major Aedes vectors of arboviruses: Advances and challenges. PLoS Negl Trop Dis. 2019;13:e0007615. 10.1371/journal.pntd.0007615 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9.Moyes CL, Vontas J, Martins AJ, Ng LC, Koou SY, Dusfour I, et al. Contemporary status of insecticide resistance in the major Aedes vectors of arboviruses infecting humans. PLoS Negl Trop Dis. 2017;11:e0005625. 10.1371/journal.pntd.0005625 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10.Shepard D, Undurraga E, Halasa Y, Stanaway J. The global economic burden of dengue: a systematic analysis. Lancet Inf Dis. 2016;16:935–41. 10.1016/S1473-3099(16)00146-8 [DOI] [PubMed] [Google Scholar]
  • 11.Cuong HQ, Vu NT, Cazelles B, Boni MF, Thai KT, Rabaa MA, et al. Spatiotemporal Dynamics of Dengue Epidemics, Southern Vietnam. Emerg Infect Dis. 2013;19:945–53. 10.3201/eid1906.121323 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12.Burke DS, Nisalak A, Johnson DE, Scott RM. A prospective study of dengue infections in Bangkok. Am J Trop Med Hyg. 1988;38:172–80. 10.4269/ajtmh.1988.38.172 [DOI] [PubMed] [Google Scholar]
  • 13.Shepard D, Undurraga E, Betancourt-Cravioto M, Guzmán MG, Halstead S, Harris E. Approaches to Refining Estimates of Global Burden and Economics of Dengue. PLoS Negl Trop Dis. 2014;8:e3306. 10.1371/journal.pntd.0003306 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14.Hung T, Clapham H, Bettis A, Cuong H, Thwaites G, Wills B, et al. The Estimates of the Health and Economic Burden of Dengue in Vietnam. Trends Parasitol. 2018;34:904–18. 10.1016/j.pt.2018.07.007 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15.Tsunoda T, Cuong T, Dong T, Yen N, Le N, Phong T, et al. Winter Refuge for Aedes aegypti and Ae. albopictus Mosquitoes in Hanoi during Winter. PLoS ONE. 2014;9:e95606. 10.1371/journal.pone.0095606 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16.Thi KLP, Briant L, Gavotte L, Labbe P, Perriat-Sanguinet M, Cornillot E, et al. Incidence of dengue and chikungunya viruses in mosquitoes and human patients in border provinces of Vietnam. Parasit Vectors. 2017;10:556. 10.1186/s13071-017-2422-z [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17.Bastos LS, Economou T, Gomes MFC, Villela DAM, Coelho FC, Cruz OG, et al. A modelling approach for correcting reporting delays in disease surveillance data. Stat Med. 2019;38:4363–77. 10.1002/sim.8303 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18.Lowe R, Coelho C, Barcellos C, Sá Carvalho M, De Castro Catão R, Coelho G, et al. Evaluating probabilistic dengue risk forecasts from a prototype early warning system for Brazil. Elife. 2016;5:e11285. 10.7554/eLife.11285 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19.Lowe R, Stewart-Ibarra AM, Petrova D, García-Díez M, Borbor-Cordova MJ, Mejía R, et al. Climate services for health: predicting the evolution of the 2016 dengue season in Machala, Ecuador. Lancet Plan Health. 2017;1:142–51. 10.1016/S2542-5196(17)30064-5 [DOI] [PubMed] [Google Scholar]
  • 20.Colón-González FJ, Harris I, Osborn T, Steiner São Bernardo C, Peres C, Hunter P, et al. Limiting global-mean temperature increase to 1.5–2°C could reduce the incidence and spatial spread of dengue fever in Latin America. Proc Natl Acad Sci U S A. 2018;115:6243–6248. 10.1073/pnas.1718945115 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21.Gage K, Burkot T, Eisen R, Hayes E. Dengue disease outbreak definitions are implicitly variable. Am J Prev Med. 2008;25:436–50. [Google Scholar]
  • 22.Reinhold J, Lazzari R, Lahondére C. Effects of the Environmental Temperature on Aedes aegypti and Aedes albopictus Mosquitoes: A Review. Insects. 2018;8:158. 10.3390/insects9040158 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 23.Kraemer M, Reiner JRC, Brady O, Messina J, Gilbert M, Pigott D, et al. Past and future spread of the arbovirus vectors Aedes aegypti and Aedes albopictus. Nat Microbiol. 2019;4(5):854–63. 10.1038/s41564-019-0376-y [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 24.Watts D, Burke D, Harrison B, Whitmire R, Nisalak A. Effect of temperature on the vector efficiency of Aedes aegypti for dengue 2 virus. Amer J of Trop Med Hyg. 1987;36(1):143–52. 10.4269/ajtmh.1987.36.143 [DOI] [PubMed] [Google Scholar]
  • 25.Mordecai EA, Cohen JM, Evans MV, Gudapati P, Johnson LR, Lippi CA, et al. Detecting the impact of temperature on transmission of Zika, dengue and chikungunya using mechanistic models. PLoS Negl Trop Dis. 2017;11:e0005568. 10.1371/journal.pntd.0005568 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 26.Lambrechts L, Paaijmans K, Fansiri T, Carrington L, Kramer L, Thomas M, et al. Impact of daily temperature fluctuations on dengue virus transmission by Aedes aegypti. Proc Natl Acad Sci U S A. 2011;108:7460–5. 10.1073/pnas.1101377108 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27.Juliano SA, OMeara GF, Morrill JR, Cutwa MM. Desiccation and thermal tolerance of eggs and the coexistence of competing mosquitoes. Oecologia. 2002;130(3):458–69. 10.1007/s004420100811 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 28.Canyon D, Muller R, Hii JL. Aedes aegypti disregard humidity-related conditions with adequate nutrition. Trop Biomed. 2013;30(1):1–8. [PubMed] [Google Scholar]
  • 29.Stewart Ibarra A, Ryan S, Beltrán E, Mejía R, Silva M, Muñoz A. Dengue Vector Dynamics (Aedes aegypti) Influenced by Climate and Social Factors in Ecuador: Implications for Targeted Control. PLoS ONE. 2013;8:e78263. 10.1371/journal.pone.0078263 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 30.Sedda L, Vilela A, Aguiar E, Gaspar C, Gonçalves A, Olmo R, et al. The spatial and temporal scales of local dengue virus transmission in natural settings: A retrospective analysis. Parasit Vectors. 2018;11(1). 10.1186/s13071-018-2662-6 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 31.Naish S, Dale P, Mackenzie JS, McBride J, Mengersen K, Tong S. Climate change and dengue: a critical and systematic review of quantitative modelling approaches. BMC Infect Dis. 2014;14:167. 10.1186/1471-2334-14-167 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 32.Lauer S, Sakrejda K, Ray E, Keegan L, Bi Q, Suangtho P, et al. Prospective forecasts of annual dengue hemorrhagic fever incidence in Thailand, 2010–2014. Proc Natl Acad Sci U S A. 2018;115:E2175–82. 10.1073/pnas.1714457115 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 33.Lowe R, Gasparrini A, Van Meerbeeck C, Lippi C, Mahon R, Trotman A, et al. Nonlinear and delayed impacts of climate on dengue risk in Barbados: A modelling study. PLoS Med. 2018;15:e1002613. 10.1371/journal.pmed.1002613 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 34.Lowe R, Barcellos C, Coelho CA, Bailey TC, Coelho GE, Graham R, et al. Dengue outlook for the World Cup in Brazil: an early warning model framework driven by real-time seasonal climate forecasts. Lancet Infect Dis. 2014;14(7):619–26. 10.1016/S1473-3099(14)70781-9 [DOI] [PubMed] [Google Scholar]
  • 35.Tompkins A, Di Giuseppe F. Potential Predictability of Malaria in Africa Using ECMWF Monthly and Seasonal Climate Forecasts. J Appl Meteor Climatol. 2015;54:521–540. [Google Scholar]
  • 36.Tompkins A, Colón-González F, Di Giuseppe F, Namanya D. Dynamical Malaria Forecasts Are Skillful at Regional and Local Scales in Uganda up to 4 Months Ahead. Geohealth. 2019;3:58–66. 10.1029/2018GH000157 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 37.Thomson MC, Mason S, Platzer B, Mihretie A, Omumbo J, Mantilla G, et al. Climate and health in Africa. Earth Perspectives. 2014;1:17. [Google Scholar]
  • 38.Yamana T, Kandula S, Shaman J. Superensemble forecasts of dengue outbreaks. J R Soc Interface. 2016;13:20160410. 10.1098/rsif.2016.0410 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 39.Eastin M, Delmelle E, Casas I, Wexler J, Intra- CS. interseasonal autoregressive prediction of dengue outbreaks using local weather and regional climate for a tropical environment in Colombia. Am J Trop Med Hyg. 2014;91:598–610. 10.4269/ajtmh.13-0303 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 40.Hii Y, Zhu H, Ng N, Ng L, Rocklöv J. Forecast of dengue incidence using temperature and rainfall. PLoS Negl Trop Dis. 2013;6:e1908. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 41.Shi Y, Liu X, Kok SY, Rajarethinam J, Liang S, Yap G, et al. Three-Month Real-Time Dengue Forecast Models: An Early Warning System for Outbreak Alerts and Policy Decision Support in Singapore. Environ Health Perspect. 2016;124:1369–75. 10.1289/ehp.1509981 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 42.Chen Y, Ong JHY, Rajarethinam J, Yap G, Ng LC, Cook AR. Neighbourhood level real-time forecasting of dengue cases in tropical urban Singapore. BMC Med. 2018;16:129. 10.1186/s12916-018-1108-5 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 43.Stolerman LM, Maia PD, Kutz JN. Forecasting dengue fever in Brazil: An assessment of climate conditions. PLoS ONE. 2019;14:e0220106. 10.1371/journal.pone.0220106 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 44.Bergmeir C, Benítez J. On the use of Cross-Validation for Time Series Predictor Evaluation. Inform Sci. 2012;191:192–213. [Google Scholar]
  • 45.Johansson MA, Apfeldorf KM, Dobson S, Devita J, Buczak AL, Baugher B, et al. An open challenge to advance probabilistic forecasting for dengue epidemics. Proc Natl Acad Sci U S A. 2019;116(48):24268–74. 10.1073/pnas.1909865116 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 46.Colón-González FJ, Lake IR, Morbey RA, Elliot AJ, Pebody R, Smith GEA. Methodological Framework for the Evaluation of Syndromic Surveillance Systems: A Case Study of England. BMC Public Health. 2018;18:544. 10.1186/s12889-018-5422-9 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 47.Lake IR, Colón-González FJ, Barker G, Morbey RA, Smith GE, Elliot AJ. Machine learning to refine decision making within a syndromic surveillance service. BMC Public Health. 2019;19:559. 10.1186/s12889-019-6916-9 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 48.Badurdeen S, Valladares D, Farrar J, Gozzer E, Kroeger A, Kuswara N, et al. Sharing experiences: towards an evidence based model of dengue surveillance and outbreak response in Latin America and Asia. BMC Public Health. 2013;13:607. 10.1186/1471-2458-13-607 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 49.Runge Ranzinger S, McCall P, Kroeger A, Horstick O. Dengue disease surveillance: an updated systematic literature review. Trop Med Int Health. 2014;19:1116–60. 10.1111/tmi.12333 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 50.Brady O, Smith D, Scott, Hay S. Dengue disease outbreak definitions are implicitly variable. Epidemics. 2015;11:91–102. 10.1016/j.epidem.2015.03.002 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 51.Noufaily A, Morbey RA, Colón-González FJ, Elliot AJ, Smith GE, et al. Comparison of statistical algorithms for daily syndromic surveillance aberration detection. Bioinformatics. 2019;17:3110–8. 10.1093/bioinformatics/bty997 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 52.Hussain-Alkhateeb L, Kroeger A, Olliaro P, Rocklöv J, Sewe MO, Tejeda G, et al. Early warning and response system (EWARS) for dengue outbreaks: Recent advancements towards widespread applications in critical settings. PLoS ONE. 2018;13:e0196811. 10.1371/journal.pone.0196811 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 53.Olliaro P, Fouque F, Kroeger A, Bowman L, Velayudhan R, Santelli AC, et al. Improved tools and strategies for the prevention and control of arboviral diseases: A research-to-policy forum. PLoS Negl Trop Dis. 2018;12:e0005967. 10.1371/journal.pntd.0005967 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 54.Hyndman R, Athanasopoulos G. Forecasting: Principles and Practice; 2014. Available from: https://www.otexts.org/book/fpp.
  • 55.CIESIN. Gridded Population of the World (GPW), v4; 2019. Available from: https://sedac.ciesin.columbia.edu/data/collection/gpw-v4.
  • 56.ESA. Land cover CCI Product User Guide Version; 2017. Available from: http://maps.elie.ucl.ac.be/CCI/viewer/download/ESACCI-LC-Ph2-PUGv2_2.0.pdf.
  • 57.Copernicus. Copernicus Climate Data Store; 2019. Available from: https://cds.climate.copernicus.eu/!/home.
  • 58.Wan Z, Hook S, Hulley G. MYD11A1 MODIS/Aqua Land Surface Temperature/Emissivity Daily L3 Global 1km SIN Grid V006; 2015. Available from: https://lpdaac.usgs.gov/products/myd11a1v006/.
  • 59.Hijmans R. TRMM (TMPA) Rainfall Estimate L3 3 hour 0.25 degree x 0.25 degree V7; 2011. Available from: https://disc.gsfc.nasa.gov/datasets/TRMM_3B42_7/summary.
  • 60.Huffman GJ, Stocker EF, Bolvin DT, Nelkin EJ, Tan J. GPM IMERG Final Precipitation L3 Half Hourly 0.1 degree x 0.1 degree V06; 2019. Available from: https://catalog.data.gov/dataset/gpm-imerg-final-precipitation-l3-half-hourly-0-1-degree-x-0-1-degree-v06-gpm-3imerghh-at-g.
  • 61.Gao BC. MODIS Atmosphere L2 Water Vapor Product; 2015. Available from: https://atmosphere-imager.gsfc.nasa.gov/MOD05-L2/index.html.
  • 62.The NCEP Climate Forecast System Version 2. Saha S, Moorthi S, Wu X, Wang J, Nadiga S, Tripp P, et al. The NCEP Climate Forecast System Version 2. J Climate. 2014;27:2185–220. [Google Scholar]
  • 63.NOAA. Monthly Atmospheric and SST Indices; 2020. Available from: https://www.cpc.ncep.noaa.gov/data/indices/sstoi.indices.
  • 64.Van Rossum G, Drake FL Jr. Python reference manual. Centrum voor Wiskunde en Informatica Amsterdam; 1995. [Google Scholar]
  • 65.WorldPop. The spatial distribution of population in Vietnam; 2018. Available from: https://cds.climate.copernicus.eu/!/home.
  • 66.MacLachlan C, Arribas A, Peterson KA, Maidens A, Fereday D, Scaife AA, et al. Description of GloSea5: the Met Office high resolution seasonal forecast system. Q J Roy Meteorol Soc. 2015;141:1072–01084. [Google Scholar]
  • 67.Scaife A, Arribas A, Blockley E, Brookshaw A, Clark RT, Dunstone N, et al. Description of GloSea5: the Met Office high resolution seasonal forecast system. Geophys Res Lett. 2014;41:2514–9. [Google Scholar]
  • 68.Bowler NE, Arribas A, Beare SE, Mylne KR, Shutts GJ. The local ETKF and SKEB: upgrades to the MOGREPS short-range ensemble prediction system. Q J Roy Meteorol Soc. 2009;41:767–76. [Google Scholar]
  • 69.Besag J, York J, Mollie A. Bayesian image restoration, with two applications in spatial statistics. An I Stat Math. 1991;43:1–20. [Google Scholar]
  • 70.Petrova D, Lowe R, Stewart-Ibarra A, Ballester J, Koopman SJ, Rodó X. Sensitivity of large dengue epidemics in Ecuador to long-lead predictions of El Niño. Climate Services. 2019;15:100096. [Google Scholar]
  • 71.Cheong YL, Burkart K, Leitao PJ, Lakes T1. Assessing Weather Effects on Dengue Disease in Malaysia. Int J Environ Res Public Health. 2013;10:6319–6334. 10.3390/ijerph10126319 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 72.Simpson D, Rue H, Riebler A, Martins TG, Sørbye SH, et al. Penalising model component complexity: A principled, practical approach to constructing priors. Stat Sci. 2017;32(1):1–28. [Google Scholar]
  • 73.Rue H, Martino S, Chopin N. Approximate Bayesian Inference for Latent Gaussian Models Using Integrated Nested Laplace Approximations (with discussion). J R Stat Soc Series B Stat Methodology. 2009;71:319–92. [Google Scholar]
  • 74.Siegert S. SpecsVerification: Forecast Verification Routines for Ensemble Forecasts of Weather and Climate; 2017. Available from: https://CRAN.R-project.org/package=SpecsVerification.
  • 75.Funk S, Camacho A, Kucharski AJ, Lowe R, Eggo RM, Edmunds WJ. Assessing the performance of real-time epidemic forecasts: A case study of Ebola in the Western Area region of Sierra Leone. PLoS Comput Biol. 2019;15(2):e1006785. 10.1371/journal.pcbi.1006785 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 76.Bosse N, Sam Abbott, Funk S. scoringutils: Utilities for Scoring and Assessing Predictions; 2020. Available from: https://CRAN.R-project.org/package=scoringutils.
  • 77.Imai C, Armstrog B, Chalabi Z, Mangtani P, Hashizume M. Time series regression model for infectious disease and weather. Environ Res. 2015;142:319–27. 10.1016/j.envres.2015.06.040 [DOI] [PubMed] [Google Scholar]
  • 78.Reich NG, Brook LC, Fox SJ, Kandula S, McGowan CJ, Moore E, et al. A collaborative multiyear, multimodel assessment of seasonal influenza forecasting in the United States. Proc Natl Acad Sci U S A. 2019;116(8):3146–54. 10.1073/pnas.1812594116 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 79.Yao Y, Vehtari A, Simpson D, Gelman A. Using Stacking to Average Bayesian Predictive Distributions. Bayesian Anal. 2018;13(3):917–1003. [Google Scholar]
  • 80.Bates JM, Granger CWJ. The Combination of Forecasts. Oper Res Q. 1969;20(4):451–68. [Google Scholar]
  • 81.Kang H. Unstable Weights in the Combination of Forecasts. Manag Sci. 1986;32(6):683–95. [Google Scholar]
  • 82.Brier GW. Verification of forecasts expressed in terms of probability. Mon Weather Rev. 1950;78(1):1–3. [Google Scholar]
  • 83.Merkle EC, Steyvers M. Choosing a Strictly Proper Scoring Rule. Decis Anal. 2013;10:292–304. [Google Scholar]
  • 84.Thornes JE, Stephenson DB. How to judge the quality and value of weather forecast products. Meteorol Appl. 2001;8:307–14. [Google Scholar]
  • 85.Richardson DS. Skill and relative economic value of the ECMWF ensemble prediction system. Q J Roy Meteorol Soc. 2000;126(563):649–67. [Google Scholar]
  • 86.Krishnamurti TN, Kumar V, Bhardwaj ASA, Ghosh T, Ross R. A review of multimodel superensemble forecasting for weather, seasonal climate, and hurricanes. Rev Geophys. 2016;54(2):336–77. [Google Scholar]
  • 87.Krishnamurti TN, Kishtawal CM, Zhang Z, LaRow T, Bachiochi D, Williford E, et al. Improved weather and seasonal climate forecasts from multimodel superensemble. Science. 1999;285:1548–1550. 10.1126/science.285.5433.1548 [DOI] [PubMed] [Google Scholar]
  • 88.Yamana T, Kandula S, Shaman J. Individual versus superensemble forecasts of seasonal influenza outbreaks in the United States. PLoS Comput Biol. 2017;13(11):e1005801. 10.1371/journal.pcbi.1005801 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 89.Held L, Meyer S, Bracher J. Probabilistic forecasting in infectious disease epidemiology: the 13th Armitage lecture. Stat Med. 2017;36(22):3443–60. 10.1002/sim.7363 [DOI] [PubMed] [Google Scholar]
  • 90.Reich NG, Lauer SA, Sakrejda K, Iamsirithaworn S, Hinjoy S, Suangtho P, et al. Challenges in Real-Time Prediction of Infectious Disease: A Case Study of Dengue in Thailand. PLoS Negl Trop Dis. 2016;10(6):e0004761. 10.1371/journal.pntd.0004761 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 91.Reiter P. Climate change and mosquito-borne disease. Environ Health Perspect. 2001;109:141–61. 10.1289/ehp.01109s1141 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 92.Bracher J, Held L Endemic-epidemic models with discrete-time serial interval distributions for infectious disease prediction. arXiv:1901.03090v3 [Preprint]. 2019. [posted 2019 Jan 10; revised 2019 Oct 4; revised 2020 Mar 13]. Available from: https://arxiv.org/abs/1901.03090.

Decision Letter 0

Artur Arikainen

29 Jun 2020

Dear Dr Colón-González,

Thank you for submitting your manuscript entitled "Probabilistic seasonal dengue forecasting in Vietnam using superensembles" for consideration by PLOS Medicine.

Your manuscript has now been evaluated by the PLOS Medicine editorial staff and I am writing to let you know that we would like to send your submission out for external peer review.

However, before we can send your manuscript to reviewers, we need you to complete your submission by providing the metadata that is required for full assessment. To this end, please login to Editorial Manager where you will find the paper in the 'Submissions Needing Revisions' folder on your homepage. Please click 'Revise Submission' from the Action Links and complete all additional questions in the submission questionnaire.

Please re-submit your manuscript within two working days, i.e. by .

Login to Editorial Manager here: https://www.editorialmanager.com/pmedicine

Once your full submission is complete, your paper will undergo a series of checks in preparation for peer review. Once your manuscript has passed all checks it will be sent out for review.

Feel free to email us at plosmedicine@plos.org if you have any queries relating to your submission.

Kind regards,

Artur Arikainen,

Associate Editor

PLOS Medicine

Decision Letter 1

Emma Veitch

5 Aug 2020

Dear Dr. Colón-González,

Thank you very much for submitting your manuscript "Probabilistic seasonal dengue forecasting in Vietnam using superensembles" (PMEDICINE-D-20-02989R1) for consideration at PLOS Medicine.

Your paper was evaluated by a senior editor and discussed among all the editors here. It was also evaluated by five independent reviewers, whose comments are enclosed below; I hope you find the reviews constructive. Any accompanying reviewer attachments can be seen via the link below:

[LINK]

In light of these reviews, I am afraid that we will not be able to accept the manuscript for publication in the journal in its current form, but we would like to consider a revised version that addresses the reviewers' and editors' comments. Obviously we cannot make any decision about publication until we have seen the revised manuscript and your response, and we plan to seek re-review by one or more of the reviewers.

In revising the manuscript for further consideration, your revisions should address the specific points made by each reviewer and the editors. Please also check the guidelines for revised papers at http://journals.plos.org/plosmedicine/s/revising-your-manuscript for any that apply to your paper. In your rebuttal letter you should indicate your response to the reviewers' and editors' comments, the changes you have made in the manuscript, and include either an excerpt of the revised text or the location (eg: page and line number) where each change can be found. Please submit a clean version of the paper as the main article file; a version with changes marked should be uploaded as a marked up manuscript.

In addition, we request that you upload any figures associated with your paper as individual TIF or EPS files with 300dpi resolution at resubmission; please read our figure guidelines for more information on our requirements: http://journals.plos.org/plosmedicine/s/figures. While revising your submission, please upload your figure files to the PACE digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email us at PLOSMedicine@plos.org.

We expect to receive your revised manuscript by Aug 26 2020 11:59PM. Please email us (plosmedicine@plos.org) if you have any questions or concerns.

***Please note while forming your response, if your article is accepted, you may have the opportunity to make the peer review history publicly available. The record will include editor decision letters (with reviews) and your responses to reviewer comments. If eligible, we will contact you to opt in or out.***

We ask every co-author listed on the manuscript to fill in a contributing author statement, making sure to declare all competing interests. If any of the co-authors have not filled in the statement, we will remind them to do so when the paper is revised. If all statements are not completed in a timely fashion this could hold up the re-review process. If new competing interests are declared later in the revision process, this may also hold up the submission. Should there be a problem getting one of your co-authors to fill in a statement we will be in contact. YOU MUST NOT ADD OR REMOVE AUTHORS UNLESS YOU HAVE ALERTED THE EDITOR HANDLING THE MANUSCRIPT TO THE CHANGE AND THEY SPECIFICALLY HAVE AGREED TO IT. You can see our competing interests policy here: http://journals.plos.org/plosmedicine/s/competing-interests.

Please use the following link to submit the revised manuscript:

https://www.editorialmanager.com/pmedicine/

Your article can be found in the "Submissions Needing Revision" folder.

To enhance the reproducibility of your results, we recommend that you deposit your laboratory protocols in protocols.io, where a protocol can be assigned its own identifier (DOI) such that it can be cited independently in the future. For instructions see http://journals.plos.org/plosmedicine/s/submission-guidelines#loc-methods.

Please ensure that the paper adheres to the PLOS Data Availability Policy (see http://journals.plos.org/plosmedicine/s/data-availability), which requires that all data underlying the study's findings be provided in a repository or as Supporting Information. For data residing with a third party, authors are required to provide instructions with contact information for obtaining the data. PLOS journals do not allow statements supported by "data not shown" or "unpublished results." For such statements, authors must provide supporting data or cite public sources that include it.

We look forward to receiving your revised manuscript.

Sincerely,

Emma Veitch, PhD

PLOS Medicine

On behalf of Clare Stone, PhD, Acting Chief Editor,

PLOS Medicine

plosmedicine.org

-----------------------------------------------------------

Requests from the editors:

*Please revise your title according to PLOS Medicine's style - this should summarise the study question but also state (preferably after colon in the subtitle) the study design (eg "A randomized controlled trial," "A retrospective study," "A modelling study," etc.).

*In the last sentence of the Abstract Methods and Findings section, please add a brief summary of any key limitation(s) of the study's methodology.

*At this stage, we ask that you include a short, non-technical Author Summary of your research to make findings accessible to a wide audience that includes both scientists and non-scientists. The Author Summary should immediately follow the Abstract in your revised manuscript. This text is subject to editorial change and should be distinct from the scientific abstract. Please see our author guidelines for more information: https://journals.plos.org/plosmedicine/s/revising-your-manuscript#loc-author-summary

*Please clarify whether the analytical approach detailed here was laid out in a prospective protocol or analysis plan? Please state this (either way) early in the Methods section.

a) If a prospective analysis plan (from your funding proposal, IRB or other ethics committee submission, study protocol, or other planning document written before analyzing the data) was used in designing the study, please include the relevant prospectively written document with your revised manuscript as a Supporting Information file to be published alongside your study, and cite it in the Methods section. A legend for this file should be included at the end of your manuscript.

b) If no such document exists, please make sure that the Methods section transparently describes when analyses were planned, and when/why any data-driven changes to analyses took place.

c) In either case, changes in the analysis-- including those made in response to peer review comments-- should be identified as such in the Methods section of the paper, with rationale.

-----------------------------------------------------------

Comments from the reviewers:

Reviewer #1: Dengue is a climate-sensitive disease, and many efforts have been attempted to predict future outbreak risk using environmental conditions and past patterns of dengue. However, it is relatively scarce to combine forecasts from multiple competing models into a superensemble to make more accurate predictions. This study aimed to develop a superensemble of probabilistic hierarchical dengue models to predict dengue risk up to six months ahead at the province level across Vietnam. It is an interesting and well-presented manuscript but several issues should be addressed to improve its clarity.

How were suspected and confirmed dengue cases diagnosed and reported? More information on each category is clearly needed. The quality and consistency of the dengue data was likely to vary with area and time, and it remains largely unknown how this problem might affect the construction of the model. Data on both disease and environmental conditions (e.g. meteorological, demographic and land-cover variables) had different spatial resolutions and it is unclear how these datasets were matched. Is there any change in the administrative boundary across 63 provinces in Vietnam during the period 2002-2019? How were population weighted monthly averages calculated for each climate variable? Why were verification metrics inconsistent (e.g. CRPS, RMSE and MAE were small but DIC and WAIC were big in Table 1)? Several explanatory variables (e.g. minimum temperature, maximum temperature, and sea surface temperature anomalies) were likely to be highly correlated (L315-319), and how was this issue resolved? Combining all best performing models into a model superensemble led to a lower CRPS and MAE than any of the five competing models, but the RMSE was an exception as the RMSE of the superensemble was higher than two of the five models (L335-337). Why is it the case? Fig 8 shows a range of cost-loss ratios and outbreak thresholds for 42 Vietnamese provinces. How about other 21 provinces as Vietnam has 63 provinces? It is important to demonstrate that their spatio-temporal superensemble could guide changes to the current practice (L479-84). However, no results were presented on this aspect.

-----------------------------------------------------------

Reviewer #2: I have mixed feelings about this paper. On the one hand, the work is really quite impressive, with a diverse set of top-notch datasets and cutting-edge methods. I found myself nodding away enthusiastically as I read the methods section and text of the results. On the other hand, though, the actual results are not very good. The superensemble of forecasts gives effectively the same performance as the better models do individually, as can be seen from table 1 but more clearly from supplementary figure 2. The text therefore feels misleading: yes, the superensemble may be technically better but effectively it was the same and I would have wanted to have seen a more balanced description of that, because this is the sort of project that other countries should be copying and they need the right information to evaluate how complex to aim for. The other reason I say the results are not very good is that the forecast at six months is really poor. Figure S6 illustrates this clearly: in March, the authors are trying to forecast the situation in August, and for HCMC can't even tell if the number of cases will be a few hundred or 20 000. If there is this level of uncertainty, the more honest thing to do would be to report that the forecast accuracy is intolerable at that forecast window and instead limit the forecast to say one to two months into the future. I thought this paper was great until I went to the supplement and had the chance to see directly what the forecasts looked like: they are really not presented transparently within the ms itself.

My second major concern is about the cost analysis, which is poorly handled and lowers the overall quality of the manuscript. I don't think the equation is correct (surely the optimal decision depends not just on the ratio of cost under action 1 and benefit under action 2 but also the impact of the two actions on the outcome and the costs of action 2 and benefits of action 1), and the statement about always or never preparing for outbreaks being the default raises alarm bells. The decision making process about interventions for outbreaks of dengue is waaaaaay more complex than the authors are making it out to be and the analysis here really comes across as the view from an ivory tower. If the authors really want to do a cost-benefit analysis, perhaps because the grant promised it, then it would behove them to do it properly. You have some really great people like Mark Jit at LSHTM, why not tap on them to see how to do a proper CBA? (Probably as a separate paper though.)

My final major criticism is on the structure of the forecast model/s themselves. Why are no lags considered for the predictors? Some such as El Nino would expect to see their presence felt a long time after. Even local features could plausibly have an impact longer than one month in the future because of the long latency caused by the cycles of breeding, infection and both incubation periods. You should be able to quickly refute this. Furthermore, why are the data at such a crude temporal resolution? If the VN MOH has provided an author, can s/he not provide weekly (or even daily! That would be fab!) data, because that would presumably greatly improve the predictive potential.

A few minor comments that may help revise the manuscript:

Abstract: "the skill of a system using different lead times in a year-round prediction system has not been previously explored" I'm not sure that is strictly true. For instance, Singapore has been doing something like this for a number of years

L13: \\textit{Ae.~aegypti} to prevent inter-sentence spacing. You can similarly use this trick after i.e.~ and so on

L10: International standard is to use a space to demarcate thousands from hundreds to avoid ambiguity about decimal point vs decimal comma

L24-32: A bit of repetitive sentence structuring, consider tweaking.

L39-48: These sentences were a bit hard to read. Consider tweaking with some extra words at the start of the sentences like "It is also known that", "In contrast," or changing the pattern a bit like "Biting rates are also affected by xxx" etc, to help with the flow

L42: bad latex: $>$ instead of > and either \\degree, \\textdegree (depending on packages) or $^{\\circ}$

L71: What is fundamentally wrong with testing a model on year 2015 after training it on 2016 (say)? Does it not still give a valid approximation of out of sample performance?

L81: Good point, perhaps worth elaborating on the difficulties of implementing such a system not in a research environment

L102: $<$ not <

L129: Possible to show time series of the data for a few provinces?

L142: Perhaps my physics is letting me down but isn't kg kg^{-1} unitless?

L152-160: Do I read correctly that you used two different datasets to create population data? How were these harmonised? If not, please clarify and remedy my confusion!

Eqn1: Should Y be Y_{i,t-1}? Should the sum be enclosed in parenthesis to prevent ambiguity about the random effects? What happens when Y=0?

L205: missing closing parenthesis

Eqn 5: \\exp \\max

T1: Interesting that so little benefit accrues from the superensemble compared to simpler models. The text hints that the superensemble is substantially better

L367: Some of this subsection should perhaps go in the methods section?

L407/8: missing spaces

L416: The proposed decision support tool is… a bit weak, isn't it? Aesthetically it's ugly, but the uncertainty in the predictions seemingly makes the forecast useless.

T2: Why is Moderate defined as being up to the 75th percentile AND mean+1SD? Mean+1SD should not be the 75th percentile.

L515: What is 95 per cent percentile? Is it the same as the 95th percentile?

F1: Not very informative. Can I propose you omit the islands (and just add a comment in the legend that islands in the East Sea are omitted for space). Then you can stack the maps in a 6x2 arrangement, making them all bigger and reducing the white space between them. At present you can hardly make out the changes in colour around HCMC and Vinh.

-----------------------------------------------------------

Reviewer #3: # Overview

In this manuscript the authors present and evaluate ensemble forecasts of

monthly dengue case numbers in each province of Vietnam for 2002-2019, which

incorporate a variety of demographic, land-cover, and climate data. These

forecasts are evaluated for lead times of 1-6 months --- looking much further

ahead than many existing dengue forecasting studies. The evaluation includes a

cost-loss analysis of whether action taken in response to the projected outbreak

probability would be beneficial. I found this to be an interesting and

well-designed study, and the prospect of further evaluation of this forecasting

system is exciting.

# Major comments

1. Regarding the "Modelling approach" section on page 6, I understand that

there were 7 seasonal climate data sets (mean monthly values were used):

(a) Minimum temperature;

(b) Maximum temperature;

(c) Daily precipitation;

(d) Specific humidity;

(e) Diurnal temperature range;

(f) Wind speed; and

(g) Sea surface temperature anomalies for the Niño 3.4 Region.

So 128 different models were considered to cover all possible combinations of

these 7 data sets. But this only became apparent after reading the first few

sentences of the results section, where this is explained to the reader. I

would have found it quite helpful to instead include these remarks in the

"Modelling approach" section.

2. "Model selection" section (page 6): "The testing set comprised the climate

hindcast data for the six months immediately after the last observation in

the training set for each geographical area."

I apologize if I've missed an obvious point here, but why were hindcasts used

instead of actual data for this period? Is the intent that in an operational

system they would be replaced by climate forecasts from the same source

(Copernicus Climate Data Store)?

3. "Accounting for autocorrelation in disease transmission" section (page 7):

"To generate dengue forecasts one month ahead, log(Y_i, t−1) for time t + 1

corresponds to the logarithm of the number of dengue cases at time t. To

generate the forecasts two months ahead we first fitted a climate naïve model

..."

It isn't clear to my why a climate-naive model was used here. Since the

testing set included climate hindcast data for the six months after the last

observation (see previous comment) then why couldn't the climate-informed

model be used to predict the number of dengue cases for time t + 1? And at

the start of the Results section (page 9), the authors write:

"Predictions up to six months ahead were made at each time step starting in

January 2007 using seasonal climate hindcasts (retrospective forecasts) data

..."

4. "Model superensemble" section (page 8): "A new set of superensemble weights

were computed each time a forecast was issued using training data from all

previous years."

Can the authors comment on how these weights evolved over time? Was the same

model consistently the highest-weighted over all of the forecast months?

Presumably in an operation system it would be highly desirable if the model

weights were stable from one month to the next? I suppose this could be

extended to allowing seasonal variation (i.e., month-to-month) if different

models are understood to perform best in different calendar months. It would

be great to see, e.g., a supplementary figure that shows how the weights

evolved.

4. "Model superensemble" section (page 8): "In addition, we evaluated the skill

of the obtained forecasts using the continuous rank probability skill score

(CRPSS). CRPSS is defined as: ..."

It appears that there should be further text after equation (6), since it

ends with a trailing comma, but instead the section ends. Is there missing

text, or should the comma be removed?

5. Results section (page 9): how did the authors decide to include precisely 5

models in the superensemble? Was there a clear divide in performance between

the top 5 models and the remaining 123 models? And were the authors surprised

that none of the models in the superensemble included daily precipitation as

a seasonal climate predictor?

6. Results section (page 10): "The predictive ability of the [superensemble]

model also varied with the month of the year (Figure 1). Overall, larger

discrepancies between observed and predicted values were observed between

July and December, when typically more cases are reported."

Did the authors consider evaluating the models using any relative error

measures, in addition to the chosen absolute error measures? This is an

important consideration for infectious diseases epidemic forecasts since an

error of, say, 10 cases has very different significance when the expected

number is 20 cases or 20,000 cases. As the authors note:

"From August to November the skill is reduced for selected provinces in the

central and southern regions characterized by larger dengue incidence

variability."

This might be primarily influenced by the choice of absolute error measures

rather than being intrinsic to the models themselves.

7. Results section (page 10): "S5 Fig. shows the observed and posterior

predictive mean dengue cases across the 63 provinces computed one month ahead

using the model superensemble. The model superensemble is able to reproduce

the spatiotemporal dynamics of dengue fever with reasonable skill although

predictions tend to underestimate the number of cases particularly during

large outbreaks."

This observation made me wonder whether the model fitting may not have

selected appropriate values for the dispersion parameters.

The authors note (top of page 6) that models were fitted "using a negative

binomial specification to account for potential over-dispersion in the data"

and that "penalising complexity priors were assumed for [...] the dispersion

parameters".

What prior distribution was used for the over-dispersion parameters, and what

were the posterior distributions when the models were fitted? I couldn't see

this mentioned in the manuscript text, nor am I able to identify a location

in the provided code or supplementary materials where this is defined. In the

provided file "04_Fit_models.R" there is the following code:

```

mu1 <- lead6$fitted.mean

E1 <- (lead6$dengue_cases - mu1) / sqrt(mu1)

N <- nrow(lead6)

p <- 6 # Max number coef

dispersion <- sum(E1^2, na.rm=TRUE) / (N - p)

dispersion <- data.table(date=j, model="ensmeble", disper=dispersion)

```

But this table doesn't seem to be used anywhere in any of the provided

scripts, so I gather this it is reporting an output statistic.

8. Discussion, page 15: "Relative to a baseline seasonal predictive model (see

Methods and Materials), the superensemble made, on average, more accurate

predictions across most provinces, and for most of the year in agreement with

previous studies [16, 39]. This observation, however, conflicts with the

results obtained by [30, 42] for whom climate-naïve models had better skill

than seasonal-climate-informed ones."

Are there any obvious reasons for this apparent disagreement? Did the

studies in refs 30 and 42 make use of different data sources, different

evaluation metrics, etc? And can the authors make any general remarks about

whether the superensemble forecasts outperformed the climate-naive models in

refs 30 and 42? I appreciate there are many reasons why a direct comparison

may not be feasible, and if that's the case then could the authors please

make this clear in the text so that reader isn't left wondering?

9. Discussion, page 15: "Third, our computations of dengue risk do not take into

consideration uncertainties due to the potential under- or misreporting of

dengue cases. Consequently, our model superensemble forecasts may

underestimate the real number of cases occurring at any given time."

That's a good point to make. Do the authors have any knowledge about whether

case ascertainment is likely to have changed substantially over the period

covered by this study (2002-2019)?

# Minor comments

1. Top of page 3, there is an unexpected character:

"Large diurnal temperature ranges (¿ 20°C) ..."

2. Top of page 4, there is an unexpected character:

"In areas where dengue incidence is typically low (e.g., ¡ 10 cases per

month) ..."

3. Figure S7: consider amending the legend to read "No Economic Value" and

"Economic Value", so that it is explicit what "value" is being considered.

-----------------------------------------------------------

Reviewer #4:

This is an interesting paper on an important subject, the development of early warning systems for vector-borne diseases, focusing here on dengue in Vietnam. While the general approach appears reasonable there are a number of clarifications and amendments required. Specific comments are given below.

The model predictions presented by the authors seem useful but they are currently somewhat of a black box. In particular, there are two directions for improving the presentation of the results. The first is based upon standard prequential theory of proper scoring rules and the paper would improve if the authors would decompose the overall prediction error for the different scoring rules into calibration and sharpness, the predictive measures corresponding to bias and variance

In addition, it would be very useful for this work, as well as future developments, to see the individual contribution to the predictive ability of the different components. For example, are the climate forecasts adding substantially to the predictive ability or the earth observation data correspond to the best part of the ensemble's predictive power?

Also, how accurate climate forecasts are the predictors of seasonal climate, i.e. how good are these forecasts on their own right?

This will be illuminating and could also guide the authors' future efforts in improving this work. It will also shed some light upon the paper's drawbacks, since absence of disease control and related data appears rather important, at least a-priori.

In addition, it will lead to a discussion on issues of model mis-specification, since moderate predictive ability of a covariate (like seasonal climate) but reasonable prediction of a function thereof (such as disease incidence) suggests issues of potential model mis-specification.

A vital aspect of this work is the predictive ability of this approach and the authors claim that this is superior to the competition. However, as far as can be seen in this paper, the comparison is not made with a number of state-of-the-art alternatives, but with a baseline model which is a sub-case of the model class used in the paper.

Is that an accurate description?

If so, are the model's conclusions a fair representation of the paper's findings?

This should be clarified and amended as appropriate.

An important clarification should be made with respect to terminology.

Is model superensemble an alternative name for bayesian model averaging (BMA)?

If so this should be clarified.

Incidentally, BMA essentially started in the fifties with the work of Savage, Lindley and Good so appropriate referencing could be added.

Non-Bayesian (e.g. AIC-based) model weights are also commonly used, including in ecology where the Burnham and Anderson's 2002 book on multimodel inference is highly influential.

In relation to the above comment, please clarify the definition of the ML weights used in equation 5. The DIC weights are often used as a Bayesian equivalent of the AIC-based ones. However, the most commonly used weights are BIC-based, essentially since they agree asymptotically with standard Bayesian theory. However, ML weights are non-standard and the authors will need to specify how are they calculated and give a reference to their theoretical justification.

The model class that this work is based upon seems reasonable. However, some details need to be added, including the specific parameterisation of the negative binomial distribution and the definition of mu (is it the mean of the log data?)

-----------------------------------------------------------

Reviewer #5: The manuscript presents a Bayesian spatio-temporal model for Dengue

counts, which incorporates weather and land-cover variables. The five

subsets of covariates that produced the best one-to-six-months forecasts

during the period 2007-2016 were combined via Bayesian model averaging.

This superensemble outperformed the ensemble members in terms of both CRPS

and point forecasts (RMSE, MAE). It was eventually used to generate

probabilistic forecasts for the 6 months ahead of April 2020.

Cooperation with key stakeholders from from public health institutions

strengthens the paper and supports transferring scientific progress into

practice.

The paper is well written overall, but I have several questions and remarks

about details of the model and how forecasts are generated.

Major comments

--------------

1. An AR1 prior is used for \\delta_{i,m[t]}. Shouldn't the model assume

a cyclic effect, i.e., have January conditioned on December?

2. Eq (1): What is the motivation behind scaling past counts by the

population, i.e., having the terms (1-\\rho) log(P) + \\rho log(Y)?

I think I haven't seen this in related models before. Did explorative

analyses indicate a better fit when incorporating past incidence

instead of past counts?

3. Related to 2.: The model fit and forecasts could probably be improved

by allowing for a region-specific effect of log(Y_{i,t-1}), i.e.,

region-specific force of infection \\rho_i. Scaling by the population

would no longer be necessary and the model would more adequately

capture regional variation in unobserved important determinants of

disease spread. A recent example of a related model with

region-specific force of infection is https://doi.org/10.1002/sim.8390 .

4. I was really confused by how k-step-ahead (k>1) forecasts are generated.

The manuscript intends to provide superensemble forecasts of regional

dengue incidence up to 6 months ahead *using seasonal climate forecasts*,

but then a *climate naive* model is used to generate forecasts for t+2

to t+6? Why make the (impressive!) effort of collecting climate data and

forecasts if they are not used to inform probabilistic forecasts?

Furthermore, how is the forecast distribution generated? Are these

simple negative binomial forecasts conditioning on the *point*

forecast for y_{t+1}, ignoring the uncertainty in that forecast? I'm not sure

if this counts as "properly reflect forecast uncertainty" (discussion).

Maybe simulation-based path forecasts should be used instead?

5. Eq (3): I think it may be useful to note that the baseline model isn't

such a simple model after all. The province-specific and time-varying

random effects will compensate for much of the missing X variables. I

think a more important "feature" of the baseline model is the missing

autoregressive term. This omission should also be kept in mind when

discussing performance relative to this model around p. 15, l. 504. I

would expect a non-dynamic model to perform particularly well in

regions with only sporadic dengue cases. Please also report

the performance of the baseline model in Table 1.

6. One model should really be added to prove the usefulness of the

superensemble for forecasting: Does the superensemble outperform a model

which contains all of the covariates of models 1 to 5 combined? If so I'm

willing to believe that "we demonstrate that using a model superensemble

results in better forecasts than using individual models" (discussion).

7. Can we conclude from Figure 1 that it would make sense to include a

non-dynamic baseline model in the ensemble? As Reich et al (2019) put

it: "building ensemble models that capitalize on the strengths of a

diverse set of individual component models will be critical to

improving accuracy and consistency of models in all infectious disease

forecasting settings." (https://doi.org/10.1073/pnas.1812594116)

8. Table 1: Please report which lags have been selected for the climate

predictors in the different models. I think this could be an

interesting result by itself.

9. The ROC analysis was primarily used to compare AUC values across different

forecast horizons (Figure 5), which is not very interesting. It seems more

relevant to compare how well the superensemble can forecast the four

different outbreak indicators (there does not seem to be much difference but

it is difficult to compare across plots). The text also says that an optimal

cut-off value for the probability is defined to maximize sensitivity and

specificity (which is no clear criterion since you cannot maximize both), but

I cannot find the resulting cut-offs.

Minor comments

--------------

1. Strong underreporting:

* Intro (first two paragraphs): The question arises why there is such

strong underreporting. Does dengue seldom require treatment? The intro

should probably mention common symptoms of dengue infection.

* At the end, the discussion briefly mentions that the model only

considers *reported* disease incidence. Thus, routine forecasting

relies on stable reporting rates. If there were efforts to reduce

underreporting at some point, endemic channels or quantiles based on

historical incidence wouldn't make sense any more. Are there any such

efforts in Vietnam? How could forecasts integrate such a (gradual)

change?

2. p. 3, l. 82: "deterministic models" are mixed up with "point forecasts"

here. ARIMA or Poisson regression used in the referenced papers are

*not* deterministic. The point is that you would like to quantify

the uncertainty of your forecasts in addition to your best (point)

estimate of future disease spread.

3. p. 4, l. 102: This sounds like CUSUM methods should be considered,

see e.g., https://doi.org/10.1016/j.prevetmed.2009.05.017 .

4. Materials and methods should also say where the shapefiles originate.

5. Ref 59 for the Worldpop project contains a wrong URL.

6. Why are two different population datasets [refs 59 and 60] used for P

in the model and to weight monthly averages of climate variables,

respectively?

7. p. 5, l. 161: land-cover variables are per province _and per year_

(please clarify this). I thought they were time-constant until I read

they vary annually later in the text.

8. p. 5, l. 175: There seem to be two data sources here; where did the

data eventually come from? Furthermore, the training period ranged from

2002 to 2006, the testing period from 2007 to 2016, and the forecasting

period from April 2020 to September 2020. How does hindcast data for

only 2012-2016 fit in here?

9. p. 6, l. 182: I'd prefer to see "month t = 1, ..., T" instead of the

more general "time t" to clarify the temporal resolution of the model.

10. On Eq (1): Please specify the distribution assumed for the temporal

random effects, probably Gaussian.

11. p. 6, l. 188: This is a log-linear model with log-transformed past

counts, so \\rho is not an autoregressive *coefficient* in the usual

sense: there is a multiplicative effect (Y_{i,t-1}/P_{i,t-1})^\\rho on

the mean. Maybe the more general term "autoregressive parameter"

should be used for \\rho.

12. p. 6, l. 215: I don't understand why annual variation is a reason for

the land-cover variables to be excluded from the selection procedure.

13. p. 6, l. 225: n is undefined.

14. p. 7, l. 230: "the best two performing models for each verification

metric" doesn't seem correct. The results show only 5 models, so it

seems only the top model for each verification metric has been

selected. But then in Table 1, model 2 seems to be best for both RMSE

and MAE, so how obtain five component models? Please clarify the

selection procedure.

15. Eq (4) and (5): It seems there is some confusion with indices i and k.

For example, M_k in (4) should be M_i. Furthermore, why is w_i a

vector and how does it sum to 1 given (5)? The "ML or DIC" part isn't

quiet clear.

16. Eq (6): CRPS_f and CRPS_b are undefined.

17. p. 9, l. 314-330: This description is repetitive and should be

dropped. The results section should concentrate on results, not

methods.

18. Figure 7: Shouldn't this be based on the 95% instead of the 75% quantile

since the former had better scores?

-----------------------------------------------------------

Any attachments provided with reviews can be seen via the following link:

[LINK]

Decision Letter 2

Artur Arikainen

20 Nov 2020

Dear Dr. Colón-González,

Thank you very much for submitting your revised manuscript "Probabilistic seasonal dengue forecasting in Vietnam: A modelling study using superensembles" (PMEDICINE-D-20-02989R2) for consideration at PLOS Medicine.

Your paper was once again evaluated by a senior editor and discussed among all the editors here. It was also sent to three of the previous independent reviewers, including a statistical reviewer. The reviews are appended at the bottom of this email and any accompanying reviewer attachments can be seen via the link below:

[LINK]

In light of these reviews, we would like to once more consider a revised version that addresses the reviewers' and editors' comments. Obviously we cannot make any decision about publication until we have seen the revised manuscript and your response, and we may seek re-review by one or more of the reviewers.

In revising the manuscript for further consideration, your revisions should address the specific points made by each reviewer and the editors. Please also check the guidelines for revised papers at http://journals.plos.org/plosmedicine/s/revising-your-manuscript for any that apply to your paper. In your rebuttal letter you should indicate your response to the reviewers' and editors' comments, the changes you have made in the manuscript, and include either an excerpt of the revised text or the location (eg: page and line number) where each change can be found. Please submit a clean version of the paper as the main article file; a version with changes marked should be uploaded as a marked up manuscript.

In addition, we request that you upload any figures associated with your paper as individual TIF or EPS files with 300dpi resolution at resubmission; please read our figure guidelines for more information on our requirements: http://journals.plos.org/plosmedicine/s/figures. While revising your submission, please upload your figure files to the PACE digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email us at PLOSMedicine@plos.org.

We expect to receive your revised manuscript by Dec 11 2020 11:59PM. Please email us (plosmedicine@plos.org) if you have any questions or concerns.

***Please note while forming your response, if your article is accepted, you may have the opportunity to make the peer review history publicly available. The record will include editor decision letters (with reviews) and your responses to reviewer comments. If eligible, we will contact you to opt in or out.***

We ask every co-author listed on the manuscript to fill in a contributing author statement, making sure to declare all competing interests. If any of the co-authors have not filled in the statement, we will remind them to do so when the paper is revised. If all statements are not completed in a timely fashion this could hold up the re-review process. If new competing interests are declared later in the revision process, this may also hold up the submission. Should there be a problem getting one of your co-authors to fill in a statement we will be in contact. YOU MUST NOT ADD OR REMOVE AUTHORS UNLESS YOU HAVE ALERTED THE EDITOR HANDLING THE MANUSCRIPT TO THE CHANGE AND THEY SPECIFICALLY HAVE AGREED TO IT. You can see our competing interests policy here: http://journals.plos.org/plosmedicine/s/competing-interests.

Please use the following link to submit the revised manuscript:

https://www.editorialmanager.com/pmedicine/

Your article can be found in the "Submissions Needing Revision" folder.

To enhance the reproducibility of your results, we recommend that you deposit your laboratory protocols in protocols.io, where a protocol can be assigned its own identifier (DOI) such that it can be cited independently in the future. For instructions see http://journals.plos.org/plosmedicine/s/submission-guidelines#loc-methods.

Please ensure that the paper adheres to the PLOS Data Availability Policy (see http://journals.plos.org/plosmedicine/s/data-availability), which requires that all data underlying the study's findings be provided in a repository or as Supporting Information. For data residing with a third party, authors are required to provide instructions with contact information for obtaining the data. PLOS journals do not allow statements supported by "data not shown" or "unpublished results." For such statements, authors must provide supporting data or cite public sources that include it.

We look forward to receiving your revised manuscript.

Sincerely,

Artur Arikainen,

Associate Editor

PLOS Medicine

plosmedicine.org

-----------------------------------------------------------

Requests from the editors:

1. Please address the reviewer’s additional comments below on the description of your methodology.

2. Please include line numbers in the margin throughout your text, including the abstract and author summary.

3. Abstract: Please include some summary quantitative data on the fit of your model to prior data and on your future predictions (with 95% CIs and p values, as appropriate).

---

Comments from the reviewers:

Reviewer #2: Thank you for addressing my previous comments. It's a nice article, well done.

Reviewer #4: The authors have revised their paper and this now represents a substantially improved manuscript, acceptable for publication

Reviewer #5: My main concern remains how k-step ahead forecasts are generated. Since forecasting is the main purpose of this manuscript, I think the corresponding description should be improved.

I admit that I didn't have the time to look at the revised version in every detail. Nevertheless, I stumbled over a few issues in the authors' reply, which I would like to bring up for consideration.

1. Eq 3+4: The math notation is too sloppy for a Bayesian inference approach. I've preferred the previous simplistic textual description of assuming a NegBin for Y_it. Now a random variable $Y$ (without $_{i,t}$) is introduced, whose meaning is undefined, and appending y >= 0 doesn't make any sense. This Bayesian model should be either formulated with mathematical rigour, or with just the core part, i.e., Eq. 4, i.e., the conditional mean of $Y_{i,t}$.

2. Eq. 3 would imply that this paper assumes a marginal NegBin for $Y_{i,t}$, but my previous reading suggests that a conditional NegBin was assumed, in particular also conditioning on the past observations just like for other covariates. Please clarify.

3. Eq. 4 now says that $\\rho$ works on past counts not past incidence. However, the text below still contains Y/P. Please check.

4. The reply confirms that k-step-ahead (k>1) forecasts as implemented here do not carry forward the uncertainty of one-step-ahead forecasts. My interpretation is that the authors eventually assume a NegBin distribution for their forecast $Y_{t+2} | Y_t, ...$, by *imputing* a point prediction for $Y_{t+1}$ to actually compute $Y_{t+2} | Y_{t+1}, ...$. AFAIK, $Y_{t+2} | Y_t$ follows a *mixture* of NegBin distributions (which is not necessarily NegBin), see, e.g., https://doi.org/10.1016/j.ijforecast.2020.07.002 (Section 3.4) for a correct approach. I understand that it is too late to fix this, but it would still be possible to include a cautious comment that all forecasts (including for k > 1) assume a conditional NegBin distribution, imputing one-step-ahead point predictions for lagged counts. I don't know if there is a name for such "naive" forecasts.

5. Table 1 should use two digits in the Bias column.

6. A "ROC curve" analysis would be used to choose a diagnostic threshold and display its effect on TPR and FPR and to display the overall diagnostic accuracy as the area under the ROC curve. I'm thus puzzled by the use of the term "ROC curve" analysis in the manuscript.

---

Any attachments provided with reviews can be seen via the following link:

[LINK]

Decision Letter 3

Artur Arikainen

12 Jan 2021

Dear Dr. Colón-González,

Thank you very much for re-submitting your manuscript "Probabilistic seasonal dengue forecasting in Vietnam: A modelling study using superensembles" (PMEDICINE-D-20-02989R3) for review by PLOS Medicine.

I have discussed the paper with my colleagues and the academic editor and it was also seen again by one of the reviewers. I am pleased to say that provided the remaining editorial and production issues are dealt with we are planning to accept the paper for publication in the journal.

The remaining issues that need to be addressed are listed at the end of this email. Any accompanying reviewer attachments can be seen via the link below. Please take these into account before resubmitting your manuscript:

[LINK]

***Please note while forming your response, if your article is accepted, you may have the opportunity to make the peer review history publicly available. The record will include editor decision letters (with reviews) and your responses to reviewer comments. If eligible, we will contact you to opt in or out.***

In revising the manuscript for further consideration here, please ensure you address the specific points made by each reviewer and the editors. In your rebuttal letter you should indicate your response to the reviewers' and editors' comments and the changes you have made in the manuscript. Please submit a clean version of the paper as the main article file. A version with changes marked must also be uploaded as a marked up manuscript file.

Please also check the guidelines for revised papers at http://journals.plos.org/plosmedicine/s/revising-your-manuscript for any that apply to your paper. If you haven't already, we ask that you provide a short, non-technical Author Summary of your research to make findings accessible to a wide audience that includes both scientists and non-scientists. The Author Summary should immediately follow the Abstract in your revised manuscript. This text is subject to editorial change and should be distinct from the scientific abstract.

We expect to receive your revised manuscript within 1 week. Please email us (plosmedicine@plos.org) if you have any questions or concerns.

We ask every co-author listed on the manuscript to fill in a contributing author statement. If any of the co-authors have not filled in the statement, we will remind them to do so when the paper is revised. If all statements are not completed in a timely fashion this could hold up the re-review process. Should there be a problem getting one of your co-authors to fill in a statement we will be in contact. YOU MUST NOT ADD OR REMOVE AUTHORS UNLESS YOU HAVE ALERTED THE EDITOR HANDLING THE MANUSCRIPT TO THE CHANGE AND THEY SPECIFICALLY HAVE AGREED TO IT.

Please ensure that the paper adheres to the PLOS Data Availability Policy (see http://journals.plos.org/plosmedicine/s/data-availability), which requires that all data underlying the study's findings be provided in a repository or as Supporting Information. For data residing with a third party, authors are required to provide instructions with contact information for obtaining the data. PLOS journals do not allow statements supported by "data not shown" or "unpublished results." For such statements, authors must provide supporting data or cite public sources that include it.

Please note, when your manuscript is accepted, an uncorrected proof of your manuscript will be published online ahead of the final version, unless you've already opted out via the online submission form. If, for any reason, you do not want an earlier version of your manuscript published online or are unsure if you have already indicated as such, please let the journal staff know immediately at plosmedicine@plos.org.

If you have any questions in the meantime, please contact me or the journal staff on plosmedicine@plos.org.  

We look forward to receiving the revised manuscript by Jan 19 2021 11:59PM.   

Sincerely,

Artur Arikainen,

Associate Editor 

PLOS Medicine

plosmedicine.org

------------------------------------------------------------

Requests from Editors:

1. Data Availability Statement (DAS): Note that a study author cannot be the contact person for a dataset. Please provide alternative details for your dengue data.

2. Abstract: Please ensure that all quantitative data are also presented in the main Results section, and that the values match.

3. Throughout, please perhaps replace “skill” with a more appropriate term (eg. ‘strength’), when describing the performance of the forecast system.

4. In the introduction, please mention dengue vaccination - although a complex and not very satisfactory situation, this could be discussed.

5. Please mention in your Methods that your study did not require separate ethical approval.

6. Line 749: "explicitly incorporate".

7. Please provide a URL or DOI for reference 6.

8. RE reference 91 listed as ”in press”, papers cannot be listed in the reference list until they have been accepted for publication or are otherwise publicly accessible (for example, in a preprint archive). The information may be cited in the text as a personal communication with the author if the author provides written permission to be named. Alternatively, please provide a different appropriate reference.

Comments from Reviewers:

Reviewer #5: Thank you for addressing my previous comments, especially for clarifying k-step-ahead as imputed one-step-ahead forecasts and including cautious comments about that.

In my opinion, this manuscript is now ready for publication. Congratulations!

Any attachments provided with reviews can be seen via the following link:

[LINK]

Decision Letter 4

Artur Arikainen

22 Jan 2021

Dear Dr Colón-González, 

On behalf of my colleagues and the guest Academic Editor, Alex R. Cook, I am pleased to inform you that we have agreed to publish your manuscript "Probabilistic seasonal dengue forecasting in Vietnam: A modelling study using superensembles" (PMEDICINE-D-20-02989R4) in PLOS Medicine.

Before your manuscript can be formally accepted you will need to complete some formatting changes, which you will receive in a follow up email. Please be aware that it may take several days for you to receive this email; during this time no action is required by you. Once you have received these formatting requests, please note that your manuscript will not be scheduled for publication until you have made the required changes.

In the meantime, please log into Editorial Manager at http://www.editorialmanager.com/pmedicine/, click the "Update My Information" link at the top of the page, and update your user information to ensure an efficient production process. 

PRESS

We frequently collaborate with press offices. If your institution or institutions have a press office, please notify them about your upcoming paper at this point, to enable them to help maximise its impact. If the press office is planning to promote your findings, we would be grateful if they could coordinate with medicinepress@plos.org. If you have not yet opted out of the early version process, we ask that you notify us immediately of any press plans so that we may do so on your behalf.

We also ask that you take this opportunity to read our Embargo Policy regarding the discussion, promotion and media coverage of work that is yet to be published by PLOS. As your manuscript is not yet published, it is bound by the conditions of our Embargo Policy. Please be aware that this policy is in place both to ensure that any press coverage of your article is fully substantiated and to provide a direct link between such coverage and the published work. For full details of our Embargo Policy, please visit http://www.plos.org/about/media-inquiries/embargo-policy/.

Thank you again for submitting to PLOS Medicine. We look forward to publishing your paper. 

Sincerely, 

Artur A. Arikainen 

Associate Editor 

PLOS Medicine

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    S1 Text. Prospective analysis plan.

    (DOCX)

    S1 Fig. Observed dengue cases across Vietnam.

    Time series of monthly dengue cases from the 63 provinces in Vietnam (August 2002 to March 2020). Provinces are ordered from north (top) to south (bottom) according to the latitude coordinates of their centroid. White boxes indicate missing data.

    (TIF)

    S2 Fig. Verification metrics of the seasonal climate forecasts evaluated across Vietnam.

    The x axis indicates the month of the year. The y axis indicates value of the CRPS for each variable. The lines indicate the lead time for the forecast. CRPS, continuous rank probability score.

    (TIF)

    S3 Fig. Optimal weights.

    Seasonal variation of the optimal weights assigned to each of the models included in the superensemble over the period January 2007 to December 2016.

    (TIF)

    S4 Fig. Observed vs. predicted dengue cases.

    Observed (dashed lines) and predicted (solid lines) dengue cases across Vietnam aggregated at the national level. Shaded areas represent the 95% credible interval. Predictions are shown for the forecast horizons of 1 (top), 3 (middle), and 6 (bottom) months ahead. Data correspond to the period January 2007 to December 2016.

    (TIF)

    S5 Fig. Prospective dengue predictions.

    Predicted dengue cases for the period May to October 2020 for 4 pilot Vietnamese provinces using a model superensemble. The forecast was issued on May 10, 2020. The x axis (top) indicates the month of the predictions. The x axis (bottom) indicates the time lead of the predictions. The y axis indicates the predicted incidence rate. Black solid lines indicate the posterior mean estimate for each of 42 forecast ensemble members. Red lines indicate the observed dengue incidence rate which was not known by the model at the time of the computation and are included here as a reference. The black dashed lines indicate the upper and lower bounds of the 95% credible intervals for the 42 ensemble members. The upper bound of the shaded areas indicates the month- and province-specific percentiles based on dengue data for previous 5 years.

    (TIF)

    S6 Fig. Month-specific variability in dengue cases across Vietnam.

    The x axis indicates the month of the year. The y axis indicates increases in the number of dengue cases (square root transformed). The upper and lower limits of each box represent the interquartile range of the distribution of dengue cases for each month. The middle solid line indicates the median value. The upper and lower whiskers indicate the maximum and minimum values of the dengue case distribution (excluding outliers which are indicated with dark purple circles). Outliers are values beyond 1.5 times the interquartile range.

    (TIF)

    Attachment

    Submitted filename: Response_to_reviewers.docx

    Attachment

    Submitted filename: Response_to_reviewers_rev2.docx

    Attachment

    Submitted filename: Response_to_editorial_requests.docx

    Data Availability Statement

    The data underlying the results presented in the study are available from: Dengue data: Dr Pham Hung, Head of Disease Control Department, Vietnam. Email: hungvncdc@gmail.com Precipitation The precipitation dataset used is a derivative work from NASA GPM and NASA TRRM, https://www.usa.gov/government-works. Wind Speed The wind speed dataset used is a derivative work from NOAA Climate Forecast System Version 2 (CFSv2) Operational Analysis, https://www.usa.gov/government-works. Land Cover The land cover dataset used is a derivative work from Land Cover, © ESA Climate Change Initiative - Land Cover led by UCLouvain (2017). https://www.esa-landcover-cci.org/?q=node/164 Population Population data was used from: WorldPop (www.worldpop.org - School of Geography and Environmental Science, University of Southampton; Department of Geography and Geosciences, University of Louisville; Departement de Geographie, Universite de Namur) and Center for International Earth Science Information Network (CIESIN), Columbia University (2018). Global High Resolution Population Denominators Project - Funded by The Bill and Melinda Gates Foundation (OPP1134076). https://dx.doi.org/10.5258/SOTON/WP00645. Licenced under Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0). GPW Population Count, v4.10 (2000, 2005, 2010, 2015, 2020). Licenced under Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0). UN WPP UN WPP-Adjusted Population Count, v4.10 (2000, 2005, 2010, 2015, 2020). Center for International Earth Science Information Network (CIESIN), Columbia University. 2017. Gridded Population of the World, Version 4 (GPWv4): Population Count Adjusted to Match 2015 Revision of UN WPP Country Totals, Revision 10. Palisades, NY: NASA Socioeconomic Data and Applications Center (SEDAC). https://doi.org/10.7927/H4JQ0XZW. Accessed 16/09/2019. Humidity The humidity dataset used is a derivative work from NASA MODIS/Terra Total Precipitable Water Vapor 5-Min L2 Swath 1km and 5km – NRT, https://www.usa.gov/government-works. Temperature The temperature dataset used is a derivative work from NASA MODIS/Terra Land Surface Temperature/Emissivity Daily L3 Global 1km SIN Grid V006 and MODIS/Aqua Land Surface Temperature/Emissivity Daily L3 Global 1km SIN Grid V006, https://www.usa.gov/government-works.


    Articles from PLoS Medicine are provided here courtesy of PLOS

    RESOURCES