Skip to main content
PLOS Computational Biology logoLink to PLOS Computational Biology
. 2020 Jul 1;16(7):e1007990. doi: 10.1371/journal.pcbi.1007990

Using information theory to optimise epidemic models for real-time prediction and estimation

Kris V Parag 1,*, Christl A Donnelly 1,2
Editor: Matthew (Matt) Ferrari3
PMCID: PMC7360089  PMID: 32609732

Abstract

The effective reproduction number, Rt, is a key time-varying prognostic for the growth rate of any infectious disease epidemic. Significant changes in Rt can forewarn about new transmissions within a population or predict the efficacy of interventions. Inferring Rt reliably and in real-time from observed time-series of infected (demographic) data is an important problem in population dynamics. The renewal or branching process model is a popular solution that has been applied to Ebola and Zika virus disease outbreaks, among others, and is currently being used to investigate the ongoing COVID-19 pandemic. This model estimates Rt using a heuristically chosen piecewise function. While this facilitates real-time detection of statistically significant Rt changes, inference is highly sensitive to the function choice. Improperly chosen piecewise models might ignore meaningful changes or over-interpret noise-induced ones, yet produce visually reasonable estimates. No principled piecewise selection scheme exists. We develop a practical yet rigorous scheme using the accumulated prediction error (APE) metric from information theory, which deems the model capable of describing the observed data using the fewest bits as most justified. We derive exact posterior prediction distributions for infected population size and integrate these within an APE framework to obtain an exact and reliable method for identifying the piecewise function best supported by available epidemic data. We find that this choice optimises short-term prediction accuracy and can rapidly detect salient fluctuations in Rt, and hence the infected population growth rate, in real-time over the course of an unfolding epidemic. Moreover, we emphasise the need for formal selection by exposing how common heuristic choices, which seem sensible, can be misleading. Our APE-based method is easily computed and broadly applicable to statistically similar models found in phylogenetics and macroevolution, for example. Our results explore the relationships among estimate precision, forecast reliability and model complexity.

Author summary

Understanding how the population of infected individuals (which may be humans, animals or plants) fluctuates in size over the course of an epidemic is an important problem in epidemiology and ecology. The effective reproduction number, R, provides an intuitive and useful way of describing these fluctuations by characterising the growth rate of the infected population. An R > 1 signifies a burgeoning epidemic whereas R < 1 indicates a declining one. Public health agencies often use R to inform or corroborate vaccination and quarantine policies. However, popular approaches to inferring R from epidemic data make heuristic choices, which may lead to visually reasonable estimates that are deceptive or unreliable. By adapting mathematical tools from information theory, we develop a general and principled scheme for estimating R in a data-justified way. Our method exposes the pitfalls of heuristic estimates and provides an easily computable correction that also maximises our ability to predict upcoming population fluctuations. Our work is widely applicable to similar inference problems found in evolution and genetics, demonstrably useful for reliably analysing emerging epidemics in real time and highlights how abstract mathematical concepts can inspire novel and practical biological solutions, showcasing the importance of multidisciplinary research.

Introduction

The time-series of newly infected cases (infecteds) observed over the course of an infectious disease epidemic is known as an incidence curve or epi-curve. These curves offer prospective insight into the spread of a disease within an animal or human population by informing on the effective reproduction number, which defines the average number of secondary infections induced by a primary one [1]. This reproduction number, denoted Rt at time t, is an important prognostic of the demographic behaviour of an epidemic. If Rt > 1, for example, we can expect and hence prepare for exponentially increasing incidence, whereas if Rt < 1, we can be reasonably confident that the epidemic has been arrested [1].

Reliably estimating meaningful changes in Rt is an important problem in epidemiology and population biology, since it can forewarn about the growth rate of an outbreak and signify the level of control effort that must be initiated or sustained [2, 3]. While we explicitly consider epidemic applications here, similar reproduction numbers (and growth rates) can be defined for many ecological problems [4], for example in species conservation where we might aim to infer species population dynamics from time-series of sample counts.

The renewal model [5] is a popular approach for inferring salient fluctuations in Rt that is based on the fundamental Euler-Lotka reproduction equation from ecology and evolution [4, 6]. This model has been used to predict Ebola virus disease case counts and assess the transmission potential of pandemic influenza and Zika virus, among others [2, 7, 8]. Currently, it is in widespread use as a tool for tracking the progress of interventions across countries in response to the COVID-19 pandemic [9]. The renewal model may be applied retrospectively to understand the past behaviour of an epidemic or prospectively to gain real-time insight into ongoing outbreak dynamics [5, 10]. We only consider the latter here (see [11] for an investigation of the former).

The prospective approach approximates Rt with a piecewise-constant function i.e. Rt is constant (stable) over some sliding window of k time units (e.g. days or weeks) into the past, beyond which a discontinuous change is assumed [10]. This formulation models the non-stationary nature of epidemics, capturing the idea that different population dynamics are expected during distinct phases (e.g. onset, growth, control) of the epidemic lifetime. The window length, k, is essentially a hypothesis about the stability of Rt underpinning the observed epi-curve. It is critical to reliably characterising the epidemic because it controls the time-scale over which Rt fluctuations are deemed significant.

Too large or small a k-value can respectively lead to over-smoothing (which ignores important changes) or to random noise being misinterpreted as meaningful. Inferring under a wrong k can appreciably affect our understanding of an epidemic, as observed in [5], where case reproduction numbers (a common k-choice) were found to over-smooth significant changes in HIV transmission, for example. Surprisingly, no principled method for optimising k exists. Current best practice either relies on heuristic choices [7] or provides implicit bounds on k [10]. Here we adapt the minimum description length (MDL) principle from information theory to develop a simple but rigorous selection framework.

The description length is defined as the number of bits needed to communicate a model (M) and the data given that model (D|M) on some channel. More complex models increase L(M) (more bits are needed to represent its extra parameters) but decrease L(D|M) (as it can better fit the data) for example. Here L(.) indicates length. The MDL model choice minimises L(M)+L(D|M) and is considered the model most justified by the data D [12]. We adapt an approximation to MDL known as the accumulated prediction error (APE) [13] to identify the k best justified by the available epi-curve, k* (see the S1 Text for further details and for the general definition of APE).

We analytically derive the posterior predictive incidence distribution of the renewal model, which allows us to evaluate cumulative log-loss prediction scores at any k exactly. The APE choice, k*, minimises these scores and, additionally, optimises short-term prediction accuracy [14]. Our method is valid at all sample sizes, easily computed for arbitrary model dimensions and, unlike many selection criteria of similar computability, includes parametric complexity [12]. Parametric complexity measures how functional relationships among parameters influence complexity, and when ignored (as in the Bayesian or Akaike information criteria) can lead to biased renewal model selection [11].

The APE metric designates the model that best predicts unseen data from the same underlying process, and not the one that best fits existing data, as optimal and of justified complexity [13]. This is equivalent to minimising what is called the generalisation or out-of-sample error in machine learning, and is known to balance under and overfitting. Our APE-based approach therefore pinpoints and characterises only those Rt fluctuations that are integral to achieving reliable short-term incidence growth predictions.

The performance, speed and computational ease of our approach makes it suitable for real-time forecasting and model selection. It could therefore serve as a stand-alone computational tool or be integrated within existing real-time frameworks, such as in [3] or [10], to provide emerging insights into infected population dynamics, or to assess the prospective efficacy of implemented interventions (e.g. vaccination or quarantine). Public health policy decisions or preparedness plans based on improperly specified k-windows could be misinformed or overconfident. Our method hopefully limits these risks.

Methods

Renewal model window-sizing problem

Let the incidence or number of newly infected cases in an epidemic at present time t, be It. The incidence curve is a historical record of these case counts from the start of the outbreak and is summarised as I1t={Is:1st}, with s as a time-indexing variable. For convenience, we assume that incidence is available on a daily scale so that I1t is a vector of t daily counts (weeks or months could be used instead). The associated effective reproduction number and total infectiousness of the epidemic are denoted Rt and Λt, respectively.

Here Rt is the number of secondary cases that are induced, on average, by a single primary case at t [1], while Λt measures the cumulative impact of past cases, from the epidemic origin at s = 1, on the present. The generation time distribution of the epidemic, which describes the elapsed time between a primary and secondary case, controls Λt (see the S1 Text). In the problems we consider, I1t and Λ1t form our data, the generation time distribution is assumed known, and R1t or groupings of this vector are the parameters to be inferred.

The renewal model [5] derives from the classic Euler-Lotka equation [6] (see the S1 Text), and defines the Poisson distributed (Poiss) relationship It ∼ Poiss(RtΛt). This means that (x)=1x!eRtΛt(RtΛt)x, with x indexing possible It values. This formulation assumes maximum epidemic non-stationarity (i.e. that demographic transmission statistics change at every time unit) and hence only uses the most recent data (It, Λt) to infer the current Rt. While this maximises the fitting flexibility of the renewal model, it often results in noisy and unreliable estimates that possess many spurious reproduction number changes [7]. Consequently, grouping is employed.

This hypothesises that the reproduction number is constant over a k-day window into the past and results in a piecewise-constant function that separates salient fluctuations (change-points) from negligible ones (the constant segments) [10, 11]. We use Rτ(t) to indicate that the present reproduction number to be inferred is stationary (constant) over the last k points of the incidence curve i.e. the time window τ(t) ≔ {t, t − 1, …, tk + 1}. The data used to estimate Rτ(t) is then (It-k+1t,Λt-k+1t). This construction allows us to filter noise and increase estimate reliability but elevates bias. In the S1 Text we show precisely how grouping achieves this bias-variance trade-off.

Choosing k therefore amounts to selecting a belief about the scale over which the epidemic statistics are meaningfully varying and can significantly influence our understanding of the population dynamics of the outbreak. Thus, it is necessary to find a principled method for balancing k. Ultimately, we want to find a k, denoted k*, that (i) is best supported by the epi-curve and (ii) maximises our confidence in making real-time, short-term predictions that can inform prospective epidemic responses. This is no trivial task as k* would be sensitive to both the specifically observed stochastic incidence of an epidemic and to how past infections propagate forward in time.

We solve (i)-(ii) by applying the accumulated prediction error (APE) metric from information theory [13], which values models on their capacity to predict unseen data from the generating process instead of their ability to fit existing data [14]. The properties and mathematical definition of the APE are provided in the S1 Text and in the subsequent section. The APE uses the window of data preceding time s, (Is-k+1s,Λs-k+1s), to predict the incidence at s + 1 and assigns a log-score to this prediction. This procedure is repeated over st and for possible k-values. The k achieving the minimum cumulated log-score is deemed optimal. Fig 1 summarises and illustrates the APE algorithm.

Fig 1. Optimal window selection using APE.

Fig 1

(A) An observed incidence curve (blue dots) is sequentially and causally predicted over time st using effective reproduction number estimates based on two possible windows lengths of k1 and k2 (blue shaded). Predictive distributions are summarised by red error bars (shown only for times t1 and t2, respectively for k1 and k2), (B) The true reproduction number (Rs, dashed black) is estimated under each window length as R^τ1(s) (blue) and R^τ2(s) (grey). Large windows (k1) smooth over fluctuations. Small ones (k2) recover more changes but are noisy. (C) The APE assesses k1 and k2 via the log-loss of their sequential predictions (i.e. from red error bars across time). The window with the smaller APE is better supported by this incidence curve. See Methods for more mathematical details.

Exact model selection using predictive distributions

To adapt APE, we require the posterior predictive incidence distribution of the renewal model, which at time s is P(x|Is-k+1s), with x indexing the space of possible one-step-ahead predictions at s + 1, and I^s+1=E[x]. We assume a gamma conjugate prior distribution on Rτ(s) as is commonly done in the renewal model frameworks of [3, 10] and because the gamma distribution can fit many unimodal variables on the positive real line. For some hyperparameters a and c this is Rτ(s)Gam(a,1c), with Gam as a shape-scale parametrised gamma distribution. The posterior distribution of Rτ(s), P(Rτ(s)|Is-k+1s), is then

Rτ(s)|Is-k+1sGam(a+iτ(s),1c+λτ(s)). (1)

For convenience we define ατ(s)a+iτ(s),βτ(s)1c+λτ(s) with iτ(s) and λτ(s) are the sum of incidence (Is) and total infectiousness (Λs) over the window τ(s) (see the S1 Text). If a variable y ∼ Gam(α, β) then (y)=yα1ey/β/βαΓ(α) and E[y]=αβ. The posterior mean estimate is therefore R^τ(t)=ατ(t)βτ(t). Applying Bayes formula and marginalising yields the posterior predictive distribution of the number of infecteds at s + 1 as

P(x|Is-k+1s)=0P(x|Rτ(s))P(Rτ(s)|Is-k+1s)dRτ(s). (2)

In Eq (2) we used the conditional independence of future incidence data from the past epi-curve, given the reproduction number to reduce P(x|Is-k+1s,Rτ(s)) to P(x|Rτ(s)), which expresses the renewal model relation x ∼ Poiss(Rτ(s)Λs+1). As Λs+1 only depends on Is-k+1s there are no unknowns. Solving using this and Eq (1) gives P(x|Is-k+1s)=ϕ1ϕ2-1 where

ϕ1Λs+1x0Rτ(s)x+ατ(s)-1e-Rτ(s)(Λs+1+1βτ(s))dRτ(s)andϕ2βτ(s)ατ(s)Γ(x+1)Γ(ατ(s)),withΓ(z)0θz-1e-zdθ. (3)

Since θz1eyθdθΓ(z)yz the integral in ϕ1 simplifies to Γ(x+ατ(s))(Λs+1+1βτ(s))x+ατ(s). Some algebra then reveals the negative binomial (NB) distribution:

x|Is-k+1sNB(ατ(s),pτ(s)Λs+1βτ(s)1+Λs+1βτ(s)). (4)

If some variable y ∼ NB(α, p) then P(y)(α+y-1y)(1-p)αpy and E[y]=pα1p. Eq (4) is a key result and relates to a framework developed in [3]. It completely describes the one-step-ahead prediction uncertainty and has mean E[x]=I^s+1=Λs+1R^τ(s), and variance var(x)=I^s+1(1+I^s+1iτ(s)+a). These relations explicate how the current estimate of Rτ(s) influences our ability to predict upcoming incidence points. At s = t the above expressions yield prediction statistics for the next (unobserved) time-point beyond the present.

The APE metric is generally defined as APEks=1t-1-logP(Is+1|Is-k+1s) [13] (see the S1 Text) and is controlled by the shape of the posterior predictive distribution over time. We specialise this metric for renewal models and integrate Eq (4) into the algorithm of Fig 1 to derive

APEk=s=1t-1Bs+1+Is+1logpτ(s)+ατ(s)log(1-pτ(s)), (5)

with Is+1 as the (true) observed incidence at time s + 1, which is evaluated within the context of the predictive space of x, and Bs+1log(Is+1+ατ(s)-1Is+1). Eq (5) is exact, easy to evaluate and offers a simple and direct means of finding the optimal data-justified window length as k* ≔ arg minkAPEk.

The summation in Eq (5) also clarifies why our approach is useful for real-time applications. As an epidemic unfolds and more data accumulate the upper limit on s will increase. We can update APEk to account for emerging data by simply including an additional term for each new data point on top of the previously calculated sum. For example, a new datum at t + 1 requires the addition of Bt+1 + It+1 log pτ(t) + ατ(t) log (1 − pτ(t)) to Eq (5). In the Results section we will demonstrate this by successively computing ks*, the optimal window length for data up to time s.

APEk can also be computed using the in-built NB routines of many software (some parametrise this distribution differently so that Eq (4) may need to be implemented as NB(ατ(s), 1 − pτ(s))). When computing Eq (5) directly, the most difficult term is Bs+1 when Is+1 and ατ(s) are large. In these cases Stirling approximations may be applied. We provide Matlab and R implementations of our renewal model APE metric at https://github.com/kpzoo/model-selection-for-epidemic-renewal-models.

Results

Optimal window selection for dynamic outbreaks

We apply our APE metric to select k* for several epidemic examples, which examine reproduction number profiles featuring stable (Fig 2A) and seasonally fluctuating (Fig 2B) transmission, as well as exponential (Fig 3A) and step-changing (Fig 3B)) control measures. These examples explore dynamically diverse epidemic scenarios and consider both large (Figs 2A and 3A) and small outbreaks (Figs 2B and 3B). Small outbreaks are fundamentally more difficult to estimate. We simulated epidemics for t = 200 days using a generation time distribution analogous to that used for Ebola virus disease predictions in [15]. We investigated a window search space of 2kt2 and computed the APE at each k, over the epidemic duration (1 ≤ st). Reproduction number estimates, R^τ(s), and demographic incidence predictions x|Is-k+1s, are presented in Figs 2 and 3. Predictions over the first s < k times use all s data points.

Fig 2. Selection for stable and fluctuating epidemics.

Fig 2

Left graphs compare R^τ(s) estimates (blue with 95% confidence intervals) at the APE window length k* to those when k is set to its upper and lower limits. Right graphs give corresponding one-step-ahead predictions I^s given the window of data Iτ(s−1) (blue with 95% prediction intervals). Dashed lines are the true Rs numbers (left) and dots are the true Is counts (right). The panels examine (A) stable (constant) and (B) periodically varying changes in Rs.

Fig 3. Selection for epidemics with interventions.

Fig 3

Left graphs compare R^τ(s) estimates (blue with 95% confidence intervals) at the APE window length k* to those when k is set to its upper or lower limits. Right graphs give corresponding one-step-ahead predictions I^s given the window of data Iτ(s−1) (blue with 95% prediction intervals). Dashed lines are the true Rs numbers (left) and dots are the true Is counts (right). The panels examine (A) exponentially rising and decaying and (B) piecewise falling Rs due to differing interventions.

We find that the APE metric balances R^τ(s) estimate accuracy against one-step-ahead predictive coverage (i.e. the proportion of time that the true Is+1 lies within the 95% prediction intervals of x|Is-k+1s) for a range of Rs dynamics (top graphs of Figs 2 and 3). This dual optimisation is central to this work. Moreover, vastly different Is+1 predictions and R^τ(s) estimates can result when k is misspecified (middle and bottom graphs). This can be especially misleading when attempting to identify the significant changes in epidemic transmissibility and can support substantially different beliefs about the infection population biology. Estimation and prediction performance also depend on the actual incidence at any time; small Is values lead to wider confidence intervals in R^τ(s) at any k (see Eq. (S3) of the S1 Text).

When k is unjustifiably large not only do we observe systematic prediction and estimation errors, but alarmingly, we tend to be overconfident in them. When k is too small, we infer rapidly and randomly fluctuating R^τ(s) values, which sometimes can deceptively underlie reasonably looking incidence predictions. Consequently, optimal k-selection is integral for trustworthy inference and prediction. Observe that small k, which implies a more complex renewal model (i.e. there are more parameters to be inferred), does not generally result in better causally predictive one-step-ahead incidence predictions. Had we instead naively picked the k that best fits the existing epi-curve, then the smallest k would always be favoured (overfitting) [11].

We emphasize and validate the predictive performance of APE in Fig 4. This shows, for all simulated examples above that minimising the APE also approximately minimises the percentage of the time st that true incidence values fall outside the 95% prediction intervals of Eq (4). This validates the APE approach and evidences its proficiency at optimising for short-term forecasting accuracy in real time. We recommend using APE to define k* for an existing epi-curve up to the present t. Applying Eq (4) with this k* should then best predict the number of cases on the (t + 1)th day. This entire procedure should be repeated for later forecasts, with k* being progressively updated. The real-time behaviour of k* as data accumulate over the course of an outbreak is investigated in the next section.

Fig 4. APE prediction accuracy.

Fig 4

We compare the APE metric (blue, left y axes) to the percentage of true incidence values, Is+1 that fall outside the 95% prediction intervals of x|Is-k+1s (red, right y axes) for various window sizes k. The dashed line is k*. Panels correspond to those of Figs 2 and 3.

Real-time performance for emerging outbreaks

During an unfolding epidemic, renewal models can support response efforts by providing real-time estimates of infectious disease transmissibility and short-term forecasts of the expected incidence of cases [9]. Here we show how the APE metric can be used within this framework as a rapid and reliable real-time tool. We focus on the key problem of diagnosing and verifying the efficacy of implemented control measures over the course of an emerging outbreak [3]. Knowing whether an intervention is working or not can inform preparedness and resource allocation. Estimating whether Rs < 1 or Rs ≥ 1 is one simple means of assessing efficacy. Effective control measures lead to the former.

We investigate how the APE metric responds in real time to swift swings in reproduction number and quantify our results over 103 simulated epi-curves with t = 150 days. We focus on step-changes in Rs and examine how the successively optimal window length, denoted ks* when computed with data up to time s, responds. If ks* is sensitive to these changes with minimum delay then it is likely a dependable means of diagnosing control efficacy in real time. In previous sections we have been computing k*=kt*. We consider four models and present our main results in Figs 5 and 6.

Fig 5. Real-time APE sensitivity to increasing transmission.

Fig 5

We simulate 103 independent epi-curves under renewal models with sharply (A) increasing and (B) recovering epidemics. Top graphs give the true (green) and predicted (blue) incidence ranges, the middle ones provide estimates of Rs under the final kt* and the bottom graphs illustrate how successive ks* choices from APE vary across time and are sensitive to real-time elevations in transmission.

Fig 6. Real-time APE sensitivity to rapid epidemic control.

Fig 6

We simulate 103 independent epi-curves under renewal models with (A) one effective and (B) two partially effective interventions. Top graphs give the true (green) and predicted (blue) incidence ranges, the middle ones provide estimates of Rs under the final kt* and the bottom graphs illustrate how successive ks* choices from APE can reliably and rapidly detect the impact of real-time control actions.

In Fig 5 we examine cases of increasing transmission for (A) an outbreak that is uncontrolled and worsening (e.g. if the infectious disease acquires an additional route of spread) and (B) an ineffectively controlled epidemic with a late-stage elevation in Rs (e.g. if quarantine measures were relaxed too quickly). Fig 6 then investigates scenarios involving effective interventions where (A) rapid control is employed (e.g. if a wide-scale lockdown is enacted) and (B) the initial action is only partially effective and so later additional controls are needed (e.g. if flight restrictions were first applied but then further transport shutdowns or mobility restrictions were required).

In both figures, top graphs overlay true Is+1 and predicted x|Is+k-1s across time and middle ones show the best R^τ(s) estimates under the final kt*. We see that the APE-selected model properly distinguishes significant changes from stable periods in all scenarios. The model in Fig 5A is the most difficult to infer as the increasing Rs does not visibly change the shape of the incidence population curve. Bottom graphs show ks* sequentially across the ongoing epidemic. Here we observe how the APE metric uses data to justify its choices in real time. As data accumulate under a stable reproduction number APE increases k*. This makes sense as there is increasing support for stationary epidemic behaviour. This is especially obvious in Fig 6A.

However, on facing a step-change the APE immediately responds by drastically reducing ks* − both in scenarios where Rs is changing from above to below 1 and vice versa. This reduction is less visible in Fig 5A because the observed epi-curve is not as dramatically altered as in the other scenarios. This rapid response recommends the APE as a suitable and sensitive real-time diagnostic tool. We dissect the impact of change-times on the actual APE values at different k in Fig 7. There we find that when the epi-curve is reasonably stable there is not much difference between the performance at various window lengths and so the APE curves are neighbouring. In contrast, when a non-stationary change occurs there is a clear unravelling of the curves and potentially large gains to be made by optimising k. Failing to properly select k here would potentially mislead our understanding of the unfolding outbreak. This increases the impetus for formal k-selection metrics such as the APE.

Fig 7. APE scores in real time.

Fig 7

The successive (in time) APE scores for each window length, k, are shown in grey. The scores for the smallest and largest k are in cyan and magenta respectively. Graphs correspond to the models in Figs 5 and 6. Dashed lines are the change-times of each model. In (5A) the change-time does not significantly affect the epi-curve shape and so the APE scores are close together. In (5B), (6A) and (6B) the change-times notably alter the epi-curve shape so that the choice of k becomes critical to performance.

Window selection on empirical data

We test our APE approach on two well-studied empirical epidemic datasets: pandemic H1N1 influenza in Baltimore from 1918 [16] and SARS in Hong Kong from 2003 [17]. For each epidemic we extract the incidence curve, total infectiousness and generation time distributions from the EpiEstim R package [10] and apply the APE metric over 2kt2 with t as the last available incidence time point. We compare the one-step-ahead Is+1 prediction fidelity and the Rτ(s) estimation accuracy obtained from the renewal model under the APE-selected k* to that from the model used in [10], which recommended weekly windows (i.e. k = 7) after visually examining several window lengths.

Our main results are in Figs 810. We benchmarked our estimates against those directly provided by EpiEstim to confirm our implementation and restrict k > 1 to avoid model identifiability issues [11, 18]. Intriguingly, we find k* = min k = 2 for both datasets. This yields appreciably improved prediction fidelity (i.e. the coverage of observed incidence values by the 95% prediction intervals), relative to the weekly window choice, as seen in Fig 8. Between 8−10% of incidence data points are better covered by using k* over k = 7. This improvement is apparent in the right graphs of Figs 9A and 10A, where the k = 7 case produces stiffer incidence predictions that cannot properly reproduce the observed epi-curve.

Fig 8. Empirical prediction accuracy.

Fig 8

We compare the APE metric (dotted blue, left y axis) to the percentage of true incidence values, Is+1, which fall outside the 95% prediction intervals of x|Is-k+1s (dotted red, right y axis) across the window search space k. The dashed line gives k* (black) and k = 7 (grey). The top graph presents results for the influenza 1918 dataset, while the bottom one is for SARS 2003 data. We find that heuristic weekly windows lead to appreciably larger forecasting error than the APE selections.

Fig 10. Selection for SARS (2003).

Fig 10

Left graphs compare R^τ(s) estimates (blue with 95% confidence intervals) at optimal APE window length k* to weekly sliding windows (k = 7), which were recommended in [10]. Right graphs give corresponding one-step-ahead predictions x|Is-k+1s (blue with 95% prediction intervals). Dashed lines are the R = 1 threshold (left) and dots are the true incidence Is (right). Panel (A) directly uses the empirical SARS (2003) data [10] while (B) smooths outliers (with a 5-day moving average) in the data as in [16].

Fig 9. Selection for pandemic influenza (1918).

Fig 9

Left graphs compare R^τ(s) estimates (blue with 95% confidence intervals) at optimal APE window length k* to weekly sliding windows (k = 7), which were recommended in [10]. Right graphs give corresponding one-step-ahead predictions x|Is-k+1s (blue with 95% prediction intervals). Dashed lines are the R = 1 threshold (left) and dots are the true incidence Is (right). Panel (A) directly uses the empirical influenza (1918) data [10] while (B) smooths outliers in the data as in [16].

Weekly windows misjudge the SARS epidemic peak, predict a multimodal SARS incidence curve that is not reflected by the actual data and systematically underestimate influenza case counts when it matters most (i.e. around the high incidence phase). However, the smaller k* window results in noisier Rτ(s) estimates (left graphs of Figs 9A and 10A, which may be questionable. These rapidly fluctuating reproduction numbers likely motivated the adoption of weekly windows in previous analyses. While the APE-based estimates are indeed more uncertain (a consequence of shorter windows), we argue that they are formally justified by the available data, especially given the weaker predictive capacity of weekly windows. While the noisier Rτ(s) estimates do not mislead our understanding of the efficacy of implemented control measures (it is still clear that the influenza epidemic is only partially under control between 40 ≤ s ≤ 65 days before recovering, while the SARS outbreak is largely arrested from s > 50 days), they likely overestimate the peak transmissibility of these diseases.

We might suspect that certain artefacts of the data could resolve this issue, rendering a more believable combination of estimated reproduction number and predicted incidence. Particularly, the influenza data seem considerably more affected by the smaller look-back windows. These k* = 2 windows are needed to help predictions get close to the peak incidence values of the data. However, these peaks seem reminiscent of outliers and in the original analysis of [16] they were attributed to possible recollection bias in patients that were questioned. Removing these biases might be expected to lead to smoother APE-justified reproduction numbers. We test this hypothesis in Figs 9B and 10B.

There we find that applying simple 5-day moving average filters to the data, as done in [16] to ameliorate outliers, still does not support a k* > 2 in either scenario, though smoother estimates do result. Weekly windows are still too inflexible to properly predict this averaged incidence. While other signal processing techniques, such as using autocorrelated smoothing prior distributions instead of independent gamma ones over the reproduction numbers, could be applied to further investigate if k* > 2 is justifiable we consider this beyond the scope of this work and somewhat biologically unmotivated. Instead we conjecture that these results support the interesting alternative hypothesis that the epi-curve population data are not Poisson distributed. We expand on this conjecture in the subsequent section.

Discussion

Inferring the dynamics of the effective reproduction number, Rt, in real time is crucial for forecasting transmissibility and the efficacy of implemented control actions and for assessing the growth of an unfolding infectious disease outbreak [3]. Renewal models provide a popular platform for prospectively estimating these reproduction numbers, which can then be used to generate incidence projections [2, 7, 15]. However, the dependence of these estimates and predictions on the look-back window size, k, which determines the renewal model dimensionality, has never been formally investigated.

Previous methods for selecting k have generally been heuristic or ad-hoc [10]. Here we have devised and validated a new, rigorous, information-theoretic approach for optimising k to available incidence data. Our method was founded on deriving an analytical expression for the renewal model posterior predictive distribution. Integrating this into the general APE metric of [13] led to Eq (5), a simple, easily-computed yet theoretically justified metric for k-selection that balances estimation fidelity with prediction accuracy.

This APE approach not only accounts for parametric complexity and is formally linked to MDL (and Bayesian model selection − see the S1 Text), but it also has several desirable properties that render it suitable for handling the eccentricities of real-time infectious disease applications:

(a) Small outbreak sample size. Early-on in an epidemic data are scarce and uncertainty is large. The APE approach is valid and computable at all sample sizes at which the renewal model is identifiable and hence is applicable to emerging outbreaks [14]. Its emphasis on using the full predictive distribution and its Bayesian formulation allow it to properly account for large uncertainties and to handle different (prior) hypotheses about Rt, as the epidemic progresses from small (e.g. onset) to large (e.g. establishment) data regimes.

(b) Non-stationary transmission. Incidence time series can change rapidly across an unfolding outbreak and hence are not independent and identically distributed [7]. Instead they are sequential, autocorrelated and possess non-stationary (time-varying) statistics. The APE formally accounts for these properties. Leave-one-out cross validation (CV) is a popular model selection approach that is related to APE. It can be defined as CVs=1t-1-logP(Is+1|Is+1c), where Is+1c means that all incidence points except Is+1 are included. Comparing this to Eq. (S4) of the S1 Text we see that APE is a modified CV that is explicitly specialised for non-stationary, accumulating time-series.

(c) Lack of a ‘true model’. It is highly unlikely that the underlying Rt is truly described by one of the piecewise-constant functions assumed within renewal models. Unlike many model selection criteria, the APE does not require a true model to exist within the set being evaluated [12]. It is only interested in finding the model that best predicts the data and so emphasises accurate forecasting and data-justified complexity. However, should a true model exist, the APE is statistically consistent i.e. as data accumulate it will select the data-generating model with probability 1 [12].

While we have specialised our results to epidemiology, our method is widely-applicable. Several popular models in macroevolution, phylogeography and genetics, such as the skyline plot, structured coalescent and sequential Markovian coalescent, all possess piecewise-Poisson statistical formulations analogous to the renewal model [18]. As shown in [11], this leads to an almost plug-and-play usability of MDL-based methodology. Moreover, the renewal model itself can be adapted to other animal and plant ecology problems where the interest is to infer a demographic growth rate from a time-series of species samples [4].

We tested our method on both simulated and empirical datasets. We started by examining various distinct, time-varying reproduction number profiles in Figs 2 and 3, for which no true model existed. By comparing the APE-optimised k* against long and short window sizes, we found that not only does APE meaningfully balance Rt estimates to achieve good prediction capacity, but also that getting k wrong could promote strikingly different conclusions about the infection population dynamics from the same dataset. This behaviour held consistent for outbreaks with both small and large infected case numbers, stable and seasonal transmission and under different dynamics of control.

We then investigated several step-changing reproduction number examples in Figs 5 and 6 to better expose the underlying mechanics of our method, and to showcase its value as a real-time inference tool for emerging epidemics with strong non-stationary transmission. These examples featured rapid changes, also known as events in information theory, caused by effective and ineffective countermeasures [3]. In all cases our metric responded rapidly whilst maintaining reliable, real-time incidence predictions. Such rapid, event-triggered responses (as in Fig 7) are known to be time-efficient [19]. A key objective of short-term forecasting is the speedy diagnosis of intervention efficacy. Our results confirmed the practical value of our method towards this objective. Importantly, we found in Figs 5 and 6 that the APE metric computed successively and causally during an unfolding outbreak, ks*, can quickly and correctly inform on changes in transmission and growth of an epidemic in real time.

While our metric is promising, we caution that much work remains to be done. Standard Poisson renewal models, such as those we have considered here, make several limiting assumptions including that (i) all cases are detected (i.e. there is no significant sampling bias), (ii) the serial interval and generation time distribution coincide and do not change over the epidemic lifetime and (iii) that heterogeneities in transmission within the infected population have negligible effect [10]. Our analysis of H1N1 influenza (1918) and SARS (2003) flagged some of these concerns. Comparing the k* to previously recommended weekly windows from [10], as in Figs 9 and 10, led to some interesting revelations.

Our APE approach selected a notably shorter window (k* = 2 days) for both datasets. While this produced noisy Rs estimates that seemed less reliable than those obtained with weekly windows, the predicted incidence values were central to understanding this discrepancy. The weekly windows were considerably worse at forecasting the observed epi-curve, often systematically biased around the peak of the epidemic and sometimes predicting multimodal curves that were not reflected by the existing data. From a model selection perspective (among all EpiEstim-type models [10]), the shorter windows are therefore justified.

However, the noisy Rt estimates are still undesirable. Both datasets are known to potentially contain super-spreading heterogeneities and other biases that may lead to outliers [10, 16]. Moving-average filters were applied in [16] to remove these artefacts from the H1N1 data. We used the same technique to regularise both datasets and then re-applied the APE. Results remained consistent, suggesting that if a Poisson model is valid then k* = 2 is indeed the correct window choice. It may also be that the mean-variance equality of Poisson models is too restrictive for these datasets and so APE compensated for this inflexibility with short windows. A negative binomial renewal model where ItNB(κ,ΛtRtΛtRt+κ), with κ as a noise parameter, would relax this equality yet still embody the Euler-Lotka dynamics of the original model. The extension of APE to these generalised models will form an upcoming study.

Real-time, model-based epidemic forecasts are quickly becoming an integral prognostic for resource allocation and strategic intervention planning [20]. As global infectious disease threats elevate, model-supported predictions, which can inform decision making as an outbreak progresses, have become imperative. In the ongoing COVID-19 pandemic, for example, modelling is being extensively used to inform policy. However, much is still unknown about the fundamentals of epidemic prediction and this uncertainty has inspired some reluctance in public health applications [20]. As a result, recent studies have called for careful evaluation of the predictive capacity of models, and demonstrated the importance of having standard metrics to compare models and qualify their uncertainties [21, 22].

Here we have presented one such metric for use in real time. Short-term forecasts aid immediate decision making and are more actionable than long-term ones (which are less reliable) [20, 21]. While we focussed on renewal models, the general APE metric (see Fig 1) can select among diverse model types and facilitate comparisons of their short-term predictive capacity and reliability. Its direct use of posterior predictive distributions within an honest, causal framework allows proper and problem-specific inclusion of uncertainty [14]. Common metrics such as the mean absolute error lack this adaptation [21]. Given its information-theoretic links and demonstrated performance we hope that our APE-based method can serve as a rigorous benchmark for model-based forecasts and help refine ongoing estimates of transmission as outbreaks unfold.

Supporting information

S1 Text. Further details and more general definitions and equations under sub-headings: Epidemic renewal models and Prospective model selection.

(PDF)

Data Availability

All code and data are available at https://github.com/kpzoo/model-selection-for-epidemic-renewal-models.

Funding Statement

KVP and CAD acknowledge joint Centre funding from the UK Medical Research Council and Department for International Development under grant reference MR/R015600/1. CAD thanks the UK National Institute for Health Research Health Protection Research Unit (NIHR HPRU) in Modelling Methodology at Imperial College London in partnership with Public Health England (PHE) for funding (grant HPRU-2012–10080). The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.

References

  • 1. Anderson R and May R. Infectious diseases of humans: dynamics and control. Oxford University Press; 1991. [Google Scholar]
  • 2. WHO Ebola Response Team. Ebola virus disease in West Africa—the first 9 months of the epidemic and forward projections. N. Engl. J. Med. 2014;371(16):1481–95. 10.1056/NEJMoa1411100 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3. Cauchemez S, Boelle P, Thomas G, and Valleron A. Estimating in real time the efficacy of measures to control emerging communicable diseases. Am. J. Epidemiol. 2006;164(6):591–7. 10.1093/aje/kwj274 [DOI] [PubMed] [Google Scholar]
  • 4. Cortes E. Perspectives on the intrinsic rate of population growth. Methods Ecol. Evol. 2016;7:1136–45. 10.1111/2041-210X.12592 [DOI] [Google Scholar]
  • 5. Fraser C. Estimating individual and household reproduction numbers in an emerging epidemic. PLOS One. 2007;8:e758 10.1371/journal.pone.0000758 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6. Wallinga J and Lipsitch M. How generation intervals shape the relationship between growth rates and reproductive numbers. Proc. R. Soc. B. 2007;274:599–604. 10.1098/rspb.2006.3754 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7. Fraser C, Cummings D, Klinkenberg D, Burke D and Ferguson N. Influenza transmission in households during the 1918 pandemic. Am. J. Epidemiol. 2011;174(5):505–14. 10.1093/aje/kwr122 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8. Ferguson N, Cucunba Z, Dorigatti I, Nedjati-Gilani G, Donnelly C, Basanez M, et al. Countering the Zika epidemic in Latin America. Science. 2016;353(6297):353–4. 10.1126/science.aag0219 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9.S Bhatia, A Cori, K Parag, S Mishra, L Cooper, K Ainslie, et al. Short-term forecasts of COVID-19 deaths in multiple countries. MRC Centre for Global Infectious Disease Analysis COVID-19 planning tools 2020 [cited 8 June 2020]. Available from https://mrc-ide.github.io/covid19-short-term-forecasts/index.html
  • 10. Cori A, Ferguson N, Fraser C and Cauchemez S. A new framework and software to estimate time-varying reproduction numbers during epidemics. Am. J. Epidemiol. 2013;178(9):1505–12. 10.1093/aje/kwt133 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11. Parag K and Donnelly C. Adaptive estimation for epidemic renewal and phylogenetic skyline models. Syst. Biol. 2020;syaa035 10.1093/sysbio/syaa035 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12. Grunwald P. The Minimum Description Length Principle. The MIT Press; 2007. [Google Scholar]
  • 13. Rissanen J. Order estimation by accumulated prediction errors. J. Appl. Prob. 1986;23:55–61. 10.2307/3214342 [DOI] [Google Scholar]
  • 14. Wagenmakers E, Grunwald P, and Steyvers M. Accumulative prediction error and the selection of time series models. J. Math. Psychol. 2006;50:149–166. 10.1016/j.jmp.2006.01.004 [DOI] [Google Scholar]
  • 15. Nouvellet P, Cori A, Garske T, Blake I, Dorigatti I, Hinsley W, et al. A simple approach to measure transmissibility and forecast incidence. Epidemics. 2018;22:29–35. 10.1016/j.epidem.2017.02.012 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16. Frost W and Sydenstricker E. Influenza in Maryland: preliminary statistics of certain localities. Public Health Rep. 1919;34:491–504. 10.2307/457505619314660 [DOI] [Google Scholar]
  • 17. Leung G, Hedley A, Ho L, Chau P, Wong I, Thach T, et al. The epidemiology of severe acute respiratory syndrome in the 2003 Hong Kong epidemic: An analysis of all 1755 patients. Ann. Intern. Med. 2004;141(9):662–73. 10.7326/0003-4819-141-9-200411020-00006 [DOI] [PubMed] [Google Scholar]
  • 18. Parag K and Pybus O. Robust design for coalescent model inference. Syst. Biol. 2019;68(5):730–43. 10.1093/sysbio/syz008 [DOI] [PubMed] [Google Scholar]
  • 19. Parag K. On signalling and estimation limits for molecular birth-processes. J. Theor. Biol. 2019;480:262–73. 10.1016/j.jtbi.2019.07.007 [DOI] [PubMed] [Google Scholar]
  • 20. Viboud C and Vespignani A. The future of influenza forecasts. PNAS. 2019;116(8):2802–4. 10.1073/pnas.1822167116 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21. Funk S, Camacho A, Kucharski A, Lowe R, Eggo R and Edmunds W. Assessing the performance of real-time epidemic forecasts: A case study of Ebola in the western area region of Sierra Leone, 2014-15. PLOS Comput. Biol. 2019;15(2):e1006785 10.1371/journal.pcbi.1006785 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22. Reich N, Brooks L, Fox S, Kandula S, McGowan C, Moore E, et al. A collaborative multiyear, multimodel assessment of seasonal influenza forecasting in the United States. PNAS. 2019;116(8):3146–54. 10.1073/pnas.1812594116 [DOI] [PMC free article] [PubMed] [Google Scholar]
PLoS Comput Biol. doi: 10.1371/journal.pcbi.1007990.r001

Decision Letter 0

Matthew (Matt) Ferrari, Virginia E Pitzer

1 Mar 2020

Dear Dr Parag,

Thank you very much for submitting your manuscript "Using information theory to optimise epidemic models for real-time prediction and estimation" for consideration at PLOS Computational Biology. As with all papers reviewed by the journal, your manuscript was reviewed by members of the editorial board and by several independent reviewers. The reviewers appreciated the attention to an important topic. Based on the reviews, we are likely to accept this manuscript for publication, providing that you modify the manuscript according to the review recommendations.

I congratulate the author's on a very nice manuscript that was well received by the reviewers. Both reviewers were complimentary of the methods, though they raised a number of concerns about the presentation. I would encourage the authors to revisit the figures, as both reviewers found that they were challenging to interpret. Specifically:

1. Please increase the font size for figure labeling

2. If possible, consider limiting the number of scenarios presented in the multi-panel figures (particularly Fig 2). This could be reduced to a smaller number of exemplary scenarios (e.g. a,f,c,e where the latter two reflect scenarios exemplary of interventions)

3. Figure 1 is quite dense and I agree with R2 that breaking this up into panels that could be labelled, and this referenced in the legend, would help with interpretation.

4. consider using a color palettes that have greater contrasts.

Both reviewers comment on the tendency for k* to be estimated at the minimal (both reviewers) or maximal (R2) values. Please elaborate on this phenomenon.

Please consider the suggestion 1 of R2. If it is possible to clarify the applicability in real-time, I suspect that would appeal to readers.

Please address the minor comments of R2 and comments 1&2 of R1.

Please prepare and submit your revised manuscript within 30 days. If you anticipate any delay, please let us know the expected resubmission date by replying to this email. 

When you are ready to resubmit, please upload the following:

[1] A letter containing a detailed list of your responses to all review comments, and a description of the changes you have made in the manuscript. Please note while forming your response, if your article is accepted, you may have the opportunity to make the peer review history publicly available. The record will include editor decision letters (with reviews) and your responses to reviewer comments. If eligible, we will contact you to opt in or out

[2] Two versions of the revised manuscript: one with either highlights or tracked changes denoting where the text has been changed; the other a clean version (uploaded as the manuscript file).

Important additional instructions are given below your reviewer comments.

Thank you again for your submission to our journal. We hope that our editorial process has been constructive so far, and we welcome your feedback at any time. Please don't hesitate to contact us if you have any questions or comments.

Sincerely,

Matthew (Matt) Ferrari

Associate Editor

PLOS Computational Biology

Virginia Pitzer

Deputy Editor

PLOS Computational Biology

***********************

A link appears below if there are any accompanying review attachments. If you believe any reviews to be missing, please contact ploscompbiol@plos.org immediately:

[LINK]

I congratulate the author's on a very nice manuscript that was well received by the reviewers. Both reviewers were complimentary of the methods, though they raised a number of concerns about the presentation. I would encourage the authors to revisit the figures, as both reviewers found that they were challenging to interpret. Specifically:

1. Please increase the font size for figure labeling

2. If possible, consider limiting the number of scenarios presented in the multi-panel figures (particularly Fig 2). This could be reduced to a smaller number of exemplary scenarios (e.g. a,f,c,e where the latter two reflect scenarios exemplary of interventions)

3. Figure 1 is quite dense and I agree with R2 that breaking this up into panels that could be labelled, and this referenced in the legend, would help with interpretation.

4. consider using a color palettes that have greater contrasts.

Both reviewers comment on the tendency for k* to be estimated at the minimal (both reviewers) or maximal (R2) values. Please elaborate on this phenomenon.

Please consider the suggestion 1 of R2. If it is possible to clarify the applicability in real-time, I suspect that would appeal to readers.

Please address the minor comments of R2 and comments 1&2 of R1.

Reviewer's Responses to Questions

Comments to the Authors:

Please note here if the review is uploaded as an attachment.

Reviewer #1: This was an interesting paper that adds to the suite of methods for analysing case report data. In particular it provides a useful measure of the “window size” over which the reproductive ratio Rt should be calculated. As such, this could prove to be a powerful tool.

On the whole I found the paper to be extremely well written, and I only have a few minor comments.

1. In the abstract, I find the phrase “most succinctly describes” to be rather vague and confusing, could the authors seek a more informative description?

2. Page 9, the choice of a Gamma distribution for the posterior should be better motivated.

3. Figures: I found these hard to read. I realise that most people view on-line and therefore can greatly magnify figures, but a printed version is unreadable. In figure 7, I wonder if plotting Rt on a log-scale would help?

4. The paper ends on a whimper, with it being unclear if the k=2 found from real data is true or an artifact of a noisy system. I’m surprised that the authors didn’t test the method against simulations that were made sequentially noisier, this would seem an obvious test. I also feel that the speculation in the last paragraph of the results would sit far better in the discussion.

Reviewer #2: See attached

**********

Have all data underlying the figures and results presented in the manuscript been provided?

Large-scale datasets should be made available via a public repository as described in the PLOS Computational Biology data availability policy, and numerical data that underlies graphs or summary statistics should be provided in spreadsheet form as supporting information.

Reviewer #1: Yes

Reviewer #2: Yes

**********

PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

Figure Files:

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email us at figures@plos.org.

Data Requirements:

Please note that, as a condition of publication, PLOS' data policy requires that you make available all data used to draw the conclusions outlined in your manuscript. Data must be deposited in an appropriate repository, included within the body of the manuscript, or uploaded as supporting information. This includes all numerical values that were used to generate graphs, histograms etc.. For an example in PLOS Biology see here: http://www.plosbiology.org/article/info%3Adoi%2F10.1371%2Fjournal.pbio.1001908#s5.

Reproducibility:

To enhance the reproducibility of your results, PLOS recommends that you deposit laboratory protocols in protocols.io, where a protocol can be assigned its own identifier (DOI) such that it can be cited independently in the future. For instructions see http://journals.plos.org/ploscompbiol/s/submission-guidelines#loc-materials-and-methods

Attachment

Submitted filename: plos_renewal_review.docx

PLoS Comput Biol. doi: 10.1371/journal.pcbi.1007990.r003

Decision Letter 1

Matthew (Matt) Ferrari, Virginia E Pitzer

13 May 2020

Dear Dr Parag,

Thank you very much for submitting your manuscript "Using information theory to optimise epidemic models for real-time prediction and estimation" for consideration at PLOS Computational Biology. As with all papers reviewed by the journal, your manuscript was reviewed by members of the editorial board and by several independent reviewers. The reviewers appreciated the attention to an important topic. Based on the reviews, we are likely to accept this manuscript for publication, providing that you modify the manuscript according to the review recommendations.

I thank the authors for their careful revisions. I do not see the need to send this back to review, but I would ask the authors to correct the legend for Figure 8 before I recommend this for publication. I assume the "(solid blue, left y axis)" and "(dotted red, right y axis)" refer to the points in the figure, but it isn't clear if I am assuming correctly or if there are elements missing from the figure. With this clarified, I will submit a final recommendation of "accept".

Please prepare and submit your revised manuscript within 30 days. If you anticipate any delay, please let us know the expected resubmission date by replying to this email. 

When you are ready to resubmit, please upload the following:

[1] A letter containing a detailed list of your responses to all review comments, and a description of the changes you have made in the manuscript. Please note while forming your response, if your article is accepted, you may have the opportunity to make the peer review history publicly available. The record will include editor decision letters (with reviews) and your responses to reviewer comments. If eligible, we will contact you to opt in or out

[2] Two versions of the revised manuscript: one with either highlights or tracked changes denoting where the text has been changed; the other a clean version (uploaded as the manuscript file).

Important additional instructions are given below your reviewer comments.

Thank you again for your submission to our journal. We hope that our editorial process has been constructive so far, and we welcome your feedback at any time. Please don't hesitate to contact us if you have any questions or comments.

Sincerely,

Matthew (Matt) Ferrari

Associate Editor

PLOS Computational Biology

Virginia Pitzer

Deputy Editor

PLOS Computational Biology

***********************

A link appears below if there are any accompanying review attachments. If you believe any reviews to be missing, please contact ploscompbiol@plos.org immediately:

[LINK]

I thank the authors for their careful revisions. I do not see the need to send this back to review, but I would ask the authors to correct the legend for Figure 8 before I recommend this for publication. I assume the "(solid blue, left y axis)" and "(dotted red, right y axis)" refer to the points in the figure, but it isn't clear if I am assuming correctly or if there are elements missing from the figure. With this clarified, I will submit a final recommendation of "accept".

Figure Files:

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email us at figures@plos.org.

Data Requirements:

Please note that, as a condition of publication, PLOS' data policy requires that you make available all data used to draw the conclusions outlined in your manuscript. Data must be deposited in an appropriate repository, included within the body of the manuscript, or uploaded as supporting information. This includes all numerical values that were used to generate graphs, histograms etc.. For an example in PLOS Biology see here: http://www.plosbiology.org/article/info%3Adoi%2F10.1371%2Fjournal.pbio.1001908#s5.

Reproducibility:

To enhance the reproducibility of your results, PLOS recommends that you deposit laboratory protocols in protocols.io, where a protocol can be assigned its own identifier (DOI) such that it can be cited independently in the future. For instructions see http://journals.plos.org/ploscompbiol/s/submission-guidelines#loc-materials-and-methods

PLoS Comput Biol. doi: 10.1371/journal.pcbi.1007990.r005

Decision Letter 2

Matthew (Matt) Ferrari, Virginia E Pitzer

27 May 2020

Dear Dr Parag,

We are pleased to inform you that your manuscript 'Using information theory to optimise epidemic models for real-time prediction and estimation' has been provisionally accepted for publication in PLOS Computational Biology.

Before your manuscript can be formally accepted you will need to complete some formatting changes, which you will receive in a follow up email. A member of our team will be in touch with a set of requests.

Please note that your manuscript will not be scheduled for publication until you have made the required changes, so a swift response is appreciated.

IMPORTANT: The editorial review process is now complete. PLOS will only permit corrections to spelling, formatting or significant scientific errors from this point onwards. Requests for major changes, or any which affect the scientific understanding of your work, will cause delays to the publication date of your manuscript.

Should you, your institution's press office or the journal office choose to press release your paper, you will automatically be opted out of early publication. We ask that you notify us now if you or your institution is planning to press release the article. All press must be co-ordinated with PLOS.

Thank you again for supporting Open Access publishing; we are looking forward to publishing your work in PLOS Computational Biology. 

Best regards,

Matthew (Matt) Ferrari

Associate Editor

PLOS Computational Biology

Virginia Pitzer

Deputy Editor

PLOS Computational Biology

***********************************************************

I thank the authors for their attention the reviewer comments and commend them on a fine manuscript.

PLoS Comput Biol. doi: 10.1371/journal.pcbi.1007990.r006

Acceptance letter

Matthew (Matt) Ferrari, Virginia E Pitzer

24 Jun 2020

PCOMPBIOL-D-20-00012R2

Using information theory to optimise epidemic models for real-time prediction and estimation

Dear Dr Parag,

I am pleased to inform you that your manuscript has been formally accepted for publication in PLOS Computational Biology. Your manuscript is now with our production department and you will be notified of the publication date in due course.

The corresponding author will soon be receiving a typeset proof for review, to ensure errors have not been introduced during production. Please review the PDF proof of your manuscript carefully, as this is the last chance to correct any errors. Please note that major changes, or those which affect the scientific understanding of the work, will likely cause delays to the publication date of your manuscript.

Soon after your final files are uploaded, unless you have opted out, the early version of your manuscript will be published online. The date of the early version will be your article's publication date. The final article will be published to the same URL, and all versions of the paper will be accessible to readers.

Thank you again for supporting PLOS Computational Biology and open-access publishing. We are looking forward to publishing your work!

With kind regards,

Sarah Hammond

PLOS Computational Biology | Carlyle House, Carlyle Road, Cambridge CB4 3DN | United Kingdom ploscompbiol@plos.org | Phone +44 (0) 1223-442824 | ploscompbiol.org | @PLOSCompBiol

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    S1 Text. Further details and more general definitions and equations under sub-headings: Epidemic renewal models and Prospective model selection.

    (PDF)

    Attachment

    Submitted filename: plos_renewal_review.docx

    Attachment

    Submitted filename: plos_revisions.docx

    Attachment

    Submitted filename: plos_revisions.docx

    Data Availability Statement

    All code and data are available at https://github.com/kpzoo/model-selection-for-epidemic-renewal-models.


    Articles from PLoS Computational Biology are provided here courtesy of PLOS

    RESOURCES