Abstract
The number of new infections per day is a key quantity for effective epidemic management. It can be estimated relatively directly by testing of random population samples. Without such direct epidemiological measurement, other approaches are required to infer whether the number of new cases is likely to be increasing or decreasing: for example, estimating the pathogen‐effective reproduction number, R, using data gathered from the clinical response to the disease. For coronavirus disease 2019 (Covid‐19/SARS‐Cov‐2), such R estimation is heavily dependent on modelling assumptions, because the available clinical case data are opportunistic observational data subject to severe temporal confounding. Given this difficulty, it is useful to retrospectively reconstruct the time course of infections from the least compromised available data, using minimal prior assumptions. A Bayesian inverse problem approach applied to UK data on first‐wave Covid‐19 deaths and the disease duration distribution suggests that fatal infections were in decline before full UK lockdown (24 March 2020), and that fatal infections in Sweden started to decline only a day or two later. An analysis of UK data using the model of Flaxman et al. gives the same result under relaxation of its prior assumptions on R, suggesting an enhanced role for non‐pharmaceutical interventions short of full lockdown in the UK context. Similar patterns appear to have occurred in the subsequent two lockdowns.
Keywords: lockdown efficacy, lockdown impact, NPI, SARS‐CoV‐2, semiparametric
1. INTRODUCTION
Clinical data on the number of cases of coronavirus disease 2019 (Covid‐19/SARS‐CoV‐2) are subject to severe temporal confounding, as the rate of testing and criteria for testing have been changing rapidly on the same time scale as the infections, particularly in the early weeks and months of the epidemic. Because these are samples of convenience where the ascertainment fraction is changing and unknown, the data can clearly not be used to infer the actual number of infections. Neither, under normal circumstances, would statisticians recommend attempting to estimate the effective reproduction number of the pathogen from such data, because given the data problems, the estimates must necessarily be driven strongly by the modelling assumptions (see, e.g. Levine et al., 2001, §1.6 for general discussion of the problems with inference from non‐random samples). Indeed generically, it is often very difficult to infer epidemiological parameters from clinical data, without the results being informed as much by the prior beliefs encoded in the model as by the data (e.g. Wood et al., 2020). Much less problematic are estimates based on randomized surveillance testing, as now conducted in the United Kingdom by the office for national statistics (see Supporting Information for discussion of inferring incidence from testing data).
However, some clinical data directly measure the quantity of epidemiological interest. This is the case for deaths with Covid‐19 and for fatal disease duration. While not perfect, these data are less compromised than the data on cases. Deaths are reliably recorded and clinical grounds for suspecting Covid‐19 are relatively clear for fatal cases, although accurately attributing death to a single cause is clearly not always possible. Good records are also often kept for such cases, with the result that there are several published studies on fatal disease duration (Verity et al., 2020; Linton et al., 2020; Wu et al., 2020, see Section 2). Although only possible with a delay of some weeks, it is of interest to establish what these relatively high‐quality data imply about the time course of infections, without strong modelling assumptions.
Two types of daily death data are available. Daily reported deaths (e.g. Worldometer, 2020) typically show marked weekly fluctuations as a result of weekly patterns in reporting delays, and may exclude deaths in some locations (such as nursing homes). Registered death data, such as the ONS data in the United Kingdom (Office for National Statistics, 2020), contain deaths in all locations and record exact date of death. NHS (2020) publishes equivalent data for hospital deaths in England. The weekly cycle is less pronounced in these data, but their release is necessarily delayed relative to the daily reported deaths, although recent work partially overcomes this delay problem, by modelling the delays to enable ‘now‐casting’ of deaths by actual death date: see Stoner et al. (2020). The right column of Figure 2 shows ONS, NHS and Swedish daily deaths by date of death (without now‐casting).
The purpose of this paper is to show how a relatively straightforward statistical approach can be used to infer the fatal infection trajectory in the United Kingdom in a data‐driven way that makes the minimum of strong modelling assumptions. The approach is also applied to data from Sweden, the western European country offering the greatest policy contrast to the United Kingdom. Sweden never implemented full lockdown, sticking to less restrictive non‐pharmaceutical intervention (NPIs) (broadly aimed at ‘optimal mitigation’ rather than ‘suppression’ in the terms used by Walker et al., 2020, who projected around 40,000 deaths for this policy). Meaningful quantification of the aggregate strength of restrictions that are intrinsically multivariate is difficult, but in terms of their aggregate economic impact, Swedish GDP dropped by about 2.9% in 2020 compared to about 9.9% for the United Kingdom. Particular questions of interest are when the decline in fatal infections started in the United Kingdom and Sweden, whether UK infections were in substantial decline before full lockdown, whether the pathogen reproduction number was below 1 before lockdown, and how the timing of fatal incidence decline relates to the timing of the easing of lockdown.
Answers to these questions may contribute to judging the proportionality of lockdown measures in the UK context, where there is strong statistical evidence for very large preventable life loss being associated with economic deprivation, and of economic deprivation being increased by economic shocks. This evidence is reviewed in detail in Marmot et al. (2020). For example, the deprivation‐related life loss that the current UK population was due to suffer before the Covid crisis was 140–240 million life years (or 2–3.5 years per capita, see Marmot et al., 2020, figure 2.3, for example). The range depends on whether the life expectancy of the lower decile or the lower half of the deprivation distribution is used as the reference for achievable life‐expectancy. In examining the effects of the 2008 financial crisis and its aftermath, Marmot documents sharp reductions in life expectancy growth in the United Kingdom, which would imply a life loss burden in the 10s of millions of years. However, attribution of such reduction‐relative‐to‐trend is obviously very difficult. Less problematic is the 9 million life year loss implied by the increase in life expectancy gap between the more and less deprived halves of the UK population since 2008 (7 weeks per capita, see Marmot et al., 2020, figure 2.5, for example): given the evidence presented in the review, this is more difficult to attribute to causes unrelated to the 2008 economic shock. The Bank of England estimates the shock to the UK economy caused by the response to Covid‐19 to have been the largest for over 300 years, so there is a clear danger of substantial life loss being caused, given the historical data for the United Kingdom. For example, a feature of the 2008 crisis already repeated in 2020 is the reliance on a large programme of quantitative easing. Quantitative easing is credibly argued to directly increase economic inequality via mechanisms related to asset price inflation (e.g., Domanski et al., 2016; Fontan et al., 2016). There is some literature attributing some short‐term life‐saving to recessions, see, for example, Anon (2020), but the effects are modest relative to the long‐term effects reviewed by Marmot. For comparison with the above figures, the life loss that might have occurred from a minimally mitigated Covid‐19 epidemic appears to be in the region of 3 million life years (2.5 weeks per capita). This is based on Office for National Statistics (2019) lifetables, ONS Covid‐19 fatality by age data, a mid‐range infection fatality rate estimate of 0.006, a somewhat high herd immunity threshold of 0.7 and a 1‐year lower bound life expectancy adjustment for co‐morbidities based on Hanlon et al. (2020). It is broadly in line with the UK government estimates (Anon, 2020). Given that 9 million life years, associated with the substantially smaller economic shock of 2008, is not negligible relative to 3 million life years potentially losable to Covid‐19, there is obviously a delicate balance to be struck in the UK context, and evidence based on assumption light inference should probably play a role in shaping that balance. Another indicator of the difficulty of achieving the right balance is that the usual UK threshold for approving a pharmaceutical intervention is £30,000 per life year saved. On the basis of economic costs detailed in OBR (2020) and the preceding life loss figures, the NPIs used in the United Kingdom appear to have a cost per life year saved that is an order of magnitude higher than this (excess government borrowing is projected to peak at £660 billion in the OBR central scenario, for example). This discrepancy in willingness to pay may lead to a problem of opportunity cost, as the same money cannot be spent on preventing other life loss, such as that associated with economic hardship.
The remainder of the paper is structured as follows. Section 2 discusses the available information on the distribution of fatal disease durations, and how to combine it while adequately characterizing the associated uncertainty. Section 3 introduces a simple generalized additive model for direct modelling of the daily deaths trajectories, and shows how it can be extended to infer the trajectory of fatal infections, either directly or by inferring the trajectory of the pathogen effective reproduction number, R, in a simple epidemic model. Since the extensions are not standard models, and are relatively expensive to compute with using standard Bayesian software, Section 4 outlines methods allowing computationally efficient inference with the models. Section 5 presents the main results on infection trajectories, and also the estimation of R. Section 6 discusses possible problems with the approach, in particular examining whether smoothness assumptions could be leading to substantial bias in inferred timings. Replication code and data are provided in the Supporting Information.
2. FATAL DISEASE DURATION
Data on the incubation period from infection to onset of symptoms are analysed in many papers, for example, Lauer et al. (2020) found that the period is 2–11 days for 95% of people, with a median of 5.2 days. A meta‐analysis by McAloon et al. (2020) suggests a log‐normal distribution with log scale mean and standard deviation of 1.63 and 0.50. The uncertainty in this distribution is negligible in comparison to the uncertainty in the distribution of times from onset of symptoms to death discussed next.
Several studies estimate the distribution of time from onset of symptoms to death, while properly controlling for the right truncation in the fatal duration data. Verity et al. (2020) found that the distribution of time from onset of symptoms to death for fatal cases can be modelled by a gamma density with mean 17.8 and standard deviation 8.44, based on 24 patients from Wuhan. Wu et al. (2020) suggested a gamma density model with mean 20 and standard deviation 10 based on 41 patients from Wuhan. Linton et al. (2020) found that a log normal model offers a slightly better fit, and estimated a mean of 20.2 days and standard deviation of 11.6 days from 34 patients internationally. These distributions are shown in the left panel of Figure 1. A simple meta‐analysis approach was used to combine the models. Samples of the correct size were simulated from each model and a log normal model was estimated by maximum likelihood for the combined resulting sample (). A further log normal was also fitted (minimizing Kullback Leibler divergence) to the infection to death distribution implied by the fitted onset to death distribution and McAloon et al. (2020) infection to onset distribution (treated as independent). This process was repeated to generate replicate distributions. These replicate distributions were treated as draws from the distribution of infection to death distributions in subsequent analysis. One hundred such draws are shown in Figure 1. The log normal was chosen because the careful analysis of Linton et al. (2020) found it to be a better model than the gamma.
In addition, under strict conditions, I was able to access data on fatal disease durations for deaths occurring in English hospitals. Access to data with hospital acquired infections filtered out was not possible, so is was necessary to treat these data as a mixture of hospital‐ and community‐acquired infections, as detailed in the Supporting Information. The resulting inferred fatal disease duration distribution for community acquired infection is plotted in blue in Figure 1 (this figure appears in colour in the electronic version of this article, and any mention of colour refers to that version), and is consistent with the published studies.
3. MODELS
This section first introduces a simple generalized additive model for daily death trajectories, and then shows how this can be extended to directly infer the trajectory of fatal infections (fatal incidence), without having to assume any particular dynamic model for the epidemic. The resulting model is no longer a generalized additive model and is the model that this paper advocates using. Its structure is such that any method for inference with the model can also be used for inference with the dynamic model of Flaxman et al. (2020), with appropriate restriction of the incidence trajectory to one representable with that model. The Flaxman model is presented to allow comparison of the results from the infection trajectory model with the apparently contradictory results of Flaxman et al., but not to advocate its use.
3.1. Basic deaths series model
Let denote the deaths or reported deaths on day i, assumed to follow a negative binomial distribution with mean and variance . Let
(1) |
where f is a smooth function of time measured in days, and is a zero mean cyclic smooth function of day of the week, , set up so that , where , or 2 denotes order of derivative. represents the underlying log death rate, whereas describes the weekly variation about that rate. The functions f and can be represented using splines with associated smoothing penalties and . Hyper‐parameters λ and control the smoothness of the functions. The model is a straightforward generalized additive model and can be estimated as part of model fitting using a standard empirical Bayes approach as described in Wood (2017). The model provides a good fit to both the reported deaths and ONS data. As expected is greatly attenuated for the ONS data (it vanishes for Swedish exact death date data).
3.2. Infection trajectory model
To estimate the daily infection trajectory, the model is extended by expressing in terms of the time course of earlier infections. Let be the function describing the variation in the number of eventually fatal infections over time. Let be the square matrix such that if and 0 otherwise. π denotes an infection‐to‐death log normal density as discussed above. For the moment, its parameters, μ and σ2, are treated as fixed but this will be relaxed in Section 4.3. Given the continuity of the log normal, the given form for can be viewed as approximating an integral of π over each day, using the midpoint of the integrand – it is straightforward to approximate the integral more accurately, but given that π is originally estimated from durations discretized to whole days, any precision gain is illusory. If and , then , where is the expected number of deaths on day i. can be represented using a spline basis, again with a cubic spline penalty. Working on the log scale ensures that is positive, but is also appealing because it means that a cubic spline penalty on can be interpreted as a first derivative penalty , acting on the epidemiologists ‘intrinsic growth rate’, r. The final infection trajectory model is then obtained by simply substituting into (1). is rank deficient, so inferring can be viewed as an inverse problem: without regularization, multiple solutions that oscillate from day‐to‐day are possible. This ambiguity is removed by the smoothing penalty on .
3.3. Relaxed Flaxman model
In the time since this work was originally undertaken in late April 2020, the work of Flaxman et al. (2020) has appeared. Flaxman et al. make inferences about the reproduction number, R, and hence incidence rates, based on death trajectories and the fatal infection duration distribution of Verity et al. (2020), but do so by modelling the pathogen‐effective reproduction number within a simple epidemic ‘renewal model’. Flaxman et al. (2020) represent as a step function with steps allowed each time the government announced new containment interventions, and a sparsity prior promoting a small number of steps. In the notation of Flaxman et al., the expected number of infections each day (now total, rather than fatal) are denoted as . Given an initial c 1, the model is iterated from as follows:
(2) |
where N is the total initially susceptible population, and for . γ is the p.d.f. of a Gamma distribution with shape parameter 6.5 × 0.622 and scale parameter 0.62−2. The values multiplied by the assumed infection fatality rate give . The level of the infection fatality rate (IFR) only matters for the damping term in the first bracket of the expression for – this has almost no effect in practice, a mid‐range value of 0.006 was used. The original assumptions about can be relaxed by representing using a spline basis, with associated penalty as for the other models, while is also treated as a free parameter. Hence, in the infection trajectory model can simply be replaced by the Flaxman model with represented as a spline function. The model is otherwise unchanged. This model is presented only to allow comparison of this paper's results with those of Flaxman et al. (2020): its simple single‐compartment structure clearly does not meet the aim of inferring incidence with minimal assumptions.
4. METHODS
The infection trajectory and Flaxman renewal models are not standard models estimable with standard software. They can be implemented in Bayesian software, such as JAGS or STAN, but inference typically takes several hours if this is done. Dealing adequately with the uncertainty in the disease duration distribution multiplies this cost by one to two orders of magnitude. To avoid these problems, an empirical Bayes approach can be employed.
4.1. Basic inferential framework
Direct inference about (1) uses the empirical Bayes approach of Wood et al. (2016) in which the smooth functions are estimated by penalized likelihood maximisation (e.g. Green and Silverman, 1994), with the smoothing parameters and θ estimated by Laplace approximate marginal likelihood maximization. Writing for the combined vector of basis coefficients for f and , the penalized version of the log likelihood, , can be written
where : and are known constant positive semi‐definite matrices. Smoothing parameters, λ and , control the smoothness of f and . Let be the maximizer of the penalized log likelihood, and its negative Hessian at . Viewing the penalty as being induced by an improper Gaussian prior, , is also the maximum a posteriori (MAP) estimate of . Furthermore, in the large sample limit,
(3) |
Writing the density in (3) as , and the joint density of y and as , the Laplace approximation to the marginal likelihood for the smoothing parameters and θ is . Nested Newton iterations are used to find the values of maximizing and the corresponding (for details, see Wood et al., 2016).
Given (3), credible intervals for f are readily computed, but it is also straightforward to make inferences about when the peak in f occurs. Simply simulate replicate coefficient vectors from (3) and find the day of occurrence of the peak for each corresponding underlying death rate function, f.
4.2. Extension for the infection and Flaxman models
Although inference about (1) using the preceding framework requires little more than a call to the gam function in R package mgcv, its application to the other models, which are not generalized additive models, requires more work. For the model formulated in terms of , this requires expressions for the negative binomial deviance (or log likelihood) and its derivative vector and Hessian matrix w.r.t. the model coefficients.
First, consider the negative binomial deviance for observation i,
These need to be transformed into derivatives w.r.t. , as follows:
Writing and for the model matrices for the smooth terms and , we have where (here , division and multiplication are applied element‐wise to vectors), and . Then , whereas
For the given representation of
When using the relaxed Flaxman model, the preceding derivatives of have to be replaced with derivatives of w.r.t. the coefficients of the spline representing . Routine application of the chain rule to (2) gives corresponding iterations for the derivatives of , and hence , w.r.t these spline coefficients and .
Given these expressions and the penalties, can be obtained by Newton iteration, given smoothing parameters. To estimate smoothing parameters, the simplest approach is to fix the negative binomial θ at its estimate from model (1), and use Wood and Fasiolo (2017), alternating generalized Fellner Schall updates of the smoothing parameters with updates of given those smoothing parameters. This finds the smoothing parameters to approximately maximize the model marginal likelihood. The non‐linearity of the renewal equation model means that some effort is required to get non‐absurd starting values. I got these by a few minutes of experimentation with simple step functions for the initial to get death trajectories of roughly the shape and amplitude of the true trajectories (a close initial fit is not required: initial deviances two orders of magnitude greater than for the final fit were unproblematic).
Given θ and the smoothing parameters, the approximate posterior (3) could be used directly, or as the basis for the proposal distribution in a simple Metropolis Hastings sampler. A fairly efficient sampler results from alternating fixed proposals based on (3) with random walk proposals based on zero mean Gaussian steps with a shrunken version of the posterior covariance matrix. By this method, effective sample sizes >5000 for each coefficient took about 40 s computing on a low specification laptop. This was the approach used for the infection trajectory model. The results were indistinguishable from those produced at the cost of several hours of computing using JAGS (Plummer, 2003; Plummer et al., 2006) to simulate from the model posterior.
4.3. Disease duration distribution uncertainty
The methods so far perform inference conditional on fixed values for the parameters μ and σ2 of the log normal density describing the infection to death duration distribution. In reality, there is uncertainty about these parameters. To incorporate this uncertainty into the infection trajectory model, inference was run for each of the 100 sample distributions shown in grey in the right‐hand panel of Figure 1, and the resulting posterior samples were pooled, to give a sample from the unconditional posterior distribution of the model.
5. RESULTS
Figure 2 shows the results of applying the model to the Office for National Statistics daily Covid‐19 death data for the United Kingdom, to the NHS England hospital data and to the daily death data for Sweden from Folkhälsomyndigheten (2020). The results include the uncertainty about the disease duration distribution shape. ONS and NHS data are up to 27th June – including later data simply narrows the uncertainty, while making negligible difference to the overall conclusions. The most notable feature of the results is that fatal infections are inferred to be in substantial decline before full lockdown (the same result was obtained by this method in early May 2020, based on the first 50 days of reported daily death data). Sweden appears most likely to have peaked only 1 or 2 days later (barring some systematic difference in fatal disease durations for Sweden), having introduced NPIs well short of full lockdown. The results also emphasize the fact that the infection trajectory is not simply a time‐shifted version of the death trajectory (assuming that it was, might lead to unwarranted delay in easing lockdown, for example). The difference in timing and shape of the inferred profile between the ONS and NHS data reflects the fact that the latter contain care home data. There is an argument for preferring hospital data for inferring community fatal infections, in that the care home epidemic is now known to have special features with at least some of the infection not coming from normal community transmission. See, in particular, Comas‐Herrera et al. (2020) for a discussion of care home deaths internationally, including the United Kingdom. In addition, in the United Kingdom, care home deaths were often attributed to Covid‐19 without a test, especially after death certification guidelines were changed to encourage reporting of suspected, rather than confirmed, Covid‐19 deaths. The care home data therefore have some under‐reporting of Covid deaths, followed by over‐reporting (the signal of this is visible in ONS data in the change in non‐Covid pneumonia deaths being reported, relative to normal, for example).
Taken together the results for the United Kingdom and Sweden raise the questions of firstly whether full lockdown was necessary to bring infections under control, or whether more limited measures might have been effective, and secondly whether the several month duration of full lockdown was appropriate. These emphasize the desirability of statistically well‐founded direct measurement of epidemic size through randomized testing. Had such testing being carried out leading up to lockdown it would have been clearer if the measures preceding lockdown (see Figures 2 and 3) were working, or whether stronger restrictions were needed. Similarly, such testing might have given earlier indication of when lockdown could be eased. Instead, management was reliant on a complex modeling synthesis of expert judgement and problematic clinical case data. Less statistically problematic reconstructions, like the one presented here, are clearly only possible weeks after the fact. Note that although it is natural to interpret these fatal infection trajectories as proportional to the overall infection trajectories, that will only be the case if the infection fatality rate is constant over time. There is evidence for improvements in hospital care from late March onwards that suggest that this is might not be the case (see Dennis et al., 2021). The Supporting Information includes a sensitivity analysis of this issue: it has the potential to right shift the peak incidence by up to a day and to lead to somewhat less rapid decay of the incidence trajectory.
5.1. Inferring R
Much public debate has focused on the effective reproduction number, R, and in theory, it is possible for a decline in the rate of infections to be only temporary as a result of R dropping but remaining above one. Could it be that the declines in seen before lockdown were of this short‐term type, and that renewed increase would therefore have occurred without full lockdown? The answer appears to be no. R is all but impossible to measure directly, so inference about it requires assumption of an epidemic model. However, given an epidemic model, it can be directly inferred from the reconstructed infection profile. For example, consider a simple susceptible exposed infectious recovered (SEIR) model: , , (here is the rate of recovery or progression to serious disease). is a direct estimate of (to within a constant of proportionality), so by solving
(from 0 initial conditions) the direct estimate is readily computed (any constant of proportionality cancels in R). A different epidemic model could be used here of course: see Diekmann et al. (1990) for calculation of R in general from a model. Figure 3 shows the results using for the English hospital data for plausible values of average time to infectivity of days and mean duration of infectiousness of days, along with sensitivity analysis for these values. The credible intervals shown include the uncertainty about the fatal disease duration distribution. R appears to be below 1 before full lockdown.
A useful feature of the R estimates is to emphasize that the analysis in this paper in no way suggests that lockdown did not have an effect on transmission. Even if R was below 1 before lockdown, full lockdown can only have reduced it further, and the estimates in Figure 3 are obviously consistent with this. Note, however, that the recovery in R after the post‐lockdown dip is to be expected, given the simple fact that R is the number of new infections created per infection, averaged over the population of infections, not the population of people. Broadly speaking, at lockdown, the population of people, and infections, was split into the locked down population, where infections could create few new infections, and the ‘unlocked’ population where the reproductive rate of the pathogen was higher (assuming lockdown had an effect). An initial average over all infections is then dominated by those infections in the locked down population, giving a low R (especially once the possibilities for infecting locked down household members have been exhausted). As the infections in the locked down population die out, the proportion of all infections that are in the unlocked population must increase – so that an average over all infections must yield a higher R again.
5.2. The Flaxman model
As noted above in Section 3, Flaxman et al. (2020) also analysed death trajectories, using a simple epidemic model, but came to conclusions apparently contradicting Figure 3. They concluded that only after full lockdown did R drop below 1, and that fatal infections continued to increase up until the eve of full lockdown. Flaxman et al. (2020) used the Verity et al. (2020) fatal disease duration distribution, so the difference in results does not lie there. To describe the epidemic dynamics, Flaxman et al. use the simple single compartment discrete renewal model (2). Within that model, they assume that R is constant between the imposition of interventions, but can undergo a step change at each intervention: the steps are free model parameters. This model for R is quite restrictive. In particular, it does not allow R to change after lockdown, despite the fact that at lockdown, the population has been stratified in a way that the renewal model does not represent, so that some compensating flexibility in R is likely to be required to avoid modelling artefacts. At the same time, the model is rather underdetermined preceding lockdown, because of the frequent intervention changes. This indeterminacy in the model is addressed by using a sparsity promoting prior on the step changes in R, which favours few larger changes, rather than several smaller changes (see the supplementary material for Flaxman et al. for a description of this prior). When using the model to simultaneously model multiple European countries, there is a further assumption that the intervention effects are the same for all countries (despite the different order of their implementation) and that only the lockdown effect varies between countries. It seems likely to be difficult to pick up effects of the interventions preceding lockdown from such a model structure.
A relaxed version of the Flaxman model in which is a continuous function is described in Section 3. The results from using this model for inference using the NHS hospital data are shown in Figure 4. The relaxation of the assumptions on R brings the results (for the United Kingdom) into alignment with those in the rest of this paper, and into broad consistency with developments later in the year, which are otherwise difficult to square with Flaxman et al. (2020).
5.3. Later infection waves
Although the initial motivation for this work was to provide reasonably timely analysis for the first wave, based on the limited data available by May 2020, the methods scale readily to the much longer data runs available by early 2021. The only change is that it makes sense to use an adaptive smoother (see, e.g. Wood, 2017, §5.3.5) for , in which the degree of smoothness is allowed to vary with time. The longer data runs make it feasible to estimate the multiple smoothing parameters that this entails. Using an adaptive smooth guards against artefacts driven by the smoothness that is appropriate on average, for all the data, not being appropriate at times of rapid change.
The results of this application are shown in Figure 5. Note that likely changes in infection fatality rate as a result of improved hospital treatment mean that the relative sizes of the fatal infection incidence curves in the first and subsequent waves cannot be interpreted as reflecting the relative sizes of total incidence (the later incidence curves would need to be scaled up somewhat). Causal over‐interpretation of the R curves should be avoided, not least because there is no reason to expect Covid‐19 not to display the seasonality in transmission common to other respiratory illnesses. However, the results are obviously inconsistent with full lockdowns having caused , because cause should not happen after effect. Further, the drop in R seen after the initial NPIs were introduced, but before full lockdown, does seem consistent with the levels of R later achieved while measures short of lockdown were in place. The interesting feature of R apparently increasing from quite early in the second lockdown might relate to the spread of the new variant, but, of course, also occurs at a time when respiratory infections generally start to increase. Likewise, the further increase until mid‐December may well be due to the new variant, but increased activity in the run up to Christmas is also likely to be a factor – incidence appears to peak over the Christmas to New Year period. Vaccine rollout seems virtually certain to be a major factor in pushing down R and fatal incidence from December. The vaccine has been given to those most at risk first, so the constant IFR assumption required to interpret fatal incidence as proportional to total incidence obviously no longer holds. This further implies that the inferred R is in some sense only the R relevant to the ‘at serious risk’ population. Of course, it could be argued that for epidemic management purposes, the fatal incidence and the corresponding R are of primary interest.
Interestingly, the pattern observed at the second lockdown and in the preceding months is consistent with the results reported by Knock et al. (2020) who analysed regionally stratified death, hospital occupancy and testing data for 2020 up until December, using a highly detailed age structured SEIR with added health service compartments. The entire trajectory up until December is also consistent with the results of Wood and Wit (2021), who reimplemented the Knock et al. model, but removed some of its very strong modelling assumptions around the first lockdown.
6. MODEL CHECKING
Although standard residual checks indicate no problem with the model from the point of view of statistical fit, there are three issues which could potentially undermine the results, and a further issue relating to interpretation.
The first relates to the infection to death interval distribution and the fact that the death data contain an unknown proportion of patients whose infection was hospital acquired. These patients are likely to have had shorter disease durations, because they were already sufficiently unwell or frail to be in hospital. This paper has inferred when the fatal infections would have occurred if they were all community generated, because it is the community infections that are of interest with respect to the effects of lockdown, social distancing, and so on. Without knowing even the proportion of deaths from hospital acquired infection, it is anyway not possible to do otherwise.
The presence of hospital infections in the death data will bias inference about the dynamics of community fatal infections if it substantially changes the shape of the deaths profile, relative to what would have occurred without hospital infection. Broadly, if the trajectory of hospital acquired infection deaths peaked earlier than the overall trajectory, then the community infection peak will be estimated to be earlier than it should be (because the true community infection death peak is then later). Conversely, if the hospital acquired infection deaths peaked later, then the community infection peak will be estimated as being later than it should be. The degree of bias will depend on the proportion of hospital acquired infections and the degree of mismatch in timings. It is difficult to judge which alternative is more likely: standard epidemiological modelling assumptions would imply that the more community acquired cases are hospitalized the more hospital infections would occur and that hospital infections will lag community cases. But against this, hospital acquired fatal disease durations are likely to contain a higher proportion of shorter durations. In any case, the proportion of hospital acquired infections in the death series would have to be quite high for the issue to substantially modify the conclusions.
The second issue is that age dependency in the duration distribution coupled with shifts in the age structure of deaths over time could also be problematic. However, as documented in the Supporting Information, the data for England and Wales show remarkably little variation in the age structure of Covid‐19 fatalities over the course of 2020, whereas analysis of English hospital data apparently shows little evidence for age dependence in the disease duration distribution.
The third issue is whether the smoothing penalty on would lead to systematic mis‐timing of the estimated peak under the scenario of a very asymmetric peak in the true infection profile around lockdown. To investigate this, data were simulated from a model in which the underlying infection rate increased geometrically, doubling every 3 days until lockdown, when the rate dropped immediately to 0.2 of its peak value, shrinking thereafter by 5% per day. Fatal infections were simulated as Poisson deviates with the given underlying rate. This model is an extreme scenario, in which measures prior to full lockdown had no effect, and the effect of lockdown was instant, as if the locked down population (i.e. those not in essential work) had isolated alone, rather than increasing their contact with members of their household while drastically reducing it with everyone else. However, it is the scenario implicit in much public discussion in the United Kingdom, at least at the time that this work was originally conducted. Under this scenario, the method does indeed tend to incorrectly estimate the infection peak as 2–3 days before lockdown, rather than the day before, as it struggles to accommodate the drop.
The naive approach to this issue is to introduce a parameter at lockdown representing an instantaneous drop in infections. However, doing so introduces a very strong structural assumption into the model, undermining the aim of avoiding strong assumptions. This approach also has the serious side effect of introducing non‐parametric smoothing boundary effects on both sides of the break. These boundary effects severely compromise inference in the most interesting region of the infection profile, while simultaneously increasing the importance of the structural assumption at the expense of the data. Indeed, when such a model is built, it estimates a large drop even from data simulated from a smooth infection profile. It also estimates such a drop if the drop's location is moved (for simulated or real data).
A better approach is to use a smooth time dilation to relax, but not eliminate the model smoothness assumptions in the vicinity of lockdown. The dilation is made sufficient that the model can accurately capture the extreme scenario in the simulation, but without imposing a break and boundary effects. In particular, and its smoothing penalty are computed with respect to a version of time which makes the day before, of and after lockdown count as 3.5, 6 and 3.5 days, respectively. Obviously, regular un‐dilated time is used for mapping infections to deaths. For the extreme simulation, the model then correctly gives most posterior probability to the day before lockdown as the peak. In contrast, the same model for the real data has very low probability of the peak being the day before lockdown rather than earlier.
Figure 6 shows the results from fitting the time‐dilated model to the extreme simulation scenario and to the NHS England hospital data. Even this model, deliberately modified to favour a very abrupt change at lockdown, suggests that infections started to decline before lockdown, with the most likely day for the peak only 1 day later than with the un‐dilated model. The Supporting Information includes similar checks for the Flaxman et al. (2020) model, with similar conclusions.
Finally, interpretation of the fatal incidence trajectories as proportional to the overall incidence trajectories rests on an assumption that the infection fatality rate is constant over time. There is evidence that the hospitalized case fatality rate declined in the 2 months or so after the peak of the first wave of infections (Dennis et al., 2021), with this effect not explicable by any detectable change in patient characteristics. However, on the ground changes in the severity threshold for admission would be very difficult to detect, seem likely at times when some hospitals were at or near capacity, and could also contribute to such a pattern. The Supporting Information includes a check of the impact that the reported improvements would have on the shape of inferred overall incidence. The peak incidence could be shifted by as much as a day later, and there would be a somewhat slower decline in incidence relative to the results plotted in Figure 2.
7. DISCUSSION
This paper does not prove that the peak in fatal infections in the United Kingdom preceded the first full lockdown by several days. Indeed, the failure to undertake the sampling that could have gathered data to directly measure infections early in the epidemic means that it will never be possible to be certain about timings then, given the substantial biases in clinical data other than deaths and fatal disease duration. What the results show is that, in the absence of strong assumptions, the currently most reliable openly available data strongly suggest that the decline in infections in the United Kingdom began before the first full lockdown, suggesting that the measures preceding lockdown may have been sufficient to bring the epidemic under control, and that community infections, unlike deaths, were probably at a low level well before the first lockdown was eased. Such a scenario would be consistent with the infection profile in Sweden, which began its decline in fatal infections shortly after the United Kingdom, but did so on the basis of measures well short of full lockdown.
The analysis does not in itself say what would have happened without full lockdown, and must obviously be weighed against other evidence. No currently available analysis will conclusively determine what would have happened without full lockdown, and the state of the art in causal inference is obviously a very long way from being able to answer this question. Models based on approximations to the mechanisms of epidemic transmission do not allow reliable answers to these causal questions either. This is particularly so given the paucity of data with which to validate their component assumptions – a paucity that only grows more acute as more detail is included in the models. These are not weather or climate models, based on the bulk properties of enormous numbers of physically well‐understood interactions of simple molecules, tested and refined against huge quantities of carefully measured calibration data collected worldwide over decades. Rather they are best working approximations constructed by experts given the limited information that could be rapidly assembled in a matter of months, and subject to all the uncertainty this implies. A model does not become a valid basis for casual inference merely by being described as mechanistic. As the above reanalysis using the Flaxman model serves to emphasize: the inclusion of model structure aiming to represent mechanism is no guarantee of improved statistical inference, and certainly not a justification for treating inference with mechanism‐based models as causal.
In the time since this work was first undertaken, other low assumption analyses have appeared, in particular looking for the coincidence of NPI introductions and changepoints in incidence, for example, in Germany and Spain. The results of this paper are in some alignment with such analyses for Germany (Wieland, 2020; Küchenhoff et al., 2021), which also suggest that a decline in incidence preceded the first full lockdown. Both are based on case data, which are problematic even in Germany which had mass (but not randomized) testing in place from the start of the epidemic. However, it seems likely that the biases in case data would lead to the start of decline in incidence being estimated as later than it really was, rather than earlier, so the qualitative conclusion is likely to be robust. In Spain, Santamaría and Hortal (2020) also identify substantial changes in rate of change of incidence before Spanish lockdown based on death series, but not sufficient to suggest a decline in incidence before lockdown. Based on pre‐print versions of the current paper, a number of researchers have also attempted to employ the basic idea of dynamic model‐free inference about incidence profiles, but via a simplified method. This method tries to impute date of infection by subtracting a random draw from the fatal duration distribution from each deceased patient's death date. This process is replicated to obtain an expected incidence profile. The method is invalid, as duration of disease is not independent of time of death, and it will tend to incorrectly show much less steep, or no, decline before lockdown. See the Supporting Information for a full discussion.
The results of applying the method to data up to mid‐February 2021 provide a picture rather consistent with the results for the first lockdown. In particular, the results preceding the first lockdown appear consistent with how the epidemic progressed under later restrictions short of lockdown. This is not the case for the published analyses suggesting high R and surging incidence on the eve of the first lockdown. The fact that school re‐opening does not appear to be followed by an increase in R is interesting: whether it relates to people deciding to keep school children apart from the vulnerable, which is anecdotally plausible, or to other factors, is unclear. While tempting, it is difficult to interpret the later patterns in terms of the new, apparently more infectious, variant that emerged in late 2020: there is confounding with seasonality of transmission, behavioural changes around the end of year holidays and with the roll out of effective vaccines from late December onwards. Greater clarity on these issues may emerge in future, particularly if the UK ONS Covid surveillance data eventually becomes public in raw form.
Supporting information
ACKNOWLEDGEMENTS
I am grateful to three referees, the editor and associate editor who provided several useful suggestions and references as well as helpful discussion relating to the analysis using the Flaxman model, and to Robert Verity (see Supporting Information). I also thank Jason Matthiopoulos, Dirk Husmeier, Anna Jewell, Peter Green, Ernst Wit, Nicole Augustin, Jonty Rougier, John Coggon and Dan Cookson, for various suggestions and discussions of aspects of the paper.
Wood, S.N. (2021) Inferring UK COVID‐19 fatal infection trajectories from daily mortality data: Were infections already in decline before the UK lockdowns? Biometrics, 1–14. 10.1111/biom.13462
DATA AVAILABILITY STATEMENT
The data that support the findings in this paper are available in the Supporting Information of this article.
REFERENCES
- Anon (2020) Direct and indirect impacts of COVID‐19 on excess deaths and morbidity . Technical report, Department of Health and Social Care, Office for National Statistics, Government Actuary's Department and Home Office.
- Comas‐Herrera, A. , Zalakaín, J. , Litwin, C. , Hsu, A.T. , Lane, N. and Fernández, J.‐L. (2020) Mortality associated with COVID‐19 outbreaks in care homes: early international evidence . International Report, LTCcovid. org, International Long‐Term Care Policy Network.
- Dennis, J.M. , McGovern, A.P. , Vollmer, S.J. and Mateen, B.A. (2021) Improving survival of critical care patients with coronavirus disease 2019 in England: a National Cohort Study, March to June 2020. Critical Care Medicine, 49, 209. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Diekmann, O. , Heesterbeek, J.A.P. and Metz, J.A. (1990) On the definition and the computation of the basic reproduction ratio R0 in models for infectious diseases in heterogeneous populations. Journal of Mathematical Biology, 28, 365–382. [DOI] [PubMed] [Google Scholar]
- Domanski, D. , Scatigna, M. and Zabai, A. (2016) Wealth inequality and monetary policy. BIS Quarterly Review , March.
- Flaxman, S. , Mishra, S. , Gandy, A. , Unwin, H.J.T. , Mellan, T.A. , Coupland, H. et al. (2020) Estimating the effects of non‐pharmaceutical interventions on COVID‐19 in Europe. Nature, 584, 257–261. [DOI] [PubMed] [Google Scholar]
- Folkhälsomyndigheten. (2020) experience.arcgis.com/experience/09f821667ce64bf7be6f9f87457ed9aa.
- Fontan, C. , Claveau, F. and Dietsch, P. (2016) Central banking and inequalities: taking off the blinders. Politics, Philosophy & Economics, 15, 319–357. [Google Scholar]
- Green, P.J. and Silverman, B.W. (1994) Nonparametric Regression and Generalized Linear Models. Boca Raton, FL: Chapman & Hall. [Google Scholar]
- Hanlon, P. , Chadwick, F. , Shah, A. , Wood, R. , Minton, J. , McCartney, G. et al. (2020) COVID‐19 – exploring the implications of long‐term condition type and extent of multimorbidity on years of life lost: a modelling study. https://wellcomeopenresearch.org/articles/5‐75. 7th May 2020. [DOI] [PMC free article] [PubMed]
- Knock, E.S. , Whittles, L.K. , Lees, J.A. , Perez Guzman, P.N. , Verity, R. , Fitzjohn, R.G. et al. (2020) The 2020 SARS‐CoV‐2 epidemic in England: key epidemiological drivers and impact of interventions . Report 41, Imperial College London. [DOI] [PMC free article] [PubMed]
- Küchenhoff, H. , Guenther, F. , Höhle, M. and Bender, A. (2021) Analysis of the early Covid‐19 epidemic curve in Germany by regression models with change points. medRxiv. Epidemiology & Infection, 149, e68, 1–7. 10.1017/S0950268821000558. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Lauer, S.A. , Grantz, K.H. , Bi, Q. , Jones, F.K. , Zheng, Q. , Meredith, H.R. et al. (2020) The incubation period of coronavirus disease 2019 (COVID‐19) from publicly reported confirmed cases: estimation and application. Annals Of Internal Medicine, 172, 577–582. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Levine, D.M. , Ramsay, P.P. and Smidt, R.K. (2001) Applied Statistics for Engineers and Scientists. Upper Saddle River, NJ: Prentice Hall. [Google Scholar]
- Linton, N.M. , Kobayashi, T. , Yang, Y. , Hayashi, K. , Akhmetzhanov, A.R. , Jung, S.‐M. et al. (2020) Incubation period and other epidemiological characteristics of 2019 novel coronavirus infections with right truncation: a statistical analysis of publicly available case data. Journal of Clinical Medicine, 9, 538. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Marmot, M. , Allen, J. , Boyce, T. , Goldblatt, P. and Morrison, J. (2020) Health Equity in England: The Marmot Review 10 Years On. London, UK: The Health Foundation. [Google Scholar]
- McAloon, C. , Collins, Á. , Hunt, K. , Barber, A. , Byrne, A. W. , Butler, F. et al. (2020) Incubation period of COVID‐19: a rapid systematic review and meta‐analysis of observational research. BMJ Open, 10, e039652. [DOI] [PMC free article] [PubMed] [Google Scholar]
- NHS (2020) www.england.nhs.uk/statistics/statistical‐work‐areas/covid‐19‐daily‐deaths/. 3rd July 2020 and 26th Feb 2021.
- OBR (2020) Office for budget responsibility fiscal sustainability report, July 2020.
- Office for National Statistics (2019) National life tables, UK: 2016 to 2018.
- Office for National Statistics (2020) Deaths registered weekly in England and Wales.
- Plummer, M (2003) JAGS: a program for analysis of Bayesian graphical models using Gibbs sampling. In Proceedings of the 3rd International Workshop on Distributed Statistical Computing (DSC 2003) , pp. 20–22.
- Plummer, M. , Best, N. , Cowles, K. and Vines, K. (2006) coda: convergence diagnosis and output analysis for MCMC. R News, 6, 7–11. [Google Scholar]
- Santamaría, L. and Hortal, J. (2020) Chasing the ghost of infection past: identifying thresholds of change during the COVID‐19 infection in Spain. Epidemiology & Infection, 148, e282, 20–12 10.1017/S0950268820002782. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Stoner, O. , Economou, T. and Halliday, A. (2020) A powerful modelling framework for nowcasting and forecasting COVID‐19 and other diseases. arXiv preprint arXiv:1912.05965.
- Verity, R. , Okell, L. C. , Dorigatti, I. , Winskill, P. , Whittaker, C. , Imai, N. et al. (2020) Estimates of the severity of coronavirus disease 2019: a model‐based analysis. The Lancet Infectious Diseases, 20, 669–677. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Walker, P. , Whittaker, C. , Watson, O. , Baguelin, M. , Ainslie, K. , Bhatia, S. et al. (2020) The global impact of COVID‐19 and strategies for mitigation and suppression. Report 12, Imperial College London.
- Wieland, T. (2020) A phenomenological approach to assessing the effectiveness of COVID‐19 related nonpharmaceutical interventions in Germany. Safety Science, 131, 104924. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Wood, S. N. (2017) Generalized Additive Models: An Introduction with R, 2nd edition. Boca Raton, FL: CRC Press. [Google Scholar]
- Wood, S.N. and Fasiolo, M. (2017) A generalized Fellner‐Schall method for smoothing parameter optimization with application to Tweedie location, scale and shape models. Biometrics, 73, 1071–1081. [DOI] [PubMed] [Google Scholar]
- Wood, S.N. , Pya, N. and Säfken, B. (2016) Smoothing parameter and model selection for general smooth models (with discussion). Journal of the American Statistical Association, 111, 1548–1575. [Google Scholar]
- Wood, S.N. and Wit, E. C. (2021) Was before the English lockdowns? On modelling mechanistic detail, causality and inference about Covid‐19. medRxiv. Cold Spring Harbor Laboratory Press. 10.1101/2021.02.03.21251112. [DOI] [PMC free article] [PubMed]
- Wood, S.N. , Wit, E.C. , Fasiolo, M. and Green, P.J. (2020) COVID‐19 and the difficulty of inferring epidemiological parameters from clinical data. The Lancet Infectious Diseases, 21, 27–28. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Worldometer (2020) www.worldometers.info/coronavirus/.
- Wu, J.T. , Leung, K. , Bushman, M. , Kishore, N. , Niehus, R. , de Salazar, P.M. et al. (2020) Estimating clinical severity of COVID‐19 from the transmission dynamics in Wuhan, China. Nature Medicine, 26, 506–510. [DOI] [PMC free article] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Supplementary Materials
Data Availability Statement
The data that support the findings in this paper are available in the Supporting Information of this article.