Skip to main content
Philosophical transactions. Series A, Mathematical, physical, and engineering sciences logoLink to Philosophical transactions. Series A, Mathematical, physical, and engineering sciences
. 2021 Nov 22;380(2214):20210117. doi: 10.1098/rsta.2021.0117

On some fundamental challenges in monitoring epidemics

Vaiva Vasiliauskaite 1, Nino Antulov-Fantulin 1,, Dirk Helbing 1,2
PMCID: PMC8607144  PMID: 34802270

Abstract

Epidemic models often reflect characteristic features of infectious spreading processes by coupled nonlinear differential equations considering different states of health (such as susceptible, infectious or recovered). This compartmental modelling approach, however, delivers an incomplete picture of the dynamics of epidemics, as it neglects stochastic and network effects, and the role of the measurement process, on which the estimation of epidemiological parameters and incidence values relies. In order to study the related issues, we combine established epidemiological spreading models with a measurement model of the testing process, considering the problems of false positives and false negatives as well as biased sampling. Studying a model-generated ground truth in conjunction with simulated observation processes (virtual measurements) allows one to gain insights into the fundamental limitations of purely data-driven methods when assessing the epidemic situation. We conclude that epidemic monitoring, simulation, and forecasting are wicked problems, as applying a conventional data-driven approach to a complex system with nonlinear dynamics, network effects and uncertainty can be misleading. Nevertheless, some of the errors can be corrected for, using scientific knowledge of the spreading dynamics and the measurement process. We conclude that such corrections should generally be part of epidemic monitoring, modelling and forecasting efforts.

This article is part of the theme issue ‘Data science approaches to infectious disease surveillance’.

Keywords: epidemic modelling, complex systems, network theory, computer simulation, data science, statistics

1. Introduction

For thousands of years, the epidemic spreading of diseases has been one of the greatest threats to societies around the world. Now, with increased mobility, diseases tend to spread faster and wider due to air traffic and, in fact, often globally. The resulting pandemics can cost the lives of millions of people and disrupt social, economic, public, political and cultural life.

These challenges are increasingly countered with new measurement and vaccination technologies [1]. In particular, digital technologies have enabled data-driven and AI-based methods [24], which have become quite popular. Some organizations even envision a future with ubiquitous health measurements, using, for example, in-body sensors [5].

Data analytics is also used for non-pharmaceutical interventions to handle an epidemic, such as lockdowns, social distancing, or the use of face masks. The effects of such interventions have been studied extensively to understand the possible impact on the trajectory of epidemics, including the current COVID-19 pandemic [612].

Already now, one can say that a data-driven approach has spread widely. This does often assume that the overall picture of the situation will become more accurate with a greater amount of data. Although one may expect that enough data will reveal the truth by itself, it is known that conventional Big Data analytics may face some issues, which are summarized in §A of the electronic supplementary material. Consequently, an improper use of Big Data may result in learning biased patterns that can imply problems and amplify uncertainties. When applied to policy making regarding epidemics, it may also sometimes lead to unnecessary lockdowns or affect the efficiency of measures taken to counter a disease.

In the following, we will investigate to what extent such issues may undermine an accurate assessment of the state of epidemics, even if a large amount of data is available. This is important, as responding to measurement-based predictions rather than to current data when taking proactive measures is becoming increasingly common. For example, one response to an anticipated increase of infections may be to engage in more testing to capture the expected rise. The motivation for this is clear: if infections are underestimated, hospitals might not be able to handle the number of emergencies, while an overestimation may lead to unnecessary lockdowns with severe socio-economic consequences. In both cases, unnecessary loss of lives may occur. However, is it always better to make more tests? Not necessarily so: as we will discuss, there may be undesirable side effects.

The precondition for accurate predictions is to have reliable measurement methods judging the actual state of epidemics well. Currently used epidemic modelling methods try to describe the disease dynamics with systems of differential equations, meta-population models or individual level simulations [1318]. Such models are, for example, used to assess the influence of population density, demographic factors, mobility or social interactions on the actual disease dynamics [68,18]. Data-driven or machine learning models make fewer assumptions about the actual dynamics and are applicable to a broader range of prediction problems, but they come at the cost of less explainability. Of course, it is also possible to combine classical modelling approaches with machine learning [19].

It seems, however, that in many policy decisions today, issues related to measurement processes are not yet sufficiently considered. Corrections for false positives and false negatives, even though these are a general problem, have been proposed in relatively few publications. For example, the authors of [20] used a statistical testing model with bias and testing errors along with estimates of certain population types in a compartmental model (e.g. the number of recovered unreported cases). Similarly, the authors of [21] used a Bayesian hierarchical model to evaluate the true number of cases from non-representative testing results. In [22], the authors studied implications of having uncertainty or intrinsic testing errors (false positive and false-negative rates). In [23], the authors proposed a method to pool mortality data from European countries to estimate the true incidence rate of the COVID-19 pandemic. [24] used a semi-Bayesian probabilistic bias analysis to account for incomplete testing and imperfect diagnostic accuracy, emphasizing the great under-estimation of COVID-19 severity in the USA. All of these works use real noisy epidemic data without a known ground truth to evaluate the performance of their proposed methods.

By contrast, our contribution will highlight problems related to monitoring epidemics by measurement processes. As it turns out, it is dangerous to assume that data-driven approaches would be largely exact, or that more measurements or tests or data would automatically give a better picture. Therefore, we will demonstrate the possible pitfalls of such an approach in the following—by analysing the estimation errors resulting from measurement errors, mean value approximations, randomness and network interactions. For the sake of quality assessment, we assume an epidemic model as ground truth.

Specifically, we analyse the limits to inferring the hidden number of infectious people within the population by means of established measurement procedures. Our measurement model considers testing errors, and it is assessed for the case when the true state of epidemics is well described by a stochastic epidemiological simulation in a networked population. In this controlled setting, we can quantify a mean-field correction and determine its—often quite significant—deviation from the truth in the presence of parameter uncertainties regarding intrinsic testing errors, testing biases, and network effects. We also discuss how to correct for biases with mean-field and Bayesian approaches, where the latter may incorporate both testing-related and network-related information. In summary, we present an investigation and quantification of fundamental limitations in estimating the state of epidemics.

2. Epidemic models

Depending on the characteristics of a disease, there are different compartmental models in epidemiology that aim to reflect different aspects of disease spreading and recovery by coupled differential equations. These are mean value equations implicitly assuming that the infection process is well characterized by averages and correlations do not matter. In the following, we will present the SEIRD model used in our analysis, whereas the simpler SIR and SEIR models are presented in §B of the electronic supplementary material.

SEIRD model

The SEIRD model [25] assumes that some people recover from the disease and are immune after recovery, while some people die. Besides the conventional categories of Susceptible, Infectious and Recovered, it considers the categories of Died and Exposed (i.e. infected, but not yet infectious, typically without symptoms). Intuitively, the numbers at time t are represented by S(t), I(t), R(t), D(t) and E(t). The increase in the number of Exposed E is proportional to the number of Susceptible S and the number I of those who can infect them. The proportionality constant is b. It depends on how infectious the disease is and the average number of contacts per individual and unit time. Exposed turn to Infectious with a rate a. Infectious recover with a rate c and die with a rate d. The differential equations describing the change of their numbers in the course of time are

S˙=bSI/N, 2.1
E˙=bSI/NaE, 2.2
I˙=aEcIdI, 2.3
R˙=cI 2.4
andD˙=dI. 2.5

Moreover, we have the normalization equation

S(t)+E(t)+I(t)+R(t)+D(t)=N, 2.6

where N is the initial population size. Note that in other publications the parameters a, b, c, d are often represented by Greek letters such as α,β,γ,δ.

3. Measuring the state of an epidemic

The core question of this paper is: how well can the state of epidemics be measured using common test and monitoring methods? When discussing this, we must consider that all test methods have false positive rates and false negative rates [26], even though they may be small. This also applies to virtual measurement methods such as those based on inference from tracing data. We assume that Infectious I are being tested with probability p and people without symptoms (S, E and R) with probability q. When the testing is focused on infectious people with symptoms (I), we will have qp or even qp.

In the following, for the sake of illustration, we will focus on the SEIRD model as ground truth. Furthermore, let the false-positive rate (FPR) when healthy people are tested be α, and the false-negative rate (FNR) be β when Infectious are tested, but β when Exposed are tested. Under these circumstances, the expected number of positive tests is

Np(t)=(1β)pI(t)+(1β)qE(t)+αq[S(t)+R(t)]. 3.1

Accordingly, the assumption that the number Np(t) of positively tested is proportional to the number I(t) of Infectious holds only if β=1 and α=0 (which is unrealistic [26]), or if q=0 (i.e. if ‘healthy’ people are not being tested). Otherwise, Np(t) can be quite misleading. In fact, the term αq(S+R) may dominate the other terms, if a large number of people without symptoms are being tested (corresponding to ‘mass testing’, addressed in §E of the electronic supplementary material). In such a case, the majority of positive tests would be false positives, and the number of positive tests would increase with the overall number

NT(t)=pI(t)+q[S(t)+E(t)+R(t)] 3.2

of tests made. Testing a large number of people without symptoms might, therefore, produce pretty arbitrary and misleading numbers, if appropriate statistical corrections are not done. As a consequence, the number of positive tests is not a good basis for policy measures. In particular, if the testing rate q is increased with the anticipated number of positive tests Np(t+Δt) at some future point in time t+Δt, this may lead to a ‘lock-in effect’. Then, for a long time, one may not find a way out of the vicious cycle of more predicted cases triggering more tests, which implies more measured and predicted cases, and so on.

It is much better to compare the number Np of positive tests given by (3.1) with the overall number of tests NT, see (3.2). Accordingly, the measured proportion of Infectious people, also known as prevalence, is

s^(t)=Np(t)NT(t)=(1β)pI+(1β)qE+αq(S+R)pI+q(S+E+R) 3.3

This equation might be seen as a naive estimate of the proportion of infectious individuals. Here, we have dropped the argument (t) on the right-hand side for simplicity. Let us now compare this measurement-based estimate with the actual (true) proportion of Infectious among living people, which is given by

s(t)=I(t)N(t), 3.4

where

N(t)=ND(t)=S(t)+E(t)+I(t)+R(t) 3.5

is the number of living people in the considered population. The relative error then becomes

ϵ=|s^(t)s(t)||s(t)|=|[(1β)pI+(1β)qE+αq(S+R)][pI+q(NI)]NI1|. 3.6

If the tests measure the Exposed as if they were ‘still healthy’, it is reasonable to assume (1β)=α.1 This simplifies the previous formula to

ϵ=|s^s||s|=|[(1β)pI+αq(NI)][pI+q(NI)]NI1|=|[(1β)ps+αq(1s)]s[ps+q(1s)]1|. 3.7

Let us investigate two limiting cases. First, if we assume that only people with symptoms are tested, then q=0 and

ϵ=|s^s||s|=|(1β)s1|. 3.8

Accordingly, the smaller the actual proportion s(t) of Infectious, the bigger the relative error will be. Second, if the number (NI)=(S+E+R) of people without symptoms typically outweighs the Infectious by far, one may assume that αq(S+E+R)=αq(NI)(1β)pI and q(NI)pI. In this case, we get

ϵ=|s^s||s||αs1|. 3.9

In both cases, we can see that the relative error increases with the inverse of the proportion s of Infectious. Accordingly, the relative error ϵ increases with the number of healthy people. This relative error might be huge. In fact, it is particularly big in cases of ‘mild’ epidemics characterized by s(t)0. Again, given the finite value of the false positive rate α, mass testing of people who feel healthy is not really advised, unless appropriate statistical corrections are done. It might lead to a large overestimation of the actual disease rate. That is, the state of the epidemic may be wrongly assessed.

However, there is a correction formula for the effect of false positives and negatives, and for biased samples with pq. Similar to the correction for testing errors and biases introduced in [20], to estimate the fraction of unhealthy people correctly, we need to find a function s^c that transforms the estimate s^ into s. Considering that the estimate s^ of s is given by (3.3) and that we have s=I/N with N=(S+E+I+R), under the previously made assumption (1β)=α we find

s^(s)=(1β)ps+αq(1s)(pq)s+q. 3.10

From this, we can derive the corrected value s^c of s^ via the inverse function of s^(s)

s^c(s^):=s(s^)=(s^α)q(1β)pαq(pq)s^. 3.11

This formula should only be used in the range 0<s^α<1αβ. For non-biased samples (p=q), the formula simplifies to

s^c(s^)=s^α1αβ. 3.12

As figure 1 shows, this correction can be very effective in fitting the true value I/N, if the parameters α, β, p and q are well known. However, we would like to point out that the above analysis and correction formula are based on a mean-value approximation. Let us, therefore, investigate the likely role of stochastic measurement errors. A common approach for this is the use of a binomial distribution. It describes the probability to have np positive tests among NT tests, if the ‘success rate’ (of having a positive test result) is s^. Accordingly, the binomial distribution,

P(np|NT)=(NTnp)s^np(1s^)NTnp, 3.13

where NT=pI+q(S+E+R) represents again the total number of tests made. For the binomial distribution, the expected number of confirmed cases is

np=NTs^=Np, 3.14

and the variance is

Var(np)=NTs^(1s^). 3.15

As a consequence, the relative standard deviation is

Var(np)np=NTs^(1s^)NTs^=1NT1s^s^. 3.16

Here, more tests are better, as expected. While the relative standard deviation may be significant if NT is small (at the beginning of an epidemics) or if s^0, for a sufficiently large number of tests, the relative variance will be arguably small, likely around 1%. Therefore, the above mean value equations and correction formula (3.11) appear to be applicable in many cases, when the uncertainty in the measurement parameters α, β, p and q is low. In reality, however, the parameter values showing up in formula (3.11) are estimates α^, β^, p^ and q^, which may deviate from the true parameter values α, β, p and q. Inserting these in formula (3.11) instead, one may find considerable under- or over-estimations of the true proportion of Infectious in a population (see figure 2 for the case p>q and figure 1 of §E in the electronic supplementary material for the ‘opposite’ case p<q, which corresponds to ‘mass testing’).

Figure 1.

Figure 1.

Estimation of the true proportion I/N of Infectious (green solid line) based on the fraction (3.3) of positive tests (blue dotted line) and the correction formula (3.11) (red dashed-dotted line). The corrected estimate tends to be very close to the true value when the parameters α, β, p and q of the test method are known exactly. The monitoring parameters are: α=0.05, β=0.15, p=0.8, q=0.06. The curves displayed here are for SEIRD dynamics with N=100000, a=1/5, b=3/14, c=1/28, d=1/28. (Online version in colour.)

Figure 2.

Figure 2.

Estimation of the true proportion I/N of Infectious (green solid line) based on the fraction (3.3) of positive tests (blue dotted line) and the correction formula (3.11) (other lines according to the colour scales on the right). In these figures, we assume that only three of the four parameters α=0.05, β=0.15, p=0.8 and q=0.06 are known exactly, while the estimate of the fourth one (represented with a hat on top) is uncertain and, therefore, varied. Specifically, in (3.11) we have replaced α by α^ in the top left figure, β by β^ in the top right figure, p by p^ in the bottom left figure and q by q^ in the bottom right figure. All figures show results for a SEIRD dynamics with N=100000, a=1/5, b=3/14, c=1/28, d=1/28. (Online version in colour.)

A further complication occurs, if the assumption (1β)=α does not hold (e.g. when the testing procedure recognizes Exposed as ‘already infectious’ people, even though they do not have symptoms yet). Then it is generally not possible to derive a correction formula such as (3.11). To derive such a formula, one would need to separately measure Infectious and Exposed, requiring two different, specific tests, but this is often not realistic to assume. Therefore, one would have to use additional assumptions or Bayesian inference.

4. Bayesian inference

Up to this point, we have assumed that the testing rates p and q are fractions determined by the health system. However, if these monitoring parameters reflect the probabilities of infectious and healthy individuals to be tested, and to be consequently observed as infected or not, we can consider them as random variables that vary from one test to another. Similarly, α and β are not fully known, but rather depend on factors such as the delay between the time an individual was exposed to a disease and being tested. It is worthwhile to note (although we do not consider this case here), that even the number of conducted tests is not exactly known, and likely varying with time.

To address these uncertainties, the monitoring parameters α^,β^,p^,q^ can be sampled from some independent distributions: α^fα^,β^fβ^,p^fp^,q^fq^ and used in modelling a likely number of Infectious I(t), given the number of positive tests Np at t. For example, α^,β^,p^,q^ may be Beta-distributed, which was the choice we studied here.

Considering that the number of Infectious and the number of positively tested people are also random variables in reality, Bayes’ rule applies. Here, our evidence is Np and our priors relate to knowledge about P(I)—the probability of the likely number of Infectious individuals. Bayes’ rule determines the probability that the hypothesis is true (here: the number of Infectious I is m), given a certain observation (here: the measured number n of positive tests)

P(I=m|Np=n)=P(Np=n|I=m)P(I=m)lP(Np=n|I=l)P(I=l). 4.1

Here, the left-hand side of the equation is the ‘posterior’ probability that we aim to infer. Assuming that the binomial distribution in (3.13) for Np is well approximated by a normal distribution, that s^i=s^(p^i,q^i,α^i,β^i) for each sample i from the prior distributions fα^,fβ^,fp^,fq^, and that all values of I are equally likely a priori, we can write the posterior probability distribution as

P(I=m|Np=n)1Mi=1M(12πNTs^i(1s^i)exp((nNTs^i)22NTs^i(1s^i))). 4.2

In practice, this formulation allows us to employ Monte Carlo sampling to infer the desired probability distribution efficiently. Overall, we consider the Bayesian posterior estimate to be

I^=P(I=m|Np=n)=m=0NmP(I=m|Np=n), 4.3

where . denotes the expectation over the values of P(I=m|Np=n). The full derivation of (4.3) is provided in §C of the electronic supplementary material.

Last but not least, network effects may also produce significant deviations from the mean-value approximations above (which implicitly assumed homogeneous mixing and a low variance of measurement errors). On the other hand, good knowledge of a contact network may be of use when inferring I^: sparse networks slow the disease spread, dense or small-world networks may speed it up. To assess the potential benefits of network priors as well as the extent to which the network can introduce a new source of error, in the following section, we will use a more sophisticated simulation approach for the SEIR/SEIRD model that considers stochastic as well as network effects.

To include network priors in the Bayesian framework, we will consider samples of epidemic trajectories from NG randomly generated networks with given network characteristics. For each of the networks, we simulate Ne epidemic trajectories, each with a randomly chosen initial infectious node, giving a prior distribution for P(I=m). The prior distribution P(It=m) is then estimated as the probability density function of trajectories in the time window [tδ,t]. We can incorporate this network prior in (4.3) such that P(I=m|Np=n)P(I=m|Np=n)P(I=m) for each sample i.

5. Stochastic simulation of epidemic spreading in social networks

In order to study stochastic SEIRD dynamics, we will consider epidemic spreading on a contact network G=(V,E), defined by a set of nodes V (corresponding to N individuals) and a set of edges E (representing their contacts, M in total). Here, we are interested in the SEIRD process, determined by the parameters (a,b,c,d), where a Susceptible individual becomes Exposed with rate b by having a contact with an Infectious neighbour, Exposed transition from the Exposed state to Infectious state with rate a, and Infectious Recovered or Died with an overall rate c+d. These processes have exponential inter-event time distributions ψ(τ)=bebτ (spreading) and ϕ(τ)=(c+d)e(c+d)τ (recovery or death), ξ(τ)=aeaτ (transition from Exposed to Infectious).

To simulate a set of Ne possible stochastic SEIRD epidemic spreading outcomes on a given contact network G, we use the SP-KMC shortest-path Kinetic Monte Carlo method described in [27,28] for the SIR model, and expand it to the SEIRD model. For both SIR and SEIRD models, the SP-KMC method works in two steps: (i) creating weighted networks {GNeSEIRD} and (ii) extracting a set of k possible dynamic realizations from {GNeSEIRD}, where the superscript indicates the type of epidemic model considered.

We then proceed with the first step by building a time-respecting weighted network instance GNeSEIRD created by taking the input network G and assigning weights to the edges of the network instance such that

ρij={ln(x)/bln(z)/aif ln(x)/bln(y)/(c+d)if ln(x)/b>ln(y)/(c+d) 5.1

Here, x,y,z are uniform random numbers from the domain [0,1]. However, the SP-KMC method can also work with non-exponential inter-event distributions [27,28]. Each weighted network GNeSEIRD is linked to a possible outcome of the epidemic spreading process, starting from a randomly selected ‘source’ node. Therefore, to extract an epidemic trajectory represented by GNeSEIRD, one needs to find the shortest paths from the source node to other nodes in this network [27,28]. We define the distance as the shortest path on the weighted network

dGNeSEIRD(vi,vj)=minχij(k,l)χijρkl, 5.2

where χij is the set of all possible paths from node vi to node vj on the network GNeSEIRD, and ρkl denotes the weights defined in (5.1). Since the length of the shortest path from the source to some other node represents the first infection time in epidemic realization, the temporal evolution of the number of infectious is extracted from the statistics of the shortest paths in GNeSEIRD (see §D of the electronic supplementary material for more details).

(a) . Network models

We studied random networks generated using two network models: (i) the Barabási–Albert model and (ii) the Erdös–Rényi model. The Erdös–Rényi (ER) network model [29] is probably the simplest random network model. In this model, one assumes no prior knowledge about the network, nodes and edges. To build a representative network of this model, one follows a simple rule: consider each pair of nodes, and with some probability π build an edge between them: each pair of nodes is connected with a probability π to give a simple graph. ER networks are ‘structure-less’, being rather dissimilar to most real-world networks [30]. Given π and the number of nodes, N, it is useful to characterize this network by the average degree of nodes, k, equal to the average number of links adjacent to a node. For a large ER network, k=Nπ and node degrees are Poisson-distributed.

Another network type we consider is called ‘scale-free’; here, the degree distribution follows a power law [31]. The term ‘scale-free’ means that there is no typical number of edges. As a result, we have many nodes with very small degree (close to one), and very few ‘winner’—nodes with very large degrees (we say that the degree heterogeneity is large) [30].

The most famous mechanism to generate scale-free networks is preferential attachment. In preferential attachment, the more connected a node is, the quicker it acquires more new connections. The Barabási–Albert (BA) [32] is one model that generates random scale-free networks using a preferential attachment mechanism. A BA network is generated iteratively, beginning with an initial connected network of μ nodes. Each new node is connected to μ existing nodes with a probability π that is proportional to the degrees of existing nodes, and this probability is defined as πkγ+c. Here, c can be interpreted as a number of edges out of μ that are attached via random attachment at each time step. If γ=1, the resulting network is scale-free and P(k)k3 [33].

(b) . Results with mean-field correction

Figure 3 shows on the left that the mean-field correction (3.11) works well if all measurement parameters α, β, p and q are known. However, for somewhat incorrect parameter estimates, we observe considerable deviations from the mean-field correction. This establishes the need for Bayesian inference.

Figure 3.

Figure 3.

Ensemble of stochastic trajectories (SP-KMC sampling) on an ensemble of 103 Barabási-Albert networks with N=100000, μ=3 and γ=1. (a) Accurate epidemic reconstruction assuming that the parameters α=0.05, β=0.15, p=0.8, q=0.06 of the test method are exactly known. (b) Inaccurate reconstruction for somewhat incorrect estimates p^=0.75 and q^=0.1, while the other parameters are assumed to be the same. The assumed parameters of the SEIRD dynamics are: a=1/5, b=3/14, c=1/28, d=1/28. Q0.01, Q0.25, Q0.75 and Q0.99 represent 1, 25, 75 and 99 per cent quantiles. The median values and error quantile bands are based on 103 simulations. (Online version in colour.)

(c) . Results for Bayes correction with measurement priors

Figure 4 shows that, using the Bayesian inference (4.3), one can improve the mean-field correction (3.11), when monitoring parameters are not fully known. Here, we consider a case when α,β,p,q are all assumed to be unknowns, sampled from independent Beta distributions, see caption of figure 4. Furthermore, in this case, we assume that the Beta distributions have correct mean values, i.e. α=α^ and similarly for β,p and q. Despite some minor over-estimation, the Bayesian correction can work well in the case when the priors are appropriate.

Figure 4.

Figure 4.

I^(t) computed from (4.3), assuming uncertainty in monitoring parameters: α^Beta(10,90), β^Beta(5,95), p^Beta(4,16), q^Beta(5,95). We used the distributions whose mean is ‘guessed’ correctly, i.e.α^=α. The true number I(t) of Infectious was obtained by simulating SEIRD dynamics with parameters a=1/5, b=3/14, c=1/28, d=1/28 on an ER network with N=105 and average degree 1.5. (Online version in colour.)

(d) . Results for Bayes correction with network priors

In figure 5, we show that using stochastic trajectories generated with networks from a correct ensemble of networks, one can improve the estimation of I^=s^N. However, when using a prior with wrong assumptions about the degree distribution of contacts, we see no improvement in the posterior estimate of the ground truth, compared to the uninformative prior. Overall, we find that both the degree distribution and the network density have effects on the quality of the Bayesian inference. Therefore, we would like to emphasize the importance of choosing correct priors for a reliable estimation of the ground truth.

Figure 5.

Figure 5.

Posterior distribution P(I|np) when a network topology prior is used (in red) and when it is not (in blue). Monitoring parameters were also considered as priors, with p^Beta(4,16), q^Beta(5,95), α^Beta(5,95), β^Beta(10,90). For the ground truth parameters, we selected α=0.05, β=0.1, p=0.2, q=0.05. The ground truth number of Infectious I=43448 was obtained from an epidemic trajectory generated on a network sampled from an ensemble of Barabási–Albert (BA) networks with μ=3, γ=1, giving a scale-free network with an average degree k=3. In scenario ‘BA prior with correct μ=3, γ=1’, we used a prior which assumes the correct network ensemble—Barabási-Albert networks with N=100000, μ=3 and γ=1. In ‘BA prior with μ=6, γ=1’, we used a network ensemble prior of BA networks with μ=6 and γ=1, which over-estimates the average node degree k. In ‘BA prior with μ=3, γ=2’, we used a prior for BA networks with μ=3 and γ=2, and in ‘ER prior with π=3×105’ we used a prior for Erdös–Rényi networks with π=3×105, giving networks with k=3, but the degree distribution P(k) is Poisson rather than a power law. In all figures, the green line shows the ‘ground truth’, against which the biased testing is performed. We observe that knowledge about the degree distribution of contacts helps to estimate the true ground truth. To estimate network priors within t[60,65], we have used 1000 networks from the ensemble. In all cases, M=1000 Monte Carlo samples were used to estimate (4.3). (Online version in colour.)

For the example shown in figure 5, we considered networks with N=100000 nodes. The ground truth networks were generated using a BA model with μ=3, γ=1, yielding scale-free networks. To study the implications of using somewhat incorrect network priors, we used networks generated with either a BA model and wrong average degree (here, we used μ=6) and wrong γ (here, we used γ=2). We also considered the case when the network prior is assumed to be an ER model generated network with π=3×105, yielding an average degree of 3 in the network.

6. Summary, conclusion, discussion and outlook

In this paper, we addressed some challenges regarding epidemic monitoring and modelling, particularly regarding (i) the modelling itself and (ii) the applicability of models to the real world. Regarding the first point, it is important to choose the right epidemiological model and to consider limitations of the mean value-based correction approach. Stochastic and network effects—in combination with the underlying nonlinear spreading dynamics—are relevant as well. Regarding the second point, one needs to consider errors of tests and estimation procedures underlying epidemic monitoring. This concerns false-positive and false-negative rates as well as biases in the measurement sample (pq). Our goal was to analyse a good approximation of reality (where we have represented the latter by an assumed ground truth based on SEIRD dynamics). Overall, we showed that data-driven approaches relying only on measured numbers of Infectious can be misleading, if reliable measurement corrections are not available or not applied.

Note that the presented Bayesian inference also makes hidden modelling assumptions such as (i) testing errors are distributed binomially or normally, (ii) parameters are time-independent, (iii) all relevant parameters are being considered in the Bayesian analysis. Under these assumptions, we have found several limitations of different statistical methods for inferring the hidden state of epidemics.

Monitoring real-world epidemics is even more difficult, because one does not know the ground truth, and the goal of all types of modelling is to infer it. Although a simple mean-field correction can work pretty well if the measurement parameters are known (figure 1), substantial deviations may occur when there is some uncertainty about them (figure 2), which is usually the case. Then, one might resort to a Bayesian correction for improvements [2024], but this will not completely remove all issues of measurement errors and uncertainties either.

For all the above reasons, forecasting the epidemic dynamics is a problem that has fundamental limitations even in the age of Big Data and Artificial Intelligence. Given the intrinsic limitations of tests, these problems are expected to stay around. Measurement problems using dynamical models [14,34] may be further amplified if the test method or the social behaviour of people are dynamically changed in response to the epidemic dynamics. Hence, data protection regulations are not the main problem for monitoring, forecasting and controlling epidemics. The limitations of a data-driven approach are a lot more fundamental in nature.

Acknowledgements

We thank Lucas Böttcher for inspiring discussions and scientific exchange in the initial phase of the project, which related to models of statistical measurement as presented in [20] and to empirical forecasting methods.

Footnotes

1

Because Exposed tested positive would then be considered ‘false positives’—they would be valued ‘true positive’, once they have transitioned to the state of ‘Infectious’.

Data accessibility

This paper is not presenting any empirical data. Additional information is available in the electronic supplementary material.

Authors' contributions

All authors developed the theoretical concepts and simulation scenarios together. D.H. initiated the study, summarized Big Data issues and worked on integrating a measurement model into models of epidemic spreading. N.A.F. and V.V. introduced the probabilistic inference parts, created the computer code used, and performed the simulations. All authors wrote, read and approved the final manuscript.

Competing interests

We declare we have no competing interests.

Funding

This work is supported by the European Union - Horizon 2020 Program under the scheme ‘INFRAIA-01-2018-2019 - Integrating Activities for Advanced Communities’, Grant Agreement no. 871042, ‘SoBigData++: European Integrated Infrastructure for Social Mining and Big Data Analytics’ (http://www.sobigdata.eu).

References

  • 1.van Riel D, de Wit E. 2020. Next-generation vaccine platforms for COVID-19. Nat. Mater. 19, 810-812. ( 10.1038/s41563-020-0746-0) [DOI] [PubMed] [Google Scholar]
  • 2.Arshadi AK et al. 2020. Artificial intelligence for COVID-19 drug discovery and vaccine development. Front. Artif. Intell. 3, 65. ( 10.3389/frai.2020.00065) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3.Vaishya R, Javaid M, Khan IH, Haleem A. 2020. Artificial intelligence (AI) applications for COVID-19 pandemic. Diabetes Metab. Syndr.: Clin. Res. Rev. 14, 337-339. ( 10.1016/j.dsx.2020.04.012) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4.Bharadwaj KK et al. 2021. Computational intelligence in vaccine design against COVID-19. In Computational intelligence methods in COVID-19: surveillance, prevention, prediction and diagnosis, pp. 311–329. Berlin, Germany: Springer. ( 10.1007/978-981-15-8534-0_16) [DOI]
  • 5.Eltorai AE, Fox H, McGurrin E, Guang S. 2016. Microchips in medicine: current and future applications. BioMed Res. Int. 2016, 1743472. ( 10.1155/2016/1743472) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6.Moore S, Hill EM, Tildesley MJ, Dyson L, Keeling MJ. 2021. Vaccination and non-pharmaceutical interventions for COVID-19: a mathematical modelling study. Lancet Infect. Dis. 21, 793-802. ( 10.1016/S1473-3099(21)00143-2) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7.Gozzi N, Bajardi P, Perra N. The importance of non-pharmaceutical interventions during the COVID-19 vaccine rollout. PLoS Comput. Biol. 17, e1009346. ( 10.1371/journal.pcbi.1009346) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.Yang W, Shaff J, Shaman J. 2021. Effectiveness of non-pharmaceutical interventions to contain COVID-19: a case study of the 2020 spring pandemic wave in New York City. J. R. Soc. Interface 18, 20200822. ( 10.1098/rsif.2020.0822) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9.Zheng Z, Xie Z, Qin Y, Wang K, Yu Y, Fu P. 2021. Exploring the influence of human mobility factors and spread prediction on early COVID-19 in the USA. BMC Public Health 21, 1-13. ( 10.1186/s12889-021-10682-3) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10.Hu S, Xiong C, Yang M, Younes H, Luo W, Zhang L. 2021. A big-data driven approach to analyzing and modeling human mobility trend under non-pharmaceutical interventions during COVID-19 pandemic. Transp. Res. Part C: Emerg. Technol. 124, 102955. ( 10.1016/j.trc.2020.102955) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11.Zhou Y, Xu R, Hu D, Yue Y, Li Q, Xiz J. 2020. Effects of human mobility restrictions on the spread of COVID-19 in Shenzhen, China: a modelling study using mobile phone data. Lancet Digit. Health 2, e417-e424. ( 10.1016/S2589-7500(20)30165-5) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12.Liang J, Yuan H-Y, Wu L, Pfeiffer DU. 2021. Estimating effects of intervention measures on COVID-19 outbreak in Wuhan taking account of improving diagnostic capabilities using a modelling approach. BMC Infect. Dis. 21, 1-10. ( 10.1186/s12879-021-06115-6) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.Vespignani A et al. 2020. Modelling COVID-19. Nat. Rev. Phys. 2, 279-281. ( 10.1038/s42254-020-0178-4) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14.Engbert R, Rabe MM, Kliegl R, Reich S. 2021. Sequential data assimilation of the stochastic SEIR epidemic model for regional COVID-19 dynamics. Bull. Math. Biol. 83, 1-16. ( 10.1007/s11538-020-00834-8) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15.Maier BF, Brockmann D. 2020. Effective containment explains subexponential growth in recent confirmed COVID-19 cases in China. Science 368, 742-746. ( 10.1126/science.abb4557) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16.Jia JS, Lu X, Yuan Y, Xu G, Jia J, Christakis NA. 2020. Population flow drives spatio-temporal distribution of COVID-19 in China. Nature 582, 389-394. ( 10.1038/s41586-020-2284-y) [DOI] [PubMed] [Google Scholar]
  • 17.Siegenfeld AF, Taleb NN, Bar-Yam Y. 2020. Opinion: what models can and cannot tell us about COVID-19. Proc. Natl Acad. Sci. USA 117, 16 092-16 095. ( 10.1073/pnas.2011542117) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18.Van den Broeck W, Gioannini C, Gonçalves B, Quaggiotto M, Colizza V, Vespignani A. 2011. The GLEaMviz computational tool, a publicly available software to explore realistic epidemic spreading scenarios at the global scale. BMC Infect. Dis. 11, 1-14. ( 10.1186/1471-2334-11-37) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19.Wu D, Gao L, Xiong X, Chinazzi M, Vespignani A, Ma Y-A, Yu R. 2021. DeepGLEAM: a hybrid mechanistic and deep learning model for COVID-19 forecasting. Preprint (https://arxiv.org/abs/2102.06684).
  • 20.Böttcher L, D’Orsogna M, Chou T. 2021. Using excess deaths and testing statistics to improve estimates of COVID-19 mortalities. Eur. J. Epidemiol 36, 545-558. ( 10.1007/s10654-021-00748-2) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21.Campbell H, de Valpine P, Maxwell L, de Jong VMT, Debray T, Jänisch T, Gustafson P, 2020. Bayesian adjustment for preferential testing in estimating the COVID-19 infection fatality rate: Theory and methods. Preprint (https://arxiv.org/abs/2005.08459).
  • 22.Bentley P. 2021. Error rates in SARS-CoV-2 testing examined with Bayes’ theorem. Heliyon 7, e06905. ( 10.1016/j.heliyon.2021.e06905) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 23.Català M et al. 2021. Robust estimation of diagnostic rate and real incidence of COVID-19 for European policymakers. PLoS ONE 16, e0243701. ( 10.1371/journal.pone.0243701) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 24.Wu SL et al. 2020. Substantial underestimation of SARS-CoV-2 infection in the United States. Nat. Commun. 11, 1-10. ( 10.1038/s41467-020-18272-4) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25.Hethcote HW. 2000. The mathematics of infectious diseases. SIAM Review 42, 599-653. ( 10.1137/S0036144500371907) [DOI] [Google Scholar]
  • 26.Banerjee A, Chitnis U, Jadhav S. et al. 2009. Hypothesis testing, type I and type II errors. Ind. Psychiatry J. 18, 127. ( 10.4103/0972-6748.62274) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27.Böttcher L, Antulov-Fantulin N. 2020. Unifying continuous, discrete, and hybrid susceptible-infected-recovered processes on networks. Phys. Rev. Res. 2, 033121. ( 10.1103/PhysRevResearch.2.033121) [DOI] [Google Scholar]
  • 28.Tolić D, Kleineberg K-K, Antulov-Fantulin N. 2018. Simulating SIR processes on networks using weighted shortest paths. Sci. Rep. 8, 1-10. ( 10.1038/s41598-018-24648-w) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 29.Erdös P, Rényi A. 1960. On the evolution of random graphs. Publ. Math. Inst. Hung. Acad. Sci. 5, 17-60. [Google Scholar]
  • 30.Vasiliauskaite V, Rosas FE. 2020. Understanding complexity via network theory: a gentle introduction. Preprint (https://arxiv.org/abs/2004.14845).
  • 31.Broido AD, Clauset A. 2019. Scale-free networks are rare. Nat. Commun. 10, 1-10. ( 10.1038/s41467-019-08746-5) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 32.Barabási A-L, Albert R. 1999. Emergence of scaling in random networks. Science 286, 509-512. ( 10.1126/science.286.5439.509) [DOI] [PubMed] [Google Scholar]
  • 33.Dorogovtsev SN, Mendes JF. 2002. Evolution of networks. Adv. Phys. 51, 1079-1187. ( 10.1080/00018730110112519) [DOI] [Google Scholar]
  • 34.Yang W, Karspeck A, Shaman J. 2014. Comparison of filtering methods for the modeling and retrospective forecasting of influenza epidemics. PLoS Comput. Biol. 10, e1003583. ( 10.1371/journal.pcbi.1003583) [DOI] [PMC free article] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Data Availability Statement

This paper is not presenting any empirical data. Additional information is available in the electronic supplementary material.


Articles from Philosophical transactions. Series A, Mathematical, physical, and engineering sciences are provided here courtesy of The Royal Society

RESOURCES