Abstract
This paper studies the SEIRD epidemic model for COVID-19. First, I show that the model is poorly identified from the observed number of deaths and confirmed cases. There are many sets of parameters that are observationally equivalent in the short run but lead to markedly different long run forecasts. Second, I show that the basic reproduction number can be identified from the data, conditional on epidemiologic parameters, and propose several nonlinear SUR approaches to estimate . I examine the performance of these methods using Monte Carlo studies and demonstrate that they yield fairly accurate estimates of . Next, I apply these methods to estimate for the US, California, and Japan, and document heterogeneity in the value of across regions. My estimation approach accounts for possible underreporting of the number of cases. I demonstrate that if one fails to take underreporting into account and estimates from the reported cases data, the resulting estimate of may be biased downward and the resulting forecasts may exaggerate the long run number of deaths. Finally, I discuss how auxiliary information from random tests can be used to calibrate the initial parameters of the model and narrow down the range of possible forecasts of the future number of deaths.
Keywords: Parameter identification, COVID-19, SEIR model, Seemingly unrelated equations
1. Introduction
The SIR (Susceptible, Infectious, Recovered) model and its variations are widely used in epidemiology to model the spread of epidemics. Since the outbreak of COVID-19, it has seen increased popularity among economists who are trying to assess the economic consequences of the coronavirus and various mitigation policies, such as Acemoglu et al. (2020), Atkeson, 2020b, Atkeson, 2020c, Avery et al. (2020), Berger et al. (2020), Eichenbaum et al. (2020), Ellison (2020), Fernandez-Villaverde and Jones (2020), Piguillem and Shi (2020), Toda (2020), and others. In this paper, I study identification and estimation of the modification of the SIR model called SEIRD (Susceptible, Exposed, Infectious, Recovered, and Dead) and present several findings.
First, I show that the SEIRD model has too many degrees of freedom and is poorly identified from the short run data on the number of deaths and confirmed cases. Conditional on the values of epidemiologic parameters, i.e. parameters that reflect the clinical progression of the disease, the only model parameter that is identified is the basic reproduction number . While governs the speed of spread of the virus, the key driver of the long run number of deaths in the model is the infection fatality rate (IFR), which is not identified separately from initial values. As a result, models that are observationally equivalent in the short run can produce markedly different long run forecasts of the number of deaths.
Second, I propose several nonlinear seemingly unrelated regressions (SUR) approaches to estimate based on the deaths and confirmed cases data. The approaches I consider differ in whether they use cumulative or daily data and how they introduce errors in the model. I study the performance of different approaches in simulations and find that the methods based on cumulative data typically outperform those based on daily data in terms of the mean squared error (MSE) of the estimate of . While there is no clear ranking of the approaches based on the data in levels or logarithms, the former do not involve trimming and thus may be more convenient in practice.
Next, I estimate the basic reproduction number for the US, California, and Japan for different values of epidemiologic parameters. I show that there is substantial heterogeneity in the values of : for the same values of epidemiologic parameters, the estimates of for the US and California are about 2–4 times higher than for Japan. Moreover, the estimates of are highly sensitive to the values of epidemiologic parameters. There is no agreement in the medical literature on the length of the incubation and infectious period for COVID-19, and different values of these parameters result in the estimates of for the US that range from under 5 to around 17. Despite these large differences in the estimates of , the resulting models lead to virtually identical fit of the observed data. These findings highlight that there is no single value of that is consistent with the data, at least in the short run. The appropriate value of depends both on the region and on the model.
My model and estimation strategy take into account possible underreporting of the number of COVID-19 cases. Even though the fraction of all cases that is reported is not identified, I show that it is important to allow it to differ from one. I demonstrate that if one does not take underreporting into account and estimates from the confirmed cases data, assuming that all cases are reported, the estimate of may be biased downward and the long run number of deaths may be overestimated.
Finally, I use the example of Iceland to show how auxiliary data can be used to narrow down the range of possible forecasts of the long run number of deaths from the epidemic. I use the results of presumably random testing conducted in Iceland to calibrate the initial conditions of the model and show that doing so results in a more than 4-fold reduction in the range of possible forecasts. This finding highlights the importance of random testing. Once more countries conduct tests of random samples of population for having COVID-19 as well as for having antibodies to it, it may become possible to calibrate the initial values better and obtain more precise forecasts about the future.
The remainder of the paper is organized as follows. Section 2 presents the SEIRD model. Section 3 describes the data I use. Section 4 discusses identification of the model. Section 5 outlines the estimation procedure. Section 6 contains Monte Carlo evidence. Section 7 presents the empirical results. Section 8 concludes. Appendix A presents additional results.
2. Model
In this paper I study a version of the SEIR model that includes dead among its compartments. Similar models have been used in epidemiology by Chowell et al. (2007), Lin et al. (2020), Wang et al. (2020), and others. More advanced versions of the model with more compartments are considered in Chowell et al., 2003, Chowell et al., 2006. I consider a model with five groups of people: susceptible (S), exposed (E), infectious (I), recovered (R), and dead (D). Susceptible are those who have not gotten the virus yet and can become infected. Exposed are those who have gotten the virus but cannot transmit it to others yet. This corresponds to the so called incubation period. Infectious are those who have the virus and are contagious. Recovered are those who were sick in the past but have recovered from the virus. Dead are those who have died because of the virus.
Including the exposed compartment in the model is important because, according to the CDC, COVID-19 involves an incubation period of up to 14 days.1 As a result, the SEIRD model should reflect the progression of the epidemic more accurately than a simpler SIRD model that does not include an incubation period.
The number of people in different groups evolves over time as follows:
(2.1) |
(2.2) |
(2.3) |
(2.4) |
(2.5) |
(2.6) |
is the population size of a given country or region. I assume that it is fixed and does not vary over time. I could model the dynamics of the population size to account for the fact that some people die from the disease, but then I would also need to model births and deaths due to other causes. In order to avoid these complications, I simply fix , as is commonly done in the literature. is the cumulative number of cases confirmed. It does not affect the model dynamics but is used to match the model to the confirmed cases data. In my main analysis, I assume that it is the people who are infectious, rather than exposed, who are tested for the virus. In my robustness checks, I replace in Eq. (2.6) with and find that the results remain virtually unchanged.
The evolution of the SEIRD model depends on several parameters. I will refer to and as epidemiologic parameters. The parameter reflects the estimated duration of illness. Its estimates in the literature vary from (e.g. Wang et al. (2020)) to (e.g. Lin et al. (2020)). The parameter reflects the estimated incubation period of the disease. Its estimates in the literature vary from (e.g. Wang et al. (2020), Lauer et al. (2020)) to (Lin et al. (2020)).
The parameter reflects the rate at which infectious people interact with others. It is often written as , where , called the basic reproduction number, measures the transmission of the disease with no mitigation efforts. Liu et al. (2020) review the literature on the estimation of for COVID-19 and conclude that the average and median estimates in the literature are around 3. However, Sanche et al. (2020) estimate that in China was equal to 5.7, much higher than found in the previous literature.
The parameter is the infection fatality rate (IFR). As discussed in Korolev (2020), the IFR has serious limitations and heavily depends on the composition of people who get sick. The IFR may also not be constant over time and can substantially increase if the health care system becomes overwhelmed. However, for simplicity, I assume that is fixed and try to estimate it. Finally, is the proportion of all COVID-19 cases that is reported. It is also estimated.
The initial conditions for the number of recovered and dead are and . Because the evolution of the model does not depend on initial number of confirmed cases , its choice does not affect my identification results. For simplicity, I set . Any other fixed value could be used, or could be estimated. Next, I need to pick the initial number of infectious and exposed . I discuss their choice later in the paper. Finally, the initial number of susceptible people is .
3. Data
In my estimation, I use the deaths and confirmed cases data for COVID-19. The country level data is collected by the Center for Systems Science and Engineering at Johns Hopkins University and is available online.2 The state level data for the US is collected by the New York Times and is also available online.3 The population of different countries and regions is taken from World Population Prospects 2019 by United Nations4 and from the US Census Bureau.5
I use observations in my sample, with the first observation being January 22, 2020 (for the US and Japan) or January 25, 2020 (for California). Around that time, cases of coronavirus were widely registered outside China, e.g. in the US (January 21),6 Germany (January 27),7 and the UK (January 31).8 However, as I show below, the initial conditions and the epidemic start date are not identified separately from the IFR and the fraction of cases reported. I discuss the identification challenges in more detail below.
I limit the sample to the first 60 observations because several states in the US issued stay home orders in March, e.g. California on March 19 and New York on March 22.9 One may be worried that these measures affected the value of the basic reproduction number . By considering the first 60 observations, i.e. the data up to March 21 (for the US and Japan) or March 24 (for California), I should be able to address this concern. I consider alternative sample sizes in the robustness checks.
One may be concerned that the number of deaths because of the virus is misreported. It could be underreported because some people who die from the virus are not tested or overreported because some people who test positive for the virus actually die from other causes. If misreporting is constant over time, then the esimate of will be correct but the estimate of will be biased. If the degree of deaths misreporting varies over time, then the estimate of may be biased. Similarly, if the fraction of cases that is reported changes over time, this may result in biased estimates of .
4. Identification
In this section, I study identification of the model parameters based on the deaths and confirmed cases data. There are several earlier papers on identification of the parameters of the SIR and related models, e.g. Marinov et al. (2014), Magal and Webb (2018), and Ducrot et al. (2019), but they are not directly applicable in the current setting. In particular, they do not study whether the parameters are identified based on the short run data only. Atkeson (2020a), written concurrently and independently of this paper, attempts to answer the question similar to mine in the context of the usual SIR model. Manski and Molinari (2020) discuss identification problems in estimating the COVID-19 infection rate, but they do not consider SIR type models studied in this paper.
In the econometrics literature, identification studies whether the parameters of the model would be known if the researcher knew the population that data is drawn from (see, e.g., Lewbel (2019)). In the context of this paper, the question is somewhat different: if we observed the evolution of deaths and confirmed cases in the short run without any noise, would we then know the parameters of the model? Or, in other words, do different parameter values lead to different realizations of the observable data in the short run? Because the SEIRD model is difficult to solve in closed form, I simulate the deaths and reported cases paths from models with different parameter values and investigate whether these paths are identical instead of studying identification theoretically. I present a more rigorous treatment of identification in the simplified SIRD, rather than SEIRD, model in Supplementary Appendix A.1 and obtain similar results.
First, I assume that epidemiologic parameters and are known constants and study identification of the remaining parameters. The parameters of the model then include the basic reproduction number , the infection fatality rate , the fraction of all cases that is reported, and the initial conditions: the number of infectious people , the number of exposed people , and the time that has passed since the epidemic started. For instance, means that the epidemic just started and the initial values correspond to the first period we observe. means that the epidemic started last period and the current period corresponds to . means that the epidemic started 9 periods ago from values and the current values are . I denote the vector of parameters and study whether these parameters can be identified based on the short run (say, 60 days) data.
The upper panel of Fig. 1 plots the simulated paths of deaths and confirmed cases for three sets of parameters: , , and . The first two sets of parameters share the same start date and reproduction number, but differ in the initial values and the values of and . Essentially, the epidemic that corresponds to the second set of parameters just scales the first epidemic up by a factor of two, but cuts the fatality rate and the reported fraction of cases in half. As a result, these two epidemics are indistinguishable in the short run. In other words, we cannot tell from the short run data whether we observe a large epidemic with a low fatality rate and large number of unreported cases, or a small epidemic with a high fatality rate and small number of unreported cases.
Fig. 1.
Parameter identification. The upper panel shows the short run number of deaths and reported cases for three sets of parameters , , and , where . The middle panel shows the long run forecasts from these models. The lower panel fixes the initial conditions and shows the short run number of deaths and reported cases for , , and .
The third epidemic starts from the same values of as the first one, but nine periods ago instead of last period. At the same time it reduces the fatality rate and the observable fraction by a factor of 2.5. It produces more cases in the current period than the first epidemic, but a smaller fraction of them is reported and a smaller fraction leads to death. As a result, the third epidemic is indistinguishable from the first one.
While the three sets of parameters are indistinguishable in the short run, the middle panel of Fig. 1 shows that the resulting epidemics lead to very different long run deaths forecasts. The epidemic with the highest fatality rate will result in about twice as many deaths as any of the other two epidemics.
Next, I study identification of . The bottom panel of Fig. 1 shows that affects the curvature of the deaths and reported cases curves, while other parameters only tilt it around the origin. As a result, can be uniquely identified from the curvature of deaths and confirmed cases.
Fig. 2 parallels Fig. 1, but plots the logarithms of deaths and reported cases rather than their levels. It starts from day 30 rather than day 1, because logarithms are very sensitive to small values of different variables that are observed in the very beginning of the epidemic. The figure shows that changes in the initial conditions, the fatality rate , or the observable fraction of cases shift the lines up or down without affecting their slope, while changes in the reproduction number change the slope of the lines. Thus, can be identified from the slope of the log series, but the remaining parameters cannot be separately identified.
Fig. 2.
Parameter identification in logarithms. The upper panel shows the logarithms of the short run number of deaths and reported cases for three sets of parameters , , and , where . The lower panel fixes the initial conditions and shows the logarithms of the short run number of deaths and reported cases for , , and .
Because one cannot separately identify , , , , and , I set and . In my empirical analysis, I will generally set , unless it leads to computational issues. The model with the lowest possible value of corresponds to the highest possible value of the IFR and yields the upper bound on the long run number of deaths. I consider alternative choices of in the robustness checks.
Next, Fig. 3 explores the role of different parameters in the evolution of the model. It demonstrates that changes in the value of primarily affect the timing of the epidemic but have little effect on the total death toll. The values of and affect the number of deaths and reported cases respectively, but they have no effect on the model dynamics. Finally, the initial values and affect the timing of the model, but to a much smaller extent than the value of . Thus, if we are interested in modeling the evolution of the epidemic and its burden in terms of the number of deaths, the primary parameters of interest are and , while the remaining model parameters can be viewed as nuisance parameters.
Fig. 3.
Role of different parameters. The upper panel shows the evolution of the number of deaths. The middle panel shows the evolution of the number of reported cases. The lower panel shows the evolution of the (unobserved) number of infectious cases. Left: , , and . Right: , , and .
5. Estimation
In this section, I turn to estimation of the SEIRD model. In order to rationalize the model with the observed data, one needs to introduce errors in the model. There are several possible ways to introduce errors, and they can lead to different estimation approaches. First, one could introduce errors directly in the cumulative numbers. They would be given by
(5.1) |
(5.2) |
where . and denote the cumulative number of deaths and reported cases in the model,10 while and denote the cumulative number of deaths and reported cases observed in the data. Because the errors and enter the cumulative equations, they would likely exhibit strong autocorrelation.
One could introduce errors that are possibly independent over time by modeling the daily numbers of deaths and reported cases. Denote by the first differences in the series . Then and would correspond to the observed daily number of deaths and reported cases. One could model them as
(5.3) |
(5.4) |
where . In fact, the model in Eqs. (5.1), (5.2) can be viewed as the model in Eqs. (5.3), (5.4) if and .
Alternatively, one could assume that additive errors enter the equations for the logarithms of the daily numbers of deaths and reported cases rather than their levels:
(5.5) |
(5.6) |
where . This model implies that
(5.7) |
(5.8) |
i.e. the errors in the daily numbers are multiplicative.
Yet another estimation approach introduces errors in the logarithms of the cumulative number of deaths and reported cases:
(5.9) |
(5.10) |
with where .
There is no agreement on the choice of the model and estimation method in the literature. Chowell et al. (2007) estimate the model using the cumulative number of cases, i.e. Eq. (5.2). In turn, Toda (2020) uses the natural logarithm of the cumulative number of cases, i.e. Eq. (5.10). Finally, Fernandez-Villaverde and Jones (2020) estimate the model using the daily number of deaths, i.e. Eq. (5.3).
All these papers use data either on the number of reported cases or on the number of deaths, but not both. In contrast, I estimate the model using both series simultaneously. All four models (5.1)–(5.2), (5.3)–(5.4), (5.5)–(5.6), and (5.9)–(5.10) can be viewed as versions of the nonlinear SUR model (see, e.g., Gallant (1975) and Chapter 9 in Davidson and MacKinnon (1993)). This model has the following form:
where corresponds to the deaths equation and corresponds to the reported cases equation. The parameters are . enters both equations, only enters the deaths equation, and only the cases equation. While it is possible to estimate the model equation by equation, this would result in two different estimates of and could lead to efficiency loss. Instead, as is usually done in the SUR literature, I estimate both equations jointly. Let , , . Then the objective function is given by
where is a 2 × 2 weighing matrix. Several choices of are possible. The simplest possible choice is , a 2 × 2 identity matrix, which assigns the same weight to both equations. Another possibility is
where is the estimate of the variance of the error terms equation by equation, are the residuals based on some preliminary estimates of . These preliminary estimates could be the equation by equation estimates (with different values of for the two equations) or the estimates based on the known weighting matrix . This choice of accounts for possibly different variances of the error terms in two equations but discards possible correlation between them.
Yet another choice is given by
where , and are the same as above. This choice of the weighting matrix accounts not only for possibly different variances of the error terms, but also for the correlation between them.
Finally, if one suspects heteroskedasticity, one could use the objective function
where is an estimate of the variance of that is allowed to vary over . Because estimating may be somewhat tricky, I do not consider this class of estimators in this paper.
While inference is also beyond the scope of this paper, I note that there are several interesting directions for future research. First, one could study how to conduct inference on the model parameters, e.g. the basic reproduction number . While there are certain methods that would seem reasonable, such as the asymptotic SUR standard errors or the bootstrap (as in Chowell et al. (2007)), one would need to carefully account for partial identification of model parameters, possible heteroskedasticity and autocorrelation in the errors, and for the fact that the estimates of and may be close to the boundary of the parameter space (0 or 1). Second, one could be able to develop specification tests based on the difference between different estimates, e.g. with different weighting matrices, along the lines of Hausman (1978) and White (1981).
6. Simulations
In this section I investigate the performance of different estimation methods described above in a number of Monte Carlo studies. I consider several data generating processes (DGPs). In the first DGP, the deterministic part of the model is exactly as in Eqs. (2.1)–(2.6), while the errors are introduced as in Eqs. (5.3)–(5.4). I generate the simulated data as follows. First, I simulate the model from Eqs. (2.1)–(2.6) for the parameter values , , . Second, I construct the daily values of deaths and reported cases from the simulated model. Next, I draw the realizations of deaths and reported cases at time from the Poisson distributions with means and respectively. In other words,
where the star superscript denotes simulated values. Thus, and , so the errors in the simulated daily deaths and reported cases have mean zero by construction. Finally, I reconstruct the cumulative series and and estimate the parameters using all of the methods discussed above. In the DGP, the initial values are , . In the estimation, I try several fixed choices of as well as attempt to estimate and . When I estimate and , I restrict my attention to the approaches based on cumulative numbers, because the approaches based on daily numbers turn out to be prone to numerical issues.11
Table 1, Table 2, Table 3, Table 4 report the mean, bias, standard deviation, and MSE of the estimates of across 500 simulation draws. I consider two estimation approaches when logarithms are used: one trims all observations with or , another one uses all observations with and . I refer to the former approach as “Trim” and to the latter as “No Trim” in the tables.
Table 1.
Deaths | Cases | SUR, Efficient | SUR, Diagonal | SUR, Identity | |
---|---|---|---|---|---|
Panel A: True and , Assumed and | |||||
Levels, Cumulative | 5.774 | 5.754 | 5.757 | 5.754 | 5.754 |
Levels, Daily | 5.781 | 5.755 | 5.759 | 5.756 | 5.756 |
Logs, Cumulative, Trim | 5.771 | 5.755 | 5.753 | 5.751 | 5.761 |
Logs, Daily, Trim | 5.799 | 5.763 | 5.757 | 5.758 | 5.776 |
Logs, Cumulative, No Trim | 5.613 | 5.786 | 5.728 | 5.727 | 5.691 |
Logs, Daily, No Trim | 4.725 | 5.786 | 5.664 | 5.640 | 5.239 |
Panel B: True and , Assumed and | |||||
Levels, Cumulative | 5.773 | 5.754 | 5.757 | 5.754 | 5.754 |
Levels, Daily | 5.780 | 5.755 | 5.759 | 5.755 | 5.755 |
Logs, Cumulative, Trim | 5.770 | 5.755 | 5.753 | 5.750 | 5.761 |
Logs, Daily, Trim | 5.796 | 5.763 | 5.757 | 5.758 | 5.775 |
Logs, Cumulative, No Trim | 5.613 | 5.786 | 5.728 | 5.727 | 5.690 |
Logs, Daily, No Trim | 4.725 | 5.786 | 5.662 | 5.638 | 5.246 |
Panel C: True and , Assumed and | |||||
Levels, Cumulative | 5.788 | 5.765 | 5.769 | 5.765 | 5.765 |
Levels, Daily | 5.801 | 5.781 | 5.789 | 5.797 | 5.781 |
Logs, Cumulative, Trim | 5.782 | 5.762 | 5.761 | 5.758 | 5.769 |
Logs, Daily, Trim | 5.857 | 5.777 | 5.770 | 5.771 | 5.792 |
Logs, Cumulative, No Trim | 5.623 | 5.789 | 5.731 | 5.730 | 5.694 |
Logs, Daily, No Trim | 4.725 | 5.793 | 5.669 | 5.644 | 5.241 |
Panel D: True and , Estimated and | |||||
Levels, Cumulative | 5.774 | 5.755 | 5.763 | 5.764 | 5.769 |
Logs, Cumulative, Trim | 5.772 | 5.756 | 5.755 | 5.754 | 5.762 |
Logs, Cumulative, No Trim | 5.617 | 5.788 | 5.732 | 5.731 | 5.693 |
The table presents the mean of the estimates of across 500 simulation draws. The true value is . The estimates in the first column are based on the deaths data only. The estimates in the second column are based on the reported cases data only. The estimates in the remaining three columns combine the deaths and reported cases data and use different weighting matrices.
Table 2.
Deaths | Cases | SUR, Efficient | SUR, Diagonal | SUR, Identity | |
---|---|---|---|---|---|
Panel A: True and , Assumed and | |||||
Levels, Cumulative | 0.019 | −0.001 | 0.002 | −0.001 | −0.001 |
Levels, Daily | 0.026 | 0.000 | 0.005 | 0.001 | 0.002 |
Logs, Cumulative, Trim | 0.016 | 0.001 | −0.001 | −0.004 | 0.006 |
Logs, Daily, Trim | 0.045 | 0.009 | 0.003 | 0.003 | 0.022 |
Logs, Cumulative, No Trim | −0.141 | 0.032 | −0.026 | −0.028 | −0.064 |
Logs, Daily, No Trim | −1.030 | 0.032 | −0.091 | −0.115 | −0.515 |
Panel B: True and , Assumed and | |||||
Levels, Cumulative | 0.018 | −0.001 | 0.002 | −0.001 | −0.001 |
Levels, Daily | 0.025 | 0.000 | 0.004 | 0.000 | 0.000 |
Logs, Cumulative, Trim | 0.016 | 0.000 | −0.001 | −0.004 | 0.006 |
Logs, Daily, Trim | 0.042 | 0.009 | 0.003 | 0.003 | 0.021 |
Logs, Cumulative, No Trim | −0.142 | 0.031 | −0.026 | −0.028 | −0.064 |
Logs, Daily, No Trim | −1.030 | 0.031 | −0.092 | −0.117 | −0.508 |
Panel Dc: True and , Assumed and | |||||
Levels, Cumulative | 0.034 | 0.011 | 0.014 | 0.011 | 0.011 |
Levels, Daily | 0.046 | 0.026 | 0.034 | 0.043 | 0.026 |
Logs, Cumulative, Trim | 0.027 | 0.007 | 0.007 | 0.003 | 0.014 |
Logs, Daily, Trim | 0.102 | 0.022 | 0.016 | 0.016 | 0.037 |
Logs, Cumulative, No Trim | −0.132 | 0.035 | −0.023 | −0.025 | −0.060 |
Logs, Daily, No Trim | −1.029 | 0.038 | −0.085 | −0.110 | −0.514 |
Panel D: True and , Estimated and | |||||
Levels, Cumulative | 0.020 | 0.000 | 0.009 | 0.009 | 0.014 |
Logs, Cumulative, Trim | 0.017 | 0.001 | 0.001 | 0.000 | 0.007 |
Logs, Cumulative, No Trim | −0.137 | 0.033 | −0.023 | −0.023 | −0.062 |
The table presents the average bias of the estimates of across 500 simulation draws. The estimates in the first column are based on the deaths data only. The estimates in the second column are based on the reported cases data only. The estimates in the remaining three columns combine the deaths and reported cases data and use different weighting matrices.
Table 3.
Deaths | Cases | SUR, Efficient | SUR, Diagonal | SUR, Identity | |
---|---|---|---|---|---|
Panel A: True and , Assumed and | |||||
Levels, Cumulative | 0.317 | 0.072 | 0.076 | 0.072 | 0.071 |
Levels, Daily | 0.512 | 0.117 | 0.118 | 0.118 | 0.119 |
Logs, Cumulative, Trim | 0.424 | 0.089 | 0.092 | 0.091 | 0.219 |
Logs, Daily, Trim | 0.570 | 0.123 | 0.119 | 0.117 | 0.281 |
Logs, Cumulative, No Trim | 0.826 | 0.251 | 0.206 | 0.224 | 0.430 |
Logs, Daily, No Trim | 0.470 | 0.230 | 0.166 | 0.217 | 0.286 |
Panel B: True and , Assumed and | |||||
Levels, Cumulative | 0.316 | 0.071 | 0.075 | 0.072 | 0.071 |
Levels, Daily | 0.511 | 0.116 | 0.114 | 0.114 | 0.116 |
Logs, Cumulative, Trim | 0.423 | 0.089 | 0.092 | 0.091 | 0.218 |
Logs, Daily, Trim | 0.567 | 0.123 | 0.118 | 0.117 | 0.280 |
Logs, Cumulative, No Trim | 0.825 | 0.251 | 0.206 | 0.224 | 0.431 |
Logs, Daily, No Trim | 0.470 | 0.230 | 0.170 | 0.220 | 0.270 |
Panel C: True and , Assumed and | |||||
Levels, Cumulative | 0.328 | 0.073 | 0.078 | 0.074 | 0.073 |
Levels, Daily | 0.500 | 0.122 | 0.175 | 0.255 | 0.123 |
Logs, Cumulative, Trim | 0.435 | 0.091 | 0.093 | 0.093 | 0.223 |
Logs, Daily, Trim | 0.719 | 0.127 | 0.123 | 0.121 | 0.293 |
Logs, Cumulative, No Trim | 0.844 | 0.253 | 0.208 | 0.226 | 0.434 |
Logs, Daily, No Trim | 0.471 | 0.235 | 0.168 | 0.220 | 0.287 |
Panel D: True and , Estimated and | |||||
Levels, Cumulative | 0.318 | 0.072 | 0.083 | 0.087 | 0.089 |
Logs, Cumulative, Trim | 0.424 | 0.090 | 0.095 | 0.098 | 0.219 |
Logs, Cumulative, No Trim | 0.831 | 0.251 | 0.212 | 0.229 | 0.429 |
The table presents the standard deviation of the estimates of across 500 simulation draws. The estimates in the first column are based on the deaths data only. The estimates in the second column are based on the reported cases data only. The estimates in the remaining three columns combine the deaths and reported cases data and use different weighting matrices.
Table 4.
Deaths | Cases | SUR, Efficient | SUR, Diagonal | SUR, Identity | |
---|---|---|---|---|---|
Panel A: True and , Assumed and | |||||
Levels, Cumulative | 0.101 | 0.005 | 0.006 | 0.005 | 0.005 |
Levels, Daily | 0.263 | 0.014 | 0.014 | 0.014 | 0.014 |
Logs, Cumulative, Trim | 0.180 | 0.008 | 0.008 | 0.008 | 0.048 |
Logs, Daily, Trim | 0.327 | 0.015 | 0.014 | 0.014 | 0.079 |
Logs, Cumulative, No Trim | 0.702 | 0.064 | 0.043 | 0.051 | 0.189 |
Logs, Daily, No Trim | 1.282 | 0.054 | 0.036 | 0.060 | 0.347 |
Panel B: True and , Assumed and | |||||
Levels, Cumulative | 0.100 | 0.005 | 0.006 | 0.005 | 0.005 |
Levels, Daily | 0.262 | 0.013 | 0.013 | 0.013 | 0.013 |
Logs, Cumulative, Trim | 0.179 | 0.008 | 0.008 | 0.008 | 0.048 |
Logs, Daily, Trim | 0.323 | 0.015 | 0.014 | 0.014 | 0.079 |
Logs, Cumulative, No Trim | 0.701 | 0.064 | 0.043 | 0.051 | 0.190 |
Logs, Daily, No Trim | 1.281 | 0.054 | 0.037 | 0.062 | 0.331 |
Panel C: True and , Assumed and | |||||
Levels, Cumulative | 0.109 | 0.005 | 0.006 | 0.006 | 0.005 |
Levels, Daily | 0.253 | 0.016 | 0.032 | 0.067 | 0.016 |
Logs, Cumulative, Trim | 0.190 | 0.008 | 0.009 | 0.009 | 0.050 |
Logs, Daily, Trim | 0.528 | 0.017 | 0.015 | 0.015 | 0.087 |
Logs, Cumulative, No Trim | 0.729 | 0.065 | 0.044 | 0.052 | 0.192 |
Logs, Daily, No Trim | 1.281 | 0.057 | 0.036 | 0.060 | 0.346 |
Panel D: True and , Estimated and | |||||
Levels, Cumulative | 0.101 | 0.005 | 0.007 | 0.008 | 0.008 |
Logs, Cumulative, Trim | 0.180 | 0.008 | 0.009 | 0.010 | 0.048 |
Logs, Cumulative, No Trim | 0.709 | 0.064 | 0.045 | 0.053 | 0.188 |
The table presents the MSE of the estimates of across 500 simulation draws. The estimates in the first column are based on the deaths data only. The estimates in the second column are based on the reported cases data only. The estimates in the remaining three columns combine the deaths and reported cases data and use different weighting matrices.
As we can see, the estimation methods based on taking the logarithms without trimming have the largest bias and standard deviation, and as a result the worst MSE. Among the remaining four approaches, the ones based on cumulative numbers (or their logarithms) tend to outperform the ones based on the daily numbers (or their logarithms). Overall, it appears that estimation based on the cumulative numbers, done in levels rather than logarithms, leads to the lowest MSE of the estimates of , and the estimation results are pretty insensitive to the choice of the weighting matrix .
Using incorrect initial conditions (e.g. instead of ) has almost no effect on the quality of resulting estimates. It introduces small bias but does not affect the standard deviation of the estimates. Because in most cases the variance term dominates the squared bias term, the MSE remains almost unchanged. Moreover, in terms of the MSE, it is actually better to use an incorrect fixed initial condition rather than to estimate it from the data. Estimating and may reduce bias, but increases variance and hence leads to slightly worse MSE.
Table 5 presents the mean of the estimates of the fatality rate for different initial conditions. In line with the identification results, when increases (decreases), the estimates of decrease (increase) proportionally. For instance, the mean of the estimates of for is eight times as large as for , but is smaller by a factor of two than for .
Table 5.
SUR, Efficient | SUR, Diagonal | SUR, Identity | |
---|---|---|---|
Panel A: True and , Assumed and | |||
Levels, Cumulative | 0.0067 | 0.0068 | 0.0067 |
Levels, Daily | 0.0068 | 0.0068 | 0.0068 |
Logs, Cumulative, Trim | 0.0068 | 0.0068 | 0.0071 |
Logs, Daily, Trim | 0.0067 | 0.0067 | 0.0071 |
Logs, Cumulative, No Trim | 0.0080 | 0.0081 | 0.0098 |
Logs, Daily, No Trim | 0.0091 | 0.0097 | 0.0181 |
Panel B: True and , Assumed and | |||
Levels, Cumulative | 0.0134 | 0.0135 | 0.0135 |
Levels, Daily | 0.0135 | 0.0136 | 0.0136 |
Logs, Cumulative, Trim | 0.0136 | 0.0137 | 0.0143 |
Logs, Daily, Trim | 0.0134 | 0.0134 | 0.0143 |
Logs, Cumulative, No Trim | 0.0160 | 0.0162 | 0.0198 |
Logs, Daily, No Trim | 0.0182 | 0.0194 | 0.0352 |
Panel C: True and , Assumed and | |||
Levels, Cumulative | 0.0008 | 0.0008 | 0.0008 |
Levels, Daily | 0.0008 | 0.0008 | 0.0008 |
Logs, Cumulative, Trim | 0.0008 | 0.0008 | 0.0009 |
Logs, Daily, Trim | 0.0008 | 0.0008 | 0.0009 |
Logs, Cumulative, No Trim | 0.0010 | 0.0010 | 0.0012 |
Logs, Daily, No Trim | 0.0011 | 0.0012 | 0.0023 |
The table presents the mean of the estimates of across 500 simulation draws. The true value is . The estimates in different columns combine the deaths and reported cases data and use different weighting matrices.
Next, I study the robustness of the estimate of to a particular form of model misspecification. In the true DGP, I replace Eq. (2.6) with , so that the reported cases are based on the number of exposed rather than infectious. However, I estimate the model as if it was generated by Eqs. (2.1)–(2.6). Table 6 reports the results. As we can see, this form of misspecification has no noticeable effect on the estimates of .
Table 6.
Robustness of Estimates of .
Deaths | Cases | SUR, Efficient | SUR, Diagonal | SUR, Identity | |
---|---|---|---|---|---|
Panel A: Mean | |||||
Levels, Cumulative | 5.775 | 5.753 | 5.753 | 5.753 | 5.753 |
Levels, Daily | 5.814 | 5.756 | 5.756 | 5.757 | 5.756 |
Logs, Cumulative, Trim | 5.750 | 5.751 | 5.752 | 5.750 | 5.749 |
Logs, Daily, Trim | 5.832 | 5.755 | 5.755 | 5.755 | 5.789 |
Logs, Cumulative, No Trim | 5.528 | 5.764 | 5.739 | 5.735 | 5.642 |
Logs, Daily, No Trim | 4.590 | 5.777 | 5.720 | 5.719 | 5.225 |
Panel B: Bias | |||||
Levels, Cumulative | 0.021 | −0.002 | −0.001 | −0.001 | −0.002 |
Levels, Daily | 0.060 | 0.002 | 0.002 | 0.003 | 0.002 |
Logs, Cumulative, Trim | −0.004 | −0.003 | −0.003 | −0.005 | −0.006 |
Logs, Daily, Trim | 0.077 | 0.001 | 0.001 | 0.000 | 0.034 |
Logs, Cumulative, No Trim | −0.226 | 0.009 | −0.015 | −0.019 | −0.113 |
Logs, Daily, No Trim | −1.165 | 0.023 | −0.035 | −0.036 | −0.530 |
Panel C: Standard Deviation | |||||
Levels, Cumulative | 0.316 | 0.041 | 0.044 | 0.042 | 0.041 |
Levels, Daily | 0.508 | 0.069 | 0.070 | 0.069 | 0.070 |
Logs, Cumulative, Trim | 0.423 | 0.055 | 0.057 | 0.055 | 0.214 |
Logs, Daily, Trim | 0.567 | 0.071 | 0.072 | 0.071 | 0.282 |
Logs, Cumulative, No Trim | 0.876 | 0.153 | 0.112 | 0.144 | 0.433 |
Logs, Daily, No Trim | 0.512 | 0.165 | 0.104 | 0.154 | 0.199 |
Panel D: MSE | |||||
Levels, Cumulative | 0.100 | 0.002 | 0.002 | 0.002 | 0.002 |
Levels, Daily | 0.262 | 0.005 | 0.005 | 0.005 | 0.005 |
Logs, Cumulative, Trim | 0.179 | 0.003 | 0.003 | 0.003 | 0.046 |
Logs, Daily, Trim | 0.327 | 0.005 | 0.005 | 0.005 | 0.081 |
Logs, Cumulative, No Trim | 0.818 | 0.023 | 0.013 | 0.021 | 0.201 |
Logs, Daily, No Trim | 1.618 | 0.028 | 0.012 | 0.025 | 0.320 |
The table presents the mean, bias, standard deviation, and MSE of the estimates of across 500 simulation draws when in Eq. (2.6) is replaced with . The true value is . Initial conditions: , . The estimates in the first column are based on the deaths data only. The estimates in the second column are based on the reported cases data only. The estimates in the remaining three columns combine the deaths and reported cases data and use different weighting matrices.
Another DGP I consider is based on the model in Eqs. (2.1)–(2.6) but involves a different way of introducing errors. I focus on the model in logarithms of the daily numbers and simulate the data as in Eqs. (5.5)–(5.6), where and are both . I then compute and by taking exponents and construct the cumulative series accordingly. Unlike the previous DGPs, which produced integer values as a result, this DGP will generally produce the number of deaths and reported cases that are not integers. In addition to this DGP I also consider its modified version that rounds the number of daily deaths and reported cases to the nearest integer. The true initial values are , . In estimation, I use these fixed values as well as attempt to estimate the initial values.
Table 7, Table 8, Table 9, Table 10 report the mean, bias, standard deviation, and mean squared error of the estimates of across 500 simulation draws. The upper two panels of each table correspond to the DGP without rounding, while the bottom two panels correspond to the DGP with rounding. For the DGP without rounding, the estimation approaches based on logarithms (either cumulative or daily) without trimming dominate all other approaches in terms of the MSE, and their performance is insensitive to the choice of the weighting matrix. This is perhaps not surprising, because the model in logarithms is indeed the true model. However, for the DGP with rounding, this is no longer true: different approaches yield pretty similar values of MSE.
Table 7.
Deaths | Cases | SUR, Efficient | SUR, Diagonal | SUR, Identity | |
---|---|---|---|---|---|
Panel A: No Rounding, True and , Assumed and | |||||
Levels, Cumulative | 5.755 | 5.755 | 5.755 | 5.755 | 5.755 |
Levels, Daily | 5.762 | 5.752 | 5.757 | 5.758 | 5.753 |
Logs, Cumulative, Trim | 5.755 | 5.753 | 5.754 | 5.754 | 5.754 |
Logs, Daily, Trim | 5.756 | 5.755 | 5.756 | 5.756 | 5.755 |
Logs, Cumulative, No Trim | 5.755 | 5.755 | 5.755 | 5.755 | 5.755 |
Logs, Daily, No Trim | 5.755 | 5.755 | 5.755 | 5.755 | 5.755 |
Panel B: No Rounding, True and , Estimated and | |||||
Levels, Cumulative | 5.757 | 5.756 | 5.759 | 5.760 | 5.766 |
Logs, Cumulative, Trim | 5.756 | 5.754 | 5.757 | 5.759 | 5.755 |
Logs, Cumulative, No Trim | 5.757 | 5.759 | 5.757 | 5.757 | 5.757 |
Panel C: Rounding, True and , Assumed and | |||||
Levels, Cumulative | 5.836 | 5.760 | 5.771 | 5.776 | 5.760 |
Levels, Daily | 5.772 | 5.752 | 5.762 | 5.762 | 5.753 |
Logs, Cumulative, Trim | 5.878 | 5.761 | 5.783 | 5.789 | 5.819 |
Logs, Daily, Trim | 5.777 | 5.755 | 5.764 | 5.764 | 5.766 |
Logs, Cumulative, No Trim | 6.424 | 5.793 | 5.772 | 5.795 | 6.104 |
Logs, Daily, No Trim | 5.483 | 5.759 | 5.750 | 5.752 | 5.620 |
Panel D: Rounding, True and , Estimated and | |||||
Levels, Cumulative | 5.837 | 5.761 | 5.777 | 5.788 | 5.770 |
Logs, Cumulative, Trim | 5.879 | 5.761 | 5.785 | 5.792 | 5.820 |
Logs, Cumulative, No Trim | 6.425 | 5.793 | 5.772 | 5.795 | 6.105 |
The table presents the mean of the estimates of across 500 simulation draws for different estimation methods. The true value is . The estimates in the first column are based on the deaths data only. The estimates in the second column are based on the reported cases data only. The estimates in the remaining three columns combine the deaths and reported cases data and use different weighting matrices.
Table 8.
Deaths | Cases | SUR, Efficient | SUR, Diagonal | SUR, Identity | |
---|---|---|---|---|---|
Panel A: No Rounding, True and , Assumed and | |||||
Levels, Cumulative | 0.0007 | 0.0004 | 0.0009 | 0.0005 | 0.0002 |
Levels, Daily | 0.0078 | −0.0027 | 0.0026 | 0.0031 | −0.0013 |
Logs, Cumulative, Trim | 0.0003 | −0.0013 | −0.0007 | −0.0009 | −0.0005 |
Logs, Daily, Trim | 0.0011 | −0.0001 | 0.0012 | 0.0011 | 0.0004 |
Logs, Cumulative, No Trim | 0.0002 | 0.0003 | 0.0002 | 0.0002 | 0.0002 |
Logs, Daily, No Trim | 0.0001 | 0.0002 | 0.0002 | 0.0002 | 0.0002 |
Panel B: No Rounding, True and , Estimated and | |||||
Levels, Cumulative | 0.0020 | 0.0014 | 0.0040 | 0.0054 | 0.0110 |
Logs, Cumulative, Trim | 0.0010 | −0.0006 | 0.0025 | 0.0040 | 0.0001 |
Logs, Cumulative, No Trim | 0.0029 | 0.0047 | 0.0022 | 0.0022 | 0.0025 |
Panel C: Rounding, True and , Assumed and | |||||
Levels, Cumulative | 0.081 | 0.005 | 0.016 | 0.021 | 0.005 |
Levels, Daily | 0.017 | −0.003 | 0.007 | 0.008 | −0.001 |
Logs, Cumulative, Trim | 0.124 | 0.006 | 0.029 | 0.034 | 0.065 |
Logs, Daily, Trim | 0.023 | 0.001 | 0.010 | 0.010 | 0.012 |
Logs, Cumulative, No Trim | 0.669 | 0.038 | 0.017 | 0.040 | 0.349 |
Logs, Daily, No Trim | −0.271 | 0.005 | −0.005 | −0.002 | −0.134 |
Panel D: Rounding, True and , Estimated and | |||||
Levels, Cumulative | 0.082 | 0.006 | 0.022 | 0.033 | 0.015 |
Logs, Cumulative, Trim | 0.125 | 0.007 | 0.030 | 0.038 | 0.066 |
Logs, Cumulative, No Trim | 0.671 | 0.038 | 0.018 | 0.041 | 0.350 |
The table presents the average bias of the estimates of across 500 simulation draws for different estimation methods. The estimates in the first column are based on the deaths data only. The estimates in the second column are based on the reported cases data only. The estimates in the remaining three columns combine the deaths and reported cases data and use different weighting matrices.
Table 9.
Deaths | Cases | SUR, Efficient | SUR, Diagonal | SUR, Identity | |
---|---|---|---|---|---|
Panel A: No Rounding, True and , Assumed and | |||||
Levels, Cumulative | 0.0364 | 0.0352 | 0.0258 | 0.0255 | 0.0351 |
Levels, Daily | 0.1110 | 0.1010 | 0.0706 | 0.0693 | 0.1026 |
Logs, Cumulative, Trim | 0.0260 | 0.0261 | 0.0195 | 0.0188 | 0.0176 |
Logs, Daily, Trim | 0.0464 | 0.0451 | 0.0343 | 0.0340 | 0.0325 |
Logs, Cumulative, No Trim | 0.0065 | 0.0069 | 0.0049 | 0.0049 | 0.0048 |
Logs, Daily, No Trim | 0.0061 | 0.0064 | 0.0044 | 0.0044 | 0.0044 |
Panel B: No Rounding, True and , Estimated and | |||||
Levels, Cumulative | 0.0366 | 0.0351 | 0.0270 | 0.0275 | 0.0396 |
Logs, Cumulative, Trim | 0.0261 | 0.0261 | 0.0227 | 0.0239 | 0.0177 |
Logs, Cumulative, No Trim | 0.0075 | 0.0079 | 0.0056 | 0.0056 | 0.0053 |
Panel C: Rounding, True and , Assumed and | |||||
Levels, Cumulative | 0.050 | 0.035 | 0.037 | 0.032 | 0.035 |
Levels, Daily | 0.112 | 0.101 | 0.070 | 0.070 | 0.103 |
Logs, Cumulative, Trim | 0.060 | 0.027 | 0.038 | 0.031 | 0.033 |
Logs, Daily, Trim | 0.082 | 0.046 | 0.045 | 0.044 | 0.047 |
Logs, Cumulative, No Trim | 0.097 | 0.021 | 0.022 | 0.021 | 0.048 |
Logs, Daily, No Trim | 0.073 | 0.024 | 0.023 | 0.024 | 0.038 |
Panel D: Rounding, True and , Estimated and | |||||
Levels, Cumulative | 0.050 | 0.035 | 0.043 | 0.043 | 0.040 |
Logs, Cumulative, Trim | 0.060 | 0.027 | 0.038 | 0.034 | 0.033 |
Logs, Cumulative, No Trim | 0.097 | 0.021 | 0.022 | 0.021 | 0.048 |
The table presents the standard deviation of the estimates of across 500 simulation draws for different estimation methods. The estimates in the first column are based on the deaths data only. The estimates in the second column are based on the reported cases data only. The estimates in the remaining three columns combine the deaths and reported cases data and use different weighting matrices.
Table 10.
Deaths | Cases | SUR, Efficient | SUR, Diagonal | SUR, Identity | |
---|---|---|---|---|---|
Panel A: No Rounding, True and , Assumed and | |||||
Levels, Cumulative | 0.0013 | 0.0012 | 0.0007 | 0.0006 | 0.0012 |
Levels, Daily | 0.0124 | 0.0102 | 0.0050 | 0.0048 | 0.0105 |
Logs, Cumulative, Trim | 0.0007 | 0.0007 | 0.0004 | 0.0004 | 0.0003 |
Logs, Daily, Trim | 0.0022 | 0.0020 | 0.0012 | 0.0012 | 0.0011 |
Logs, Cumulative, No Trim | 0.00004 | 0.00005 | 0.00002 | 0.00002 | 0.00002 |
Logs, Daily, No Trim | 0.00004 | 0.00004 | 0.00002 | 0.00002 | 0.00002 |
Panel B: No Rounding, True and , Estimated and | |||||
Levels, Cumulative | 0.0013 | 0.0012 | 0.0007 | 0.0008 | 0.0017 |
Logs, Cumulative, Trim | 0.0007 | 0.0007 | 0.0005 | 0.0006 | 0.0003 |
Logs, Cumulative, No Trim | 0.00007 | 0.00008 | 0.00004 | 0.00004 | 0.00003 |
Panel C: Rounding, True and , Assumed and | |||||
Levels, Cumulative | 0.009 | 0.001 | 0.002 | 0.001 | 0.001 |
Levels, Daily | 0.013 | 0.010 | 0.005 | 0.005 | 0.011 |
Logs, Cumulative, Trim | 0.019 | 0.001 | 0.002 | 0.002 | 0.005 |
Logs, Daily, Trim | 0.007 | 0.002 | 0.002 | 0.002 | 0.002 |
Logs, Cumulative, No Trim | 0.457 | 0.002 | 0.001 | 0.002 | 0.124 |
Logs, Daily, No Trim | 0.079 | 0.001 | 0.001 | 0.001 | 0.020 |
Panel D: Rounding, True and , Estimated and | |||||
Levels, Cumulative | 0.009 | 0.001 | 0.002 | 0.003 | 0.002 |
Logs, Cumulative, Trim | 0.019 | 0.001 | 0.002 | 0.003 | 0.005 |
Logs, Cumulative, No Trim | 0.459 | 0.002 | 0.001 | 0.002 | 0.125 |
The table presents the MSE of the estimates of across 500 simulation draws for different estimation methods. The estimates in the first column are based on the deaths data only. The estimates in the second column are based on the reported cases data only. The estimates in the remaining three columns combine the deaths and reported cases data and use different weighting matrices.
Overall, based on the simulation results, methods based on the cumulative numbers seem to outperform methods based on daily values. Relative performance of different estimation approaches (e.g. based on levels versus logarithms) depends on the true DGP. One advantage of the approach based on levels is that it does not involve trimming.
The use of the SUR approach to estimation with the efficient weighting matrix may be motivated by the desire to obtain more efficient parameter estimates. Interestingly, this is not always the case in my simulations. While joint estimation of the model based on the deaths and reported cases data yields better (in terms of the MSE) estimates than estimation based on the deaths data alone, the MSE of the SUR estimates is often very similar to the MSE of the estimates based on the reported cases data alone. Moreover, the choice of the weighting matrix does not always affect the MSE of the resulting estimates: the estimates based on the naive weighting matrix are often as good as the estimates based on the efficient choice of .
7. Empirical results
This section presents the empirical results. Before I move on to the main results, I discuss computational issues associated with estimation of the model. Based on the arguments from the previous sections, when initial parameters change, the estimate of should remain virtually unchanged, while the estimates of and should change proportionally to the changes in initial parameters. In practice, however, this is not always the case. Both and are constrained to lie between 0 and 1, and when these constraints are binding or close to binding, changes in the initial values can have a substantial effect on the estimate of .
Table 11, Table 12 compare the estimation results for USA and California for the values of epidemiologic parameters and when the initial conditions change from , to , . The parameter estimates for California behave as expected: the estimate of remains virtually unchanged, while the estimates of and are cut in half. However, for the US as a whole the picture is different. The estimates of change substantially as the initial conditions change, while the changes in the estimates of and do not follow the expected pattern. For instance, when the model is estimated in levels, the estimates of and remain virtually unchanged as the initial condition changes. This example illustrates that when the parameter estimates are close to the boundary, changes in initial conditions may lead to somewhat unexpected estimation results.
Table 11.
Estimates of , , and for USA.
|
|
|||||
---|---|---|---|---|---|---|
SUR, Efficient |
SUR, Diagonal |
SUR, Identity |
SUR, Efficient |
SUR, Diagonal |
SUR, Identity |
|
Panel A: Levels, Cumulative | ||||||
9.265 | 9.519 | 9.743 | 8.806 | 9.144 | 9.291 | |
0.00006 | 0.00004 | 0.00003 | 0.00006 | 0.00004 | 0.00003 | |
0.0033 | 0.0024 | 0.0018 | 0.0031 | 0.0020 | 0.0017 | |
Panel B: Logs, Cumulative, Trim | ||||||
8.599 | 8.443 | 8.368 | 8.343 | 8.371 | 8.386 | |
0.00016 | 0.00019 | 0.00021 | 0.00011 | 0.00011 | 0.00010 | |
0.0068 | 0.0083 | 0.0091 | 0.0047 | 0.0046 | 0.0045 | |
Panel C: Logs, Cumulative, No Trim | ||||||
10.595 | 9.584 | 8.311 | 10.415 | 9.673 | 8.331 | |
0.00003 | 0.00007 | 0.00030 | 0.00002 | 0.00003 | 0.00014 | |
0.0007 | 0.0020 | 0.0082 | 0.0004 | 0.0009 | 0.0040 |
The table presents the estimates of , , and for , , and different initial values.
Table 12.
Estimates of , , and for California.
|
|
|||||
---|---|---|---|---|---|---|
SUR, Efficient |
SUR, Diagonal |
SUR, Identity |
SUR, Efficient |
SUR, Diagonal |
SUR, Identity |
|
Panel A: Levels, Cumulative | ||||||
5.062 | 5.094 | 5.058 | 5.064 | 5.097 | 5.060 | |
0.0040 | 0.0038 | 0.0041 | 0.0020 | 0.0019 | 0.0020 | |
0.217 | 0.206 | 0.219 | 0.108 | 0.103 | 0.109 | |
Panel B: Logs, Cumulative, Trim | ||||||
4.890 | 4.842 | 5.060 | 4.888 | 4.840 | 5.064 | |
0.0054 | 0.0059 | 0.0040 | 0.0027 | 0.0029 | 0.0020 | |
0.292 | 0.318 | 0.218 | 0.147 | 0.159 | 0.109 | |
Panel C: Logs, Cumulative, No Trim | ||||||
5.277 | 5.308 | 5.601 | 5.279 | 5.309 | 5.602 | |
0.0024 | 0.0023 | 0.0015 | 0.0012 | 0.0012 | 0.0008 | |
0.154 | 0.148 | 0.099 | 0.077 | 0.074 | 0.049 |
The table presents the estimates of , , and for , , and different initial values.
Table 13 presents the estimates of for the US, California, and Japan. All three panels of the table use cumulative data to estimate the model. The upper panel uses levels, the middle panel uses logarithms with trimming, and the lower panel uses logarithms without trimming. Different rows within a panel correspond to different choices of the weighting matrix .
Table 13.
Estimates of .
, | , | , |
, , |
||
---|---|---|---|---|---|
Panel A: Levels | |||||
USA | SUR, Efficient | 4.834 | 9.265 | 16.698 | 5.947 |
SUR, Diagonal | 4.963 | 9.519 | 17.399 | 6.089 | |
SUR, Identity | 5.053 | 9.743 | 18.249 | 6.167 | |
California | SUR, Efficient | 3.043 | 5.062 | 8.536 | 3.852 |
SUR, Diagonal | 3.058 | 5.094 | 8.599 | 3.872 | |
SUR, Identity | 3.041 | 5.058 | 8.528 | 3.850 | |
Japan | SUR, Efficient | 1.716 | 2.685 | 4.566 | 2.249 |
SUR, Diagonal | 1.717 | 2.684 | 4.559 | 2.248 | |
SUR, Identity | 1.714 | 2.681 | 4.587 | 2.245 | |
Panel B: Logs, Trim | |||||
USA | SUR, Efficient | 4.501 | 8.599 | 15.175 | 5.571 |
SUR, Diagonal | 4.552 | 8.443 | 15.190 | 5.640 | |
SUR, Identity | 4.581 | 8.368 | 15.198 | 5.679 | |
California | SUR, Efficient | 2.960 | 4.890 | 8.199 | 3.748 |
SUR, Diagonal | 2.937 | 4.842 | 8.105 | 3.719 | |
SUR, Identity | 3.043 | 5.060 | 8.536 | 3.855 | |
Japan | SUR, Efficient | 1.688 | 2.569 | 4.215 | 2.179 |
SUR, Diagonal | 1.689 | 2.570 | 4.217 | 2.180 | |
SUR, Identity | 1.700 | 2.642 | 4.502 | 2.219 | |
Panel C: Logs, No Trim | |||||
USA | SUR, Efficient | 5.504 | 10.595 | 19.545 | 6.682 |
SUR, Diagonal | 5.124 | 9.584 | 17.543 | 6.285 | |
SUR, Identity | 4.544 | 8.311 | 15.038 | 5.638 | |
California | SUR, Efficient | 3.143 | 5.277 | 8.958 | 3.979 |
SUR, Diagonal | 3.157 | 5.308 | 9.018 | 3.997 | |
SUR, Identity | 3.294 | 5.601 | 9.595 | 4.167 | |
Japan | SUR, Efficient | 1.862 | 2.816 | 4.710 | 2.328 |
SUR, Diagonal | 1.859 | 2.823 | 4.731 | 2.329 | |
SUR, Identity | 1.879 | 2.821 | 4.991 | 2.329 |
The table presents the estimates of for different countries and different values of epidemiologic parameters and . Initial conditions: for USA and California, for Japan. The middle panel trims all observations with or . The lower panel uses all observations with and .
Because there is no agreement in the medical literature on the appropriate values of epidemiologic parameters and , I consider several scenarios in the table: “fast” with , , “medium” with , , and “slow” with , . I also consider a modified version of the SEIRD model that replaces in Eqs. (2.1), (2.2) with for . This version of the model assumes that people in the exposed compartment can be contagious, but to a lesser extent that people in the infectious compartment. While these exact choices are somewhat arbitrary, considering several values of epidemiologic parameters instead of just one allows me to better understand their effect on the estimation results and forecasts.
The initial conditions are , for the US and California and , for Japan. The choice of the initial conditions for Japan is tricky because the estimate of is at the upper bound of 1, and the choice of initial conditions affects the estimate of . Given that Bommer and Vollmer (2020) estimate that the detection rate in Japan in March was around 20%–25%, even the initial condition may be too low.
As we can see from the table, the estimates of for a given country or region change a lot as epidemiologic parameters change. For example, the estimate of for the US ranges from under 5 in the “fast” scenario to around 15–18 in the “slow” one. Moreover, for the fixed values of epidemiologic parameters and , the estimates of differ a lot between regions. For instance, in the “medium” scenario with and , the estimates of vary from 2.6 for Japan to around 9.5 for the US. While the magnitude of these differences might be surprising, heterogeneity itself is not. If various mitigation or suppression policies can reduce , then one could expect that different countries have different values of due to the differences in their approaches to dealing with COVID-19, as well as differences in social norms, population density, etc.
Next, I present my results graphically. I focus on the results for California because they are least prone to numerical issues, as discussed above. For the sake of space I only present the results from the model estimated in levels. Additional results are presented in the online supplement.
The upper panel of Fig. 4 plots the fitted values of deaths and reported cases for California from the models with different values of epidemiologic parameters. The estimates of are based on the efficient weighting matrix . As we can see, even though the four models have different values of epidemiologic parameters and different estimates of , they appear to be indistinguishable in the short run: the resulting paths of deaths and reported cases are identical. However, in the long run the story is different. The lower panel of Fig. 4 demonstrates that the predicted total number of deaths from the COVID-19 epidemic in the four models ranges from around 70 thousand to more than 300 thousand.
Fig. 4.
Results for California. The upper panel shows the fit of the actual cumulative deaths and reported cases by models with four different values of epidemiologic parameters and . The middle panel shows the fit of the logarithms of the actual cumulative deaths and reported cases for the same four models. The lower panel shows the forecasts from the same four models.
Next, Fig. 5 fixes the values of epidemiologic parameters and and considers the pessimistic and optimistic scenarios, given by different initial conditions, for which the resulting models are observationally equivalent. The pessimistic scenario corresponds to the initial condition , , while the optimistic scenario corresponds to , . Intuitively, the lower the initial values, the lower the cumulative number of people who have had the virus, the higher the estimated fatality rate, and the higher the forecasted death toll.
Fig. 5.
Pessimistic and optimistic scenarios for California. The upper panel shows the fit of the actual cumulative deaths and reported cases by models with different initial conditions. The middle panel shows the fit of the logarithms of the actual cumulative deaths and reported cases by these models. The lower panel shows the forecasts from these models. The values of epidemiologic parameters are , .
We can see that different initial conditions lead to observationally equivalent models in the short run. However, there are large differences in estimates of unobserved variables and in long run forecasts. For instance, the model with estimates that the number of people with COVID-19 (here I count people both in the exposed and infectious compartments) in California on March 22 was around 50 thousand and predicts over 150 thousand deaths in the long run. In contrast, the model with estimates that there were around 1.6 million people with COVID-19 and predicts less than 5 thousand deaths in the long run, a 32-fold difference.
Next, I demonstrate empirically why it is important to allow the fraction of all cases that is reported to differ from one. Fig. 6 presents the results for different estimates of in the “medium” scenario with and . One model corresponds to the pessimistic case scenario from Fig. 5 with and . The forecasted long run number of deaths for that model is around 150 thousand. The remaining two models estimate from the confirmed cases data assuming that all cases are reported, i.e. , and then recover from the deaths data. The resulting estimates of are lower, 4.18 when the initial condition is , , and 2.90 when the initial condition is , . The estimates of , , and for these models are reported in Table 14. As we can see from the upper panel of the figure, the resulting models, especially the latter one, provide a poorer fit of the observed data: they cannot generate enough curvature because of the low .
Fig. 6.
Results for California for different values of . The upper panel shows the fit of the actual cumulative deaths and reported cases by models with and without underreporting and different initial conditions. The middle panel shows the fit of the logarithms of the actual cumulative deaths and reported cases by these models. The lower panel shows the forecasts from these models. The values of epidemiologic parameters are , .
Table 14.
Estimates of for California, with and without Underreporting.
Possible Underreporting, |
No Underreporting, |
No Underreporting, |
|
---|---|---|---|
5.062 | 4.176 | 2.908 | |
0.004 | 0.018 | 0.018 | |
0.217 | 1 | 1 |
The table presents the estimates of , , and for California for , . The left panel allows for underreporting of the number of cases and estimates all parameters from the data on deaths and reported cases jointly. The right two panels assume that all cases are reported, estimate from the reported cases data, and then recover , conditional on the estimated value of , from the deaths data.
The bottom panel of Fig. 6 shows that the forecasted long run number of deaths from both these models is over 600 thousand, a lot higher than in the pessimistic model with higher that fits the data well. Thus, estimating based on the confirmed cases data under the assumption that all cases are reported leads to the downward bias in the estimate of , poor fit of the observed data, and severe overestimation of the long run number of deaths.
Finally, I study whether additional information can help calibrate the initial conditions using Iceland as an example. Iceland is an interesting country to study because it was among the first countries to launch wide-scale random, or nearly random, testing of its population.12 While it is debatable whether testing in Iceland is completely random, my goal here is to demonstrate how information from these tests could in principle be used to calibrate the initial values. I fix the epidemiologic parameter values at and . Because the number of deaths in the data is very low, I use the first 70, rather than 60, observations to estimate the model.
The tests done in Iceland by deCode Genetics between March 13 and March 21 found 48 positives among 5,571 people who were tested for COVID-19, for the positive test rate of 0.86%. The 95% Wilson score confidence interval for the positive test rate is .13 I assume that the fraction of Iceland’s population who had COVID-19 on March 21, when the results were published, was the same as in the test. I then use the test results to calibrate the initial values in the model such that the fraction of Iceland’s population with COVID-19 on March 21 in the model is the same as in the test.
Given the population of 341,250, the positive test rate of 0.86% translates into 2,940 cases, with the 95% Wilson score confidence interval of . I hold and calibrate so that the sum of exposed and infectious people on March 21 matches the these numbers. yields 2,939 cases, yields 2,217 cases, and yields 3,892 cases on March 21. For simplicity, I do not require that be an integer.
The results are presented in the right panel of Fig. 7. For comparison, the left panel of Fig. 7 plots the results for and . As we can see, both in the left and right panel all models are indistinguishable on the available data. However, in the left panel, the forecasted death toll varies from under 77 for to 592 for ; in the right panel, it varies from 152 for to 264 for . Thus, the use of additional information leads to a more than 4-fold reduction in the range of forecasted deaths for observationally equivalent models. This result demonstrates the value of auxiliary information that becomes available due to random testing.
Fig. 7.
Results for Iceland. The figure presents the results for Iceland. The left panel does not use any additional information. The right panel matches the number of active COVID-19 cases on March 21 to the one estimated based on testing a random sample of population. The upper panel shows the cumulative deaths fit by models with different initial values . The middle panel shows the cumulative reported cases fit by these models. The lower panel shows the deaths forecasts from these models.
8. Conclusion
In this paper, I show that the SEIRD model for COVID-19 is poorly identified from the short run data on deaths and reported cases. There can be many different models that are indistinguishable in the short run but result in markedly different long run forecasts. For instance, the forecasted number of deaths in California in observationally equivalent models ranges from under 5 thousand to over 150 thousand. Thus, this paper highlights that long run forecasts for COVID-19 heavily depend on arbitrary choices made by the researcher. Available data cannot be used to determine which model is correct because there are many models that look identical in the short run.
Next, I propose several nonlinear SUR approaches to estimate the basic reproduction number , which is identified conditional on the values of epidemiologic parameters. Unlike most papers in the literature, which use data either on deaths or on reported cases, the proposed estimation methods combine these two series. Simulations suggest that the proposed methods lead to precise estimates of .
I then estimate for the US, California, and Japan for different values of epidemiologic parameters. The resulting estimates of heavily depend on the epidemiologic parameters and are heterogeneous across regions: they are 2–4 times higher in the US and California than in Japan.
My model takes into account possible underreporting of the number of cases. I demonstrate that the estimates of based on the confirmed cases data under the assumption that all cases are reported may be biased downward. The resulting models may be inconsistent with the observed data and may dramatically overestimate the long run number of deaths.
Finally, I demonstrate that auxiliary information from random tests for COVID-19 can help calibrate the initial values of the model and reduce the range of possible forecasts that are consistent with the observed data. Random, or nearly random, tests were conducted in Iceland, and utilizing the information from these tests leads to a more than 4-fold reduction in the range of the forecasted number of deaths.
The model I consider is fairly simplistic and does not take into account important factors such as possible overloading of the health care system, mitigation efforts, behavioral responses to the epidemic, etc. There are more sophisticated and realistic epidemic models that may be able to predict the spread of COVID-19 and the long run number of deaths better than the model studied here. However, those models usually have even more parameters, so one may worry that their identification would be even more troublesome.
Acknowledgments
I thank the editor, an anonymous referee, Andy Atkeson, Eric Fisher, Jeremy Fox, Oleg Itskhoki, Chad Jones, Dan McFadden, David Slichter, Jim Stock, Ping Yan, and Tom Zohar for their comments and suggestions. All remaining errors are mine.
Footnotes
https://github.com/CSSEGISandData/COVID-19/tree/master/csse_covid_19_data/csse_covid_19_time_series.
The dependence of the modeled number of deaths and reported cases on the parameters is now made explicit.
Supplementary Appendix A.2 presents some additional computational details.
The confidence interval based on the asymptotic normal approximation is .
Supplementary material related to this article can be found online at https://doi.org/10.1016/j.jeconom.2020.07.038.
Appendix A. Additional results
A.1. Lack of identification in SIRD models
In this section, I study an approximate solution of the simplified SIRD model to illustrate which model parameters can be identified from the data. The model is:
(A.1) |
(A.2) |
(A.3) |
(A.4) |
(A.5) |
During the early stages of the epidemic, , so that the equation for the evolution is, approximately,
The solution is given by . It is possible to show then that the approximate solutions for and are given by
(A.6) |
(A.7) |
Because and depend on , , and through the products and , and cannot be identified separately from . can be identified for a fixed value of , but it cannot be identified separately from in general: one can increase and decrease , so that remains unchanged, and adjust and accordingly so that and remain unchanged.
Finally, note that for a large enough , , so that . Thus, for a large enough ,
(A.8) |
(A.9) |
These equations demonstrate that the log series are approximately linear in and that and affect the slope of the log series, while , , and only affect the level.
A.2. Computational challenges
In this section, I describe some of the computational challenges associated with estimating the SEIRD model. Estimation consists of two steps. First, I simulate the model for a given choice of , , and using the DifferentialEquations package in Julia, developed by Rackauckas and Nie (2017). More specifically, the routine takes the model (2.1)–(2.6) and simulates the paths of , , , , , and for the given values of , , and . Next, I compute the difference between the modeled and observed quantities (cumulative or daily, in levels or logarithms) and minimize the appropriate nonlinear objective function using the Optim package.14 I use the Nelder–Mead algorithm to find the solution, as I found that it outperforms other algorithms available in Optim, such as simulated annealing or particle swarm.
The Nelder–Mead algorithm in Optim does not allow for bounds on the parameters. Because is constrained to be nonnegative and and are constrained to lie between 0 and 1, I use a reparametrization to ensure that all parameter estimates satisfy the constraints. Namely, I write , , , where are the parameters of the routine. The choice of the parameters , , , used in simulations, corresponds to .
Appendix B. Supplementary material
The following is the Supplementary material related to this article.
Additional results.
References
- Acemoglu D., Chernozhukov V., Werning I., Whinston M.D. National Bureau of Economic Research; 2020. Optimal targeted lockdowns in a multi-group SIR model. (Working Paper Series). [Google Scholar]
- Atkeson A. National Bureau of Economic Research; 2020. How deadly is COVID-19? Understanding the difficulties with estimation of its fatality rate. (Working Paper Series). [Google Scholar]
- Atkeson A. UCLA; 2020. Lockdowns and GDP: Is there a tradeoff? [Google Scholar]
- Atkeson A. National Bureau of Economic Research; 2020. What will be the economic impact of COVID-19 in the US? Rough estimates of disease scenarios. (Working Paper Series). [Google Scholar]
- Avery C., Bossert W., Clark A., Ellison G., Ellison S.F. Policy implications of models of the spread of coronavirus: perspectives and opportunities for economists. COVID Econ. 2020;12:21–68. [Google Scholar]
- Berger D.W., Herkenhoff K.F., Mongey S. National Bureau of Economic Research; 2020. An SEIR infectious disease model with testing and conditional quarantine. (Working Paper Series). [Google Scholar]
- Bommer C., Vollmer S. Working Paper. University of Göttingen; 2020. Average detection rate of SARS-CoV-2 infections has improved since our last estimates but is still as low as nine percent on March 30th. [Google Scholar]
- Chowell G., Ammon C., Hengartner N., Hyman J. Transmission dynamics of the great influenza pandemic of 1918 in geneva, Switzerland: Assessing the effects of hypothetical interventions. J. Theoret. Biol. 2006;241(2):193–204. doi: 10.1016/j.jtbi.2005.11.026. [DOI] [PubMed] [Google Scholar]
- Chowell G., Fenimore P., Castillo-Garsow M., Castillo-Chavez C. SARS Outbreaks in Ontario, Hong Kong and Singapore: the role of diagnosis and isolation as a control mechanism. J. Theoret. Biol. 2003;224(1):1–8. doi: 10.1016/S0022-5193(03)00228-5. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Chowell G., Nishiura H., Bettencourt L.s.M. Comparative estimation of the reproduction number for pandemic influenza from daily case notification data. J. R. Soc. Interface. 2007;4(12):155–166. doi: 10.1098/rsif.2006.0161. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Davidson R., MacKinnon J.G. Oxford University Press; 1993. Estimation and inference in econometrics. [Google Scholar]
- Ducrot A., Magal P., Nguyen T., Webb G. Identifying the number of unreported cases in SIR epidemic models. Math. Med. Biol. 2019 doi: 10.1093/imammb/dqz013. [DOI] [PubMed] [Google Scholar]
- Eichenbaum M.S., Rebelo S., Trabandt M. National Bureau of Economic Research; 2020. The macroeconomics of epidemics. (Working Paper Series). [Google Scholar]
- Ellison G. National Bureau of Economic Research; 2020. Implications of heterogeneous SIR models for analyses of COVID-19. (Working Paper Series). [Google Scholar]
- Fernandez-Villaverde J., Jones C. National Bureau of Economic Research; 2020. Estimating and simulating a sird model of covid-19 for many countries, states, and cities. (Working Paper Series). [DOI] [PMC free article] [PubMed] [Google Scholar]
- Gallant A. Seemingly unrelated nonlinear regressions. J. Econometrics. 1975;3(1):35–50. [Google Scholar]
- Hausman J.A. Specification tests in econometrics. Econometrica. 1978;46(6):1251–1271. [Google Scholar]
- Korolev I. Binghamton University; 2020. What does the infection fatality ratio really measure? [Google Scholar]
- Lauer S.A., Grantz K.H., Bi Q., Jones F.K., Zheng Q., Meredith H.R., Azman A.S., Reich N.G., Lessler J. The incubation period of coronavirus disease 2019 (COVID-19) from publicly reported confirmed cases: estimation and application. Annal. Internal Med. 2020 doi: 10.7326/M20-0504. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Lewbel A. The identification zoo: Meanings of identification in econometrics. J. Econ. Lit. 2019;57(4):835–903. [Google Scholar]
- Lin Q., Zhao S., Gao D., Lou Y., Yang S., Musa S.S., Wang M.H., Cai Y., Wang W., Yang L. A conceptual model for the coronavirus disease 2019 (COVID-19) outbreak in Wuhan, China with individual reaction and governmental action. Intl. J. Infect. Dis. 2020 doi: 10.1016/j.ijid.2020.02.058. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Liu Y., Gayle A.A., Wilder-Smith A., Rocklöv J. The reproductive number of COVID-19 is higher compared to SARS coronavirus. J. Travel Med. 2020 doi: 10.1093/jtm/taaa021. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Magal P., Webb G. The parameter identification problem for SIR epidemic models: identifying unreported cases. J. Math. Biol. 2018;77(6–7):1629–1648. doi: 10.1007/s00285-017-1203-9. [DOI] [PubMed] [Google Scholar]
- Manski C.F., Molinari F. Estimating the COVID-19 infection rate: Anatomy of an inference problem. J. Econometrics. 2020 doi: 10.1016/j.jeconom.2020.04.041. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Marinov T.T., Marinova R.S., Omojola J., Jackson M. Inverse problem for coefficient identification in SIR epidemic models. Comput. Math. Appl. 2014;67(12):2218–2227. Efficient Algorithms for Large Scale Scientific Computations. [Google Scholar]
- Piguillem F., Shi L. Einaudi Institute for Economics and Finance (EIEF); 2020. The optimal COVID-19 quarantine and testing policies. [Google Scholar]
- Rackauckas C., Nie Q. Differentialequations. jl – a performant and feature-rich ecosystem for solving differential equations in julia. J. Open Res. Softw. 2017;5(1) [Google Scholar]
- Sanche S., Lin Y.T., Xu C., Romero-Severson E., Hengartner N., Ke R. High contagiousness and rapid spread of severe acute respiratory syndrome coronavirus 2. Emerg. Infect. Diseases. 2020 doi: 10.3201/eid2607.200282. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Toda A.A. 2020. Susceptible-infected-recovered (SIR) dynamics of COVID-19 and economic impact. ArXiv preprint arXiv:2003.11221. [Google Scholar]
- Wang H., Wang Z., Dong Y., Chang R., Xu C., Yu X., Zhang S., Tsamlag L., Shang M., Huang J. Phase-adjusted estimation of the number of coronavirus disease 2019 cases in Wuhan, China. Cell Discov. 2020;6(1):1–8. doi: 10.1038/s41421-020-0148-0. [DOI] [PMC free article] [PubMed] [Google Scholar]
- White H. Consequences and detection of misspecified nonlinear regression models. J. Amer. Statist. Assoc. 1981;76(374):419–433. [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Supplementary Materials
Additional results.