One-sentence summary
In shaping school-related policy for the COVID-19 pandemic, a useful body of evidence can only be built if we ask the right questions regarding appropriate causal contrasts in relevant settings
Policy questions such as how best to re-open schools for in-person learning during a pandemic are incredibly important - but also incredibly hard to answer in an evidence-based fashion. As with many other policy decisions during the COVID-19 pandemic, school re-opening policies were generally crafted with minimal direct evidence. Recently, however, an increasing number of empirical studies are shaping an evidence base about the risks and benefits of in-person schooling. For example, Ertem et al. (2021) use data from the United States to examine the effects of in-person schooling in selected districts on SARS-CoV-2 cases in surrounding counties. In a similar vein, Fukumoto et al. (2021) provide data comparing SARS-CoV-2 cases in Japanese municipalities with schools that opened for in-person learning to similar municipalities that kept schools closed. As school policymakers must now weigh multiple studies on pandemic-related closure policies (often with apparently conflicting results), it is important to keep three considerations front-and-center: the causal question being asked, the treatment conditions being contrasted, and the context.
For policy decisions we are almost always interested in a causal question - that is, one that compares outcomes under two different potential states of the world. In one, a (well-defined) group experiences the “treatment” of interest (for example, re-opening schools for in-person learning); in the other, it experiences a “comparison” condition (for example, continued virtual learning). This comparison immediately raises the “fundamental problem of causal inference” -- that one can only observe outcomes under one condition, while the other is unobserved, or “counterfactual” (Holland, 1986). Thus, we are forced to use data from different groups to estimate differences that would be seen in the same group under different conditions (a “causal contrast”). Appropriate causal inference therefore requires strong study designs -- such as randomization, difference-in-differences/event study approaches (as in Ertem et al., 2021), and/or well-selected comparison groups (as in Fukumoto et al., 2021) -- to estimate the expected difference in outcomes for a given population under different policy choices.
While others have appropriately highlighted the importance of study design in answering pandemic-related causal policy questions (e.g., Bonvini et al., 2021; Goodman-Bacon and Marcus, 2020; Haber et al., 2021), we argue that policymakers should “keep it simple.” Specifically, most causally-focused studies can be evaluated in terms of their question, contrast (comparison), and context. By asking whether these three components of a given study seem reasonable - and the degree to which they apply to a current decision - policymakers without extensive methodological expertise can make a rapid assessment as to the relevance of a particular study.
Take, for example, a county school board evaluating the results of Ertem et al. (2021) in deciding whether to restrict in-person learning in the face of a new pandemic wave. This study speaks to county-level decisions to open or close in-person learning - not, for example, decisions at the state or national level. The contrast studied would only be directly applicable to an “all-or-nothing” closure decision, as few counties adopted an approach of keeping elementary schools open but middle and high schools closed. And in terms of context, different results were seen in the South than in other regions - and results in the United States might or might not be transportable to other countries. But by focusing on the question, contrast, and context, non-expert decision-makers could reasonably assess the relevance of this study to their policy decision. We should encourage this sort of thinking - and make it more accessible by highlighting these three elements in any analysis that seeks to estimate a policy-relevant causal effect.
Sadly there is often a disconnect between the questions, contrasts, and contexts addressed in research studies, and those that policymakers must consider. Regarding the question being asked, Ertem et al. (2021) and Fukumoto et al. (2021) both consider area-level policy decisions; other work on in-person schooling (e.g., Lessler et al., 2021) has focused on individual household behaviors. But these may not be the questions that local policymakers are being forced to answer; as highlighted by Orr et al. (2019), even randomized trials in schools are not always immediately relevant for local decision making. In terms of contrasts, some studies have compared “school reopening” to “school closure”overall, while others have aimed to estimate the effects of specific mitigation strategies. But rarely are the strategies formally studied the only policy options on the table. For example, Viner and Koirala (2021) compared daily testing versus isolation of contacts - but many school systems might be interested in less frequent testing, or different strategies for children versus staff members. With respect to context, analyses can - and should - evaluate differences in estimated effects across contexts, but these explorations are often limited by the available data. For example, Ertem et al. (2021) highlight interesting variation in estimated effects of school closures across regions of the United States, but also note an inability to differentiate between potential explanations for these differences, such as different mitigation strategies in place, weather-related differences, or differences in underlying incidence. And as a final consideration, all research studies use data from the past, whereas policy decisions must be made in the present. These challenges highlight the importance of performing research that is as close as possible in question, contrast, and context to actual policy decisions that are being considered. If these diverge too greatly, policymakers will default to decision-making in the absence of evidence, thus invalidating the considerable effort being made to bring an evidence base to bear in this process. Rarely are blanket conclusions such as “reopening schools does not fuel SARS-CoV-2 transmission” appropriate.
It is therefore critical that, in informing evidence-based decision-making, we maintain focus on the causal questions, contrasts, and contexts - while making use of the most appropriate data and study designs available. In the social sciences the UTOSTi (units, treatment, outcomes, settings, and times) framework (Cook, 2014) has helped articulate some of these considerations; we need a similarly simple guide for scientists and decision-makers asking policy-relevant questions regarding the COVID-19 pandemic. No one study will be relevant to all policy questions; we therefore must urgently build a diverse evidence base that mirrors the causal questions, contrasts, and contexts most likely to be encountered - and communicate those results to decision-makers in real-time, using language that can be broadly understood.
Footnotes
Conflicts of interest
The authors report no conflicts of interest. Dr. Dowdy’s time was supported by a Hopkins Business of Health Initiative pilot grant and a Johns Hopkins University Catalyst Award. Dr. Stuart’s time was supported by a Johns Hopkins University Discovery Award and by National Institutes of Health award P50MH115842 (PI: Daumit).
References
- Bonvini M, Kennedy E, Ventura V, and Wasserman L (2021). Causal inference in the time of Covid-19. https://arxiv.org/abs/2103.04472
- Cook TD (2014). Generalizing causal knowledge in the policy sciences: External validity as a task of both multi-attribute representation and multi-attribute extrapolation. Journal of Policy Analysis and Management 33(2): 527–536. [Google Scholar]
- Ertem Z, Schechter-Perkins EM, Oster E, van den Berg P, Epshtein I, Chaiyakunapruk N, Wilson FA, Perencevich E, Pettey WBP, Branch-Elliman W, and Nelson RE (2021). The impact of school opening model on SARS-CoV-2 community incidence and mortality. Forthcoming in Nature Medicine. [DOI] [PubMed] [Google Scholar]
- Espana G, Cavany S, Oidtman R, Barbera C, Costello A, Lurch A, Poterek M, Tran Q, Wieler A, Moore S, and Perkins TA (2021). Impacts of K-12 school reopening on the COVID-19 epidemic in Indiana, USA. Epidemics 37 (December 2021): 100487. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Fukumoto K, McClean CT, and Nakagawa K (2021). No causal effect of school closures on the spread of COVID-19. Forthcoming in Nature Medicine. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Goodman-Bacon A, and Marcus J (2020). Using difference-in-differences to identify causal effects of COVID-19 policies. Survey Research Methods 14(2), 153–158. 10.18148/srm/2020.v14i2.7723 [DOI] [Google Scholar]
- Haber N, Clarke-Deeder E, Salomon J, Feller A, and Stuart EA (in press). Policy evaluation in COVID-19: A guide to common design issues. Forthcoming in American Journal of Epidemiology. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Holland PW (1986). Statistics and causal inference. Journal of the American Statistical Association 81(396): 945–960. [Google Scholar]
- Lessler J, Grabowski MK, Grantz KH, Badillo-Goicoechea E, Metcalf CJE, Lupton-Smith C, Azman A, and Stuart EA (2021). Household COVID-19 risk and in-person schooling. Science. Published online 29 April 2021. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Orr LL, Olsen RB, Schmid I, Shivji A, Bell S, and Stuart EA (2019). Using the results from rigorous multi-site evaluations to inform local policy decisions. Journal of Policy Analysis and Management. 38:4: 978–1003. 10.1002/pam.22154 [DOI] [Google Scholar]
- Viner RM, and Koirala A (2021). Daily antigen testing to reduce disruption when schools return. The Lancet 398(10307): 1196–1197. [DOI] [PMC free article] [PubMed] [Google Scholar]