Skip to main content
Biostatistics (Oxford, England) logoLink to Biostatistics (Oxford, England)
. 2023 Dec 6;25(3):867–884. doi: 10.1093/biostatistics/kxad030

A Bayesian multivariate factor analysis model for causal inference using time-series observational data on mixed outcomes

Pantelis Samartsidis 1,, Shaun R Seaman 2, Abbie Harrison 3, Angelos Alexopoulos 4,5, Gareth J Hughes 6, Christopher Rawlinson 7, Charlotte Anderson 8, André Charlett 9, Isabel Oliver 10, Daniela De Angelis 11,12
PMCID: PMC11247182  PMID: 38058013

Summary

Assessing the impact of an intervention by using time-series observational data on multiple units and outcomes is a frequent problem in many fields of scientific research. Here, we propose a novel Bayesian multivariate factor analysis model for estimating intervention effects in such settings and develop an efficient Markov chain Monte Carlo algorithm to sample from the high-dimensional and nontractable posterior of interest. The proposed method is one of the few that can simultaneously deal with outcomes of mixed type (continuous, binomial, count), increase efficiency in the estimates of the causal effects by jointly modeling multiple outcomes affected by the intervention, and easily provide uncertainty quantification for all causal estimands of interest. Using the proposed approach, we evaluate the impact that Local Tracing Partnerships had on the effectiveness of England’s Test and Trace programme for COVID-19.

Keywords: Causal inference, Contact tracing, Data augmentation, Factor analysis, Policy evaluation

1 Introduction

This article considers the problem of estimating from observational data the causal effect of an intervention (or treatment/policy) on an outcome of interest in applications where: (i) there are multiple sample units, some of which receive the intervention; (ii) for each unit, the outcome of interest is measured at multiple time points; and (iii) the units receiving the intervention and the times at which the intervention is implemented are not randomized. This problem arises often in epidemiology (Barnard et al., 2022), public health (Bruhn et al., 2017), economics (Hsiao et al., 2012), marketing (Brodersen et al., 2015), political science (Xu, 2017), and social studies (Ben-Michael et al., 2023). Often the sample units in such applications represent groups of individuals; for example, the units might be hospitals or administrative regions. See Ferman et al. (2020) and Samartsidis et al. (2019) for a list of real-world examples.

Methods for causal evaluation in settings where (i)–(iii) apply can be broadly classified as causal factor analysis (or “matrix completion”) (Kim and Oka, 2014; Gobillon and Magnac, 2016; Xu, 2017; Athey et al., 2021; Nethery et al., 2023; Samartsidis et al., 2020; Pang et al., 2021, among others) or synthetic control (Abadie et al., 2010; Hsiao et al., 2012; Brodersen et al., 2015; Robbins et al., 2017; Ben-Michael et al., 2021, among others) approaches. For an overview of these methods and an explanation of how they account for potential confounding, see Clarke et al. (2023), Xu (2023), and Samartsidis et al. (2019).

Existing approaches have limitations. First, for some of them, it is hard to obtain uncertainty intervals for the causal effects of interest. Second, most of them are designed to deal with continuous outcomes. An exception is Nethery et al. (2023), who develop a factor model for a single count outcome. However, these authors do not consider binomial outcomes, which arise frequently in applications; nor do they account for uncertainty in the number of latent factors. Third, few of these methods allow joint modeling of multiple outcomes affected by the intervention. Such joint modeling may improve statistical efficiency, which is particularly valuable when evaluating policies in contexts where data are sparse, i.e. only available annually or quarterly over a limited number of years. In recent work (Samartsidis et al., 2020), we used the multivariate factor analysis model of De Vito et al. (2021) to jointly model multiple normally distributed outcomes. However, this model cannot be applied to binomial/count data. It also assumes that variability shared across any of the multiple outcomes is shared across all the multiple outcomes.

Here, we develop a general approach to causal inference from time-series observational data that tackles these limitations. This method uses a multivariate factor analysis model that can be applied to multiple outcomes of mixed type (continuous/binomial/count) and allows for sharing of the loadings parameters across these outcomes. The model is Bayesian and thus automatically provides uncertainty quantification for all quantities of interest. Our model accounts for uncertainty in the number of latent factors and allows sharing of loadings across any subsets of the multiple outcomes. The Markov chain Monte Carlo (MCMC) algorithm we propose for fitting the model combines modern data augmentation and sampling techniques to overcome slow mixing issues stemming from the well-known identifiability problems of factor analysis models. The potential benefits of joint outcome modeling are demonstrated via a simulation study. We apply the proposed method to real data on a new policy for COVID-19 contact tracing in UK, providing valuable insights on the effectiveness of this policy. Finally, although our model is developed for intervention evaluation problems, it can be of use in other fields where multivariate factor analysis models are of interest, such as cancer biology (Avalos-Pacheco et al., 2022).

The remainder is structured as follows. In Section 2, we introduce the motivating COVID-19 contact tracing policy evaluation problem. Section 3 presents our Bayesian model. In Section 4, we discuss model fitting. Section 5 presents a simulation study that demonstrates the benefit of jointly modeling multiple outcomes. Section 6 contains results from our motivating data set. Finally, Section 7 summarizes the article’s findings and lists some directions for future research.

2 COVID-19 contact tracing policy evaluation problem

The NHS Test & Trace (TT) programme was launched in May 2020 as part of the containment strategy for the COVID-19 pandemic in UK. One of its main functions was to ensure that individuals who tested positive for COVID-19 (the cases) were aware1 of their infection, and to inform them of their legal duty to self-isolate to prevent onward transmission. When TT communicated with a case (henceforth this communication will be referred to as tracing), the case was asked to list the individuals (contacts) with whom he/she had been in close physical proximity within 48 h of his/her symptom onset (for asymptomatic cases, date of positive test is used). TT would then attempt to reach these contacts to provide advice and tell them to self-isolate. Multiple studies have provided evidence in favor of TT (Fetzer and Graeber, 2021).

Starting in July 2020, local authorities gradually introduced local tracing partnerships (LTPs). These LTPs involved the local authority working with national TT to improve tracing of cases and contacts. An LTP team was composed of staff based locally, whose job was to trace cases resident in that local authority who had not been able to be traced by TT. This is done mainly by telephone, often using the team’s local knowledge and, in several authorities, by visiting a case’s home if telephone communication had not been successful. The question we address here is whether LTPs (the intervention) have had an impact on the effectiveness of TT. Which local authorities introduced LTPs, and when, depended mainly on the interest expressed by local authorities, rather than being randomized, which complicates estimation of the causal effects of LTPs.

We judge the effectiveness of TT in terms of four outcomes measured daily: (i) the proportion of cases on that day that were ultimately completed, where a case is considered complete if TT manage to contact him/her; (ii) the proportion of completed cases whose completion was timely, i.e., occurs within 48 h of the case being registered with TT; (iii) the total number of contacts elicited from cases completed on that day; and (iv) the proportion of these elicited contacts who were subsequently completed. We shall refer to these as case completion, timely case completion, number of contacts and contact completion, respectively. We have daily measurements of these four outcomes on 181 spatial units during the study period July 01, 2020 to November 15, 2020 (i.e., 138 days). Each of these units is either an upper-tier local authority (UTLA) or part of a UTLA.

Figure 1 provides some graphical summaries of the data. The roll-out of LTPs is summarized in Fig. 1(a). The first LTP was formed on July 15, 2020 and by November 15, 2020, 118 (66%) of the units were covered by an LTP. Figure 1(b) shows the number of cases for each day and unit; during July–September the number of cases is low for most of the units, but it increases thereafter. The data on the four outcomes (Fig. D.1 of the supplementary material available at Biostatistics online) show considerable variability across units for the same day and across time within the same unit. The existing methods for policy evaluation listed in Section 1 cannot be used to investigate whether LTPs improved some of these outcomes because they are not applicable to binomial data.

Fig. 1.

Fig. 1

Graphical summaries of the data. (a) Proportion of units (out of 181) that had already formed an LTP by each day. (b) Total number of new cases for each day and unit, where units have been ordered by average number of new cases.

3 Methods for estimation of causal effects

3.1 Latent factor model and causal effects

Let i=1,,N index the sample unit and t=1,T denote time (e.g., in days) relative to a fixed calendar time (e.g., July 1st 2020 in our application). Let N1 and N2=NN1 be the numbers of control units and treated units, respectively (i.e., units that do not and units that do experience the intervention). Without loss of generality, assume units are ordered so that 1,,N1 are control units. Let Ti denote the last time at which unit i has not experienced the intervention (for control units, Ti = T). Let Tmin=min{T1,TN} and N˜t=i=N1+1NI(Ti<t).

Suppose there are D1 continuous, D2 binomial and D3 count outcomes. Let yitd be the observed value of the dth continuous outcome (d=1,,D1). For the dth binomial outcome (d=1,,D2), let kitd be the observed number of successes in a known number, nitd, of independent Bernoulli trials. Let zitd be the observed value of the dth count outcome (d=1,,D3).

We shall define the causal effect of the intervention on any one of the D=D1+D2+D3 outcomes in unit i at time t in terms of the difference between its observed value and the outcome that unit i would have if it were not treated at time t. The latter value is called a potential untreated outcome (Holland, 1986). We denote the potential untreated continuous, binomial, and count outcomes as yitd(0),kitd(0), and zitd(0), respectively. We also define the potential treated outcomesyitd(s),kitd(s), and zitd(s) (s=1,Ti) as the outcomes that unit i would have at time t if time s had been the last time at which unit i had not experienced the intervention. We make the stable unit treatment value assumption (SUTVA), which implies that the potential outcomes of one unit do not depend on whether, or at what time, the intervention is applied to other units. We further make the no-treatment-anticipation assumption (Athey and Imbens, 2021) which means that yitd(s)=yitd(0),kitd(s)=kitd(0), and zitd(s)=zitd(0) for all ts. With these assumptions, the observed outcomes are related to the potential outcomes by yitd=yitd(0),kitd=kitd(0), and zitd=zitd(0) if tTi; and yitd=yitd(Ti),kitd=kitd(Ti), and zitd=zitd(Ti) if t>Ti.

There is a growing interest in using factor analysis (FA) to model the potential untreated outcomes (see Section 1 for references), or as the data generating mechanism to study the properties of synthetic control-type approaches (Abadie et al., 2010; Hsiao et al., 2012; Ben-Michael et al., 2021; Ferman, 2021). This is due to the FA model’s ability to adjust for observed covariates and to allow for the presence of unmeasured confounding with a particular structure. Here, we propose a multivariate, mixed-outcome FA model for the potential untreated outcomes. Specifically, for each i=1,,N and t=1,,T, we assume

yitd(0)N(μitd,σid2),μitd=λiftd+η1,dxit,(d=1,,D1)kitd(0)Bin(nitd,pitd),logit(pitd)=λigtd+η2,dxit,(d=1,,D2)zitd(0)NegBin(witdqitdξd1,(1+ξd)1),log(qitd)=λihtd+η3,dxit.(d=1,,D3). (3.1)

Here, ftd,gtd,htd are J-vectors of unobserved factors at time t; λi is an unobserved J-vector of loadings for unit i; xit is a P-vector of (possibly time-dependent) exogenous (not affected by the intervention) covariates; and η,d are its regression coefficients (=1,2,3; d=1,,D). NegBin(a,b) denotes the negative binomial distribution with mean a(1b)/b and variance a(1b)/b2. It follows from (3.1) that E(zitd)=witdqitd and Var(zitd)=witdqitd(1+ξd). So, ξd>0 is the degree of overdispersion relative to the Poisson distribution. Here, witd is a known offset; if there is no offset, witd = 1. In some applications, there might be enough information to estimate a separate dispersion for each unit; in such cases, we can replace ξd by ξid.

We assume that. for all i=1,,N,

Ti{(yit1(0),,yitD1(0),kit1(0),,kitD2(0),zit1(0),,zitD3(0)):t=1,,T}|xi1,,xiT,λi.

This implies that after controlling for observed potential confounders xi1,,xiT and unobserved potential confounders λi, there is no confounding of the causal effects of the intervention on the dth continuous, dth binary and dth count outcome of unit i>N1 at time t>Ti, defined as

αitd=yitd(Ti)yitd(0)=yitdyitd(0),γitd=kitd(Ti)kitd(0)=kitdkitd(0),δitd=zitd(Ti)zitd(0)=zitdzitd(0), (3.2)

respectively. An alternative to γitd would be to define the causal effect on the dth binomial outcome of unit i at time t in terms of the success probability of the Bernoulli trials. To do this, we would assume kitd(Ti)=kitdBin(nitd,pitd(Ti)) (for i>N1,t>Ti) and define the causal effect as

βitd=pitd(Ti)pitd. (3.3)

For the dth continuous outcome, we define (with slight abuse of notation) the average (over time) causal effect in unit i as αid=t=Ti+1Tαitd/(TTi) and the overall (over both time and units) causal effect as αd=(i=N1+1Nt=Ti+1Tαitd)/(t=1TN˜t). Corresponding average causal effects, βid, βd, γid, γd, δid, and δd, for binary and count outcomes are defined analogously.

Policy makers are also interested in identifying treated units whose response to the intervention (i.e., causal effect) is extreme compared to other treated units. This could be useful, for example, to identify units to which alternative interventions should be applied. For the dth continuous outcome, unit i and time t>Ti, define ritd(α) as the (scaled) rank of the αitd values of the N˜t units treated by time t, i.e.,

ritd(α)=j:Tj<tI(αjtdαitd)N˜t+1. (3.4)

In (3.4), we have scaled by (N˜t+1)1 to ensure that ranks are comparable between any times t and s (both >Tmin) for which N˜tN˜s. Further, let rid(α)=t=Ti+1Tritd(α)/(TTi) be the average (over time) rank of unit i. Ranks for β, γ, and δ are defined analogously.

3.2 LTP application

In our motivating problem, where intervention is the introduction of an LTP, N = 181, T = 138, D1=0 (no continuous outcomes), D2=3 (case completion, timely case completion and contact completion), D3=1 (number of contacts), and N1=63 units have no LTP during study period.

For the first binary outcome (case completion), nit1 is the number of cases registered on day t for unit i, and kit1 is the number of these cases that were ultimately completed. For the second binary outcome (timely case completion), nit2 is the number of cases completed on day t for unit i, and kit2 is the number of these cases that were completed within 48 h. For the count outcome (number of contacts), wit1 is the number of cases that were completed on day t for unit i, and zit1 is the number of contacts elicited from these wit1 cases. For the third binary outcome (contact completion), kit3 is the number of contacts that were ultimately completed from the nit3=zit1 contacts. No covariates xit are considered.

In Section 3.1, we assumed nitd and witd are not affected by the intervention. In the LTP application, however, nit2=kit1 and nit3=zit1, which may both be affected. Also, wit1 is related to kij1 and kij2 for jt, and so may be affected. It is also possible that nit1 is affected (successful tracing can prevent new cases). This changes the interpretation of the causal effects γit1,γit2,γit3, and δit1. As shown in Section A of the supplementary material available at Biostatistics online, they can be interpreted (under some additional assumptions) as separable direct effects on kit1,kit2,kit3, and zit1, respectively. These are different from the corresponding total effects, defined as γ˜itd=kitdk˜itd(0) and δ˜itd=zitdz˜itd(0), respectively, where k˜itd(0) and z˜itd(0) are obtained by replacing nitd and witd with their potential untreated outcomes in (3.1). We do not consider the total effects (even though they can be estimated), because we believe they can be misleading. For example, suppose that for some i>N1 and t>Ti,pit3(Ti)=pit3 and nit3>n˜it3, that is, LTP has no impact on contact completion probability but increases the number of contacts elicited. Then kit3>k˜it3(0) (because larger number of trials), and hence γ˜it3>0, suggesting that the LTP increased the number of contacts completed. In contrast, γit3, interpreted as the total number of additional contacts completed thanks to the LTP, would be zero.

3.3 Outline of inference

In order to estimate the causal effects defined in Section 3.1, we need to estimate the potential untreated outcomes of the treated units i>N1 in their postintervention periods t>Ti. We do this under the Bayesian paradigm as follows.

First, we set prior distributions for the parameters of the multivariate FA model (3.1). These are detailed in Section 3.4. We then fit, using the MCMC algorithm in Section 4, the FA model to pre-intervention data only (i.e., for each i, we only use data up to time Ti). The posterior samples that we obtain allow us to draw samples from the posterior predictive distribution of the potential untreated outcomes conditional on xit, nitd and witd. We transform them using (3.2), to obtain samples from the posterior distribution of causal effects αitd, γitd, and δitd. For βitd, we further need to account for uncertainty in pitd(Ti). For each i>N1,t>Ti, and d=1,,D2, we a priori assume that pitd(Ti)Beta(ap,bp). Thus, conditional on postintervention data, pitd(Ti)Beta(ap+kitd,bp+nitdkitd). We obtain samples from the posterior of βitd using (3.3), i.e., by subtracting the draws from the posterior of pitd from the samples obtained from the beta posterior of pitd(Ti).

Let αitd() denote the th sample (=1,,L) from the posterior distribution of αitd. We estimate αitd as =1Lαitd()/L (the posterior mean) and calculate 95% posterior credible intervals (CIs) using the 2.5% and 97.5% percentiles of the αitd(). Samples from the posterior (and hence point estimates and 95% CIs) of αid, αtd, αd, and ritd(α) can be obtained from the αitd() using the expressions provided in Section 3.1. For binomial and count outcomes, point estimates and 95% CIs are obtained analogously.

3.4 Prior distributions

One of the main challenges in fitting the FA model (3.1) is that the number of latent loadings J is not known. To account for uncertainty in J, we start by setting it to J, an upper limit for the number of factors we expect. Following Gao et al. (2016), we assign a three-level three-parameter beta prior to the loadings. In particular, for j=1,,J we let

λijN(0,1ϕij1),ϕijTPB(aλ,bλ,1ζj1),ζjTPB(cλ,dλ,1ρ1),ρTPB(eλ,fλ,ν), (3.5)

where xTPB(a,b,c) means that x(0,1) has density π(x|a,b,c)=Γ(a+b)Γ(a)Γ(b)cbxb1(1x)a1(1+(c1)x)ab. Following Zhao et al. (2016), we set aλ,,fλ=0.5 and ν=0.1.

Let Λ be the N×J matrix with rows λi. The prior (3.5) induces three levels of regularization in Λ (Zhao et al., 2016). The overall matrix shrinkage is controlled by ρ. The ζj induce loading-specific shrinkage. When ζs is estimated as 1 for some 1sJ, we will also have that λis0 for all i, thus effectively removing the sth column of Λ. This allows us to recover the effective number of factors supported by the data as the number of nonzero2 columns in Λ. Finally, the ϕij allow the scale of the jth loading to differ across the N units. This is useful because given the interpretation of λij as potential unobserved confounders, one may expect them not to be similar across units, especially between control and treated units.

It is possible that each element λij of λi affects only some of the outcomes. To address this issue, one potential solution would be to assume a different loadings vector for each outcome. However, this is inefficient, especially when Ti’s are small (see simulation study of Section 5). Another solution would be to assume that there are loadings that are specific to each outcome and loadings that are shared by all outcomes (De Vito et al., 2021). However, such an approach does not allow for loadings that are shared by subsets of the outcomes. Grabski et al. (2023) propose a model for normal outcomes that allows for shared loadings among a subset of outcomes. However, vj are effectively either zero or one, thus not allowing loadings to affect the outcomes to different extents (which is the case when 0<vj<1). Here, we allow for this via the prior on the factor parameters.

Let ftdj, gtdj, and htdj denote the elements of ftd,gtd, and htd, respectively (j=1,,J). We assume that ftdjN(0,s1,dj),gtdjN(0,s2,dj) and htdjN(0,s3,dj). For each j, let vj be a vector of size D with elements vj defined as

vj={s1,,j,D1s2,D1,j,D1<D1+D2s3,D1D2,j,>D1+D2.

We introduce variables Mj{1,,D} which, for each j, indicate the outcome “most affected” by loading j. Conditional on Mj = l, we set vjl = 1 and assume vjUni[0,1] for each l. When vj is close to zero for some , the values of the corresponding factor will be close to zero, and hence the effect of loading j on outcome will be small. An alternative (which we test on our real data) is to let vjUni[0,1] for all j and . This introduces more free parameters than needed but has worked well in previous studies (Roy et al., 2021).

For fixed d and j, we have chosen not to impose any temporal structure on {ftdj}t=1T (or {gtdj}t=1T,{htdj}t=1T) to simplify posterior computations. We expect the loss in efficiency to be small. The reason is that in policy evaluation problems, the number of control units at each time point is typically sufficiently high to allow accurate estimation of factor parameters without the need to impose further structure, see e.g., simulation study in Samartsidis et al. (2020).

For the remaining model parameters, we specify weakly informative prior distributions. For all i and d=1,,D1, we assume σid2Uni(0,102). We prefer this prior to the conjugate IG(ϵ,ϵ) with small ϵ, because the latter encourages values of σid2 near zero. This can lead to over-fitting as non-systematic fluctuations are encouraged to be absorbed by the loadings/factors term. For all =1,2,3 and d, we assume that regression coefficients η,dN(0,102I). For d=1,,D3, we let ξdUni(0,20), i.e., we assume that the variance of the counts is at most 20 times higher compared to a Poisson model. For each loading j, we set P(Mj=)=1/D for =1,,D.

Finally, we set ap=bp=1 i.e., let pitd(Ti)Uni(0,1). The choice of prior on pitd(Ti) is not crucial for the LTP data as nitd (t>Ti) are large, allowing us to estimate these parameters accurately. However, when nitd are low, the posterior uncertainty in pitd(Ti) under independent uniform priors will be high, leading to high uncertainty in βitd (see (3.3)). In such cases, efficiency can be gained by imposing structure on pitd(Ti). For example, one can assume that for fixed i and d, logit(pitd(Ti)) arise from an AR(1) process.

4 Posterior computations

We are interested in drawing samples from the high-dimensional posterior with density

π({λi,{ξid}d=1D3}i=1N,{{gtd}t=1T,η2,d}d=1D2,{{htd}t=1T,η3,d}d=1D3,θ|data), (4.6)

where

θ={{{ftd}t=1T,{σid2}i=1N}d=1D1,{ξd}d=1D3,{{ϕij}i=1N,ζj,{vj,l}=1D,Mj}j=1J,ρ}

and

data={{{yitd}d=1D1,{kitd,nitd}d=1D2,{zitd,witd}d=1D3,xit}t=1Ti}i=1N.

The reason for grouping some parameters in θ is that they are easy to sample. Posterior (4.6) is analytically intractable; we thus use MCMC to sample from it. We propose a Metropolis-within-Gibbs sampler in which subsets of parameters are drawn (updated) in blocks from their full conditionals. Here, we provide an outline of the algorithm; for further details see Section B of the supplementary material available at Biostatistics online.

The parameters in θ are straightforward to update. We update each σid2 (i=1,,N and d=1,,D1) using a Metropolis–Hastings step. The ftd (t=1,,T and d=1,,D1) and η1,d (d=1,,D1) are drawn from their multivariate normal full conditional distributions. For the loadings shrinkage parameters, we follow Gao et al. (2016). They show that by using an equivalent representation of the prior (3.5), one can update all ϕij (i=1,,N and j=1,,J), ζj (j=1,,J) and ρ using Gibbs steps. We draw each Mj (j=1,,J) from its full conditional (these parameters can only take D values), after integrating out vdj (d=1,,D). Conditional on Mj, the vdj (d=1,,D and j=1,,J) are either set to one or drawn from their truncated inverse-gamma full conditionals (see Section B of the supplementary material available at Biostatistics online).

The most challenging step in the proposed MCMC algorithm is to update λi,gtd, and htd. Their full conditionals are not available in closed form and hence Gibbs updates are not possible. Moreover, standard algorithms which employ a constant variance–covariance matrix in their proposal (e.g., MALA, preconditioned MALA, HMC) can be inefficient in FA models. This due to the label-switching, rotation, and scale ambiguity problems (Zhao et al., 2016, among others). Consider, for example, the choice of a proposal variance for an element htdj of htd. Tuning this value is nontrivial as htdj may represent a different factor (due to label-switching) or change its scale (due to scale ambiguity) at each iteration of MCMC. To overcome these problems, we make use of data augmentation techniques. In particular, we employ methods developed in the case of univariate models (Polson et al., 2013; Zhou and Carin, 2015). As explained below, introducing auxiliary variables enables efficient sampling of λi,gtd, and htd. Finally, we note that under the proposed data augmentation scheme, drawing the remaining parameters η2,d,η3,d, and ξd is also conducted efficiently, see the remaining of this section.

Let PG(a,b) denote the Pólya–Gamma distribution with parameters a and b. Following Polson et al. (2013), for each i=1,,N,tTi and d=1,,D2, we introduce latent variables ωitd which are a prioriPG(nitd,0) distributed. Using results from Polson et al. (2013), we can show that the likelihood contribution of each binomial data point conditional on ωitd is

π(kitd|nitd,λi,gtd,η2,d,xit,ωitd)exp{ωitd2(κitdωitdλigtdη2,dxit)2}, (4.7)

where κitd=kitdnitd/2. The likelihood (4.7) is a quadratic form of both gtd and η2,d, and these parameters have multivariate normal priors. Therefore, introducing ωitd allows us to draw each gtd (d=1,D2 and t=1,,T) and η2,d (d=1,,D2) from their multivariate normal full conditionals (see SM Section B). Further, the full conditional of ωitd is a PG(nitd,λigtd+η2,dxit) (Polson et al., 2013), a result which allows us to update these parameters easily.

Let CRT(a,b) denote the Chinese restaurant table distribution with parameters a and b. Following Zhou and Carin (2015), for each i=1,,N,tTi and d=1,,D3, we introduce latent variables LitdCRT(zitd,witdqitd/ξd). These can be drawn as Litd=l=1zitdbl, where blBernoulli(witdqitd/ξdwitdqitd/ξd+l1). Recall that qitd=exp(λihtd+η3,dxit). From Section 3.1 of Zhou and Carin (2015), we have that

π(zitd,Litd|witd,qitd,xit,ξd)=CRT(Litd;zitd,witdqitdξd)NegBin(zitd;witdqitdξd,11+ξd)=ξdzitdLitd!|S(zitd,Litd)|(1+ξd)zitd(log(1+ξd))LitdPois(Litd;witdqitdξdlog(1+ξd)), (4.8)

where S(a, b) are Stirling numbers of the first kind. In (4.8), qitd only appear in the second term (Poisson probability mass function). As shown in Section A of the supplementary material available at Biostatistics online, this enables us to update htd (d=1,,D3, and t=1,,T), η3,d (d=1,,D3) and λi (i=1,,N) using the simplified manifold Metropolis adjusted Langevin (SMMALA) algorithm (Girolami and Calderhead, 2011), which would be very hard unless introducing the latent variables Litd and ωitd. The SMMALA algorithm exploits the geometry of the parameter space to adapt the variance–covariance matrix of its normal proposal at each iteration of MCMC, thus mitigating the problems caused by label-switching and scale ambiguity explained above. In our simulation studies, we found that SMMALA massively outperformed MALA and HMC, for which convergence was extremely slow.

Finally, we update each negative binomial dispersion parameter ξd (d=1,,D3) using the Barker method (Livingstone and Zanella, 2022). To reduce dependence on the Litd (t=1,,Ti), we integrate these parameters out when updating ξd. This is straightforward using (4.8), see Section A of the supplementary material available at Biostatistics online. The Barker method requires the specification of a stepsize parameter; following Livingstone and Zanella (2022), we tune this for each i during the burn-in phase of the MCMC to achieve an acceptance rate near 40%.

5 Simulation studies

We perform a simulation study to demonstrate benefits, in terms of efficiency of causal effect estimates, that can be obtained by modeling multiple outcomes jointly (i.e., loadings are shared across outcomes) rather than individually (i.e., each outcome has its own loadings matrix). See Section C of the supplementary material available at Biostatistics online for additional simulations in settings that differ from the one considered here.

5.1 Data generating mechanism

We simulate B = 2, 500 synthetic data sets. For each one, we generate the untreated outcomes from the FA model of (3.1) with D1=D2=D3=1, T = 24 (2 years of monthly data), N = 80 and λi,ft,gt,htR7. Since we consider only one outcome of each type, we omit the outcome index d for the remainder of this section to ease notation.

For all i, we set λi1=1 (to control the mean), and draw λijN(0,1) for j=2,,7. Let f˜j=(f1j,,fTj),g˜j=(g1j,,gTj), and h˜j=(h1j,,hTj). For each j, we generate f˜j,g˜j and h˜j from a N(m1j1,s1j2R),N(m2j1,s2j2R) and N(m3j1,s3j2R), respectively, where 1 is a T-vector of ones and R is a T × T correlation matrix. The values of mdj and sj (=1,2,3) are shown in Table 1. These specifications imply that E(μit)=5,E(pit)=0.6, and E(qit)=4. Some of the sj are set to zero so that each loading j affects only a subset of the outcomes. For fixed outcome, the nonzero sj are all equal for simplicity. The values of s are chosen such that the 97.5% quantiles of μit, pit, and zit (over units, times, and simulations) are roughly 7.5, 0.85, and 10, respectively. R has elements Rts=exp(log(0.8)|ts|). This nondiagonal R introduces temporal correlations within the data on each unit.

Table 1.

Mean and standard deviation used to simulate the seven factors of Section 5.

Mean
Standard deviation
j Normal Binomial Count Normal Binomial Count
1 m11=5 m21=logit(0.6) m31=log(4) 0 0 0
2 0 0 0 s 1 s 2 0
3 0 0 0 s 1 0 s 3
4 0 0 0 0 s 2 s 3
5 0 0 0 s 1 0 0
6 0 0 0 0 s 2 0
7 0 0 0 0 0 s 3

The value of σ is chosen such that λift accounts for 80% of the variability in the yit. We draw the nit from a Pois(n˜it) distribution. For each i, n˜i1=5,n˜iT is drawn from a Uni(50,200) distribution and n˜it=n˜i1+t1T1(n˜iTn˜i1) for 1<t<T. Similarly, we draw the wit from a Pois(w˜it) distribution where for each i, w˜i1=5,w˜iT is drawn from a Uni(25,75) distribution and w˜it=w˜i1+t1T1(w˜iTw˜i1) for 1<t<T. We have made n˜it and w˜it increasing to mimic the LTP data of Section 5. Finally, we set ξ = 2.

In each data set, Ti are chosen as follows. For every i and t, we draw uitUni(0,1) and let

ϖit={expit(κ0+κ1pit+κ2qit)t>tmin0ttmin, (5.9)

where expit(·)=exp(·)/(1+exp(·)). Then, for each i we set Ti=min{t:uit<ϖit}. We set the minimum number of pre-intervention time points to tmin=8. The value of κ0 is chosen such that the average over simulated data sets N1 is 40. The values of κ1 and κ2 control the degree of unobserved confounding of the effects βit and δit, respectively. We set κ1 such that the average (over simulated data sets) value of the empirical completion probability kit/nit is roughly 10% higher in controls units for all t>tmin. The value of κ2 is chosen such that on average (over simulated data sets), zit/wit is roughly 10% higher in controls units for every t>tmin.

For each simulated data set b, we perform a multivariate (MV) and a univariate (UV) analysis. MV analysis is carried out by fitting the FA model of Section 3 to all three outcomes jointly. UV analysis is carried out by fitting the FA model of Section 3 to each one of the outcomes individually. In both cases, the models that we fit are correctly specified. For both MV and UV analyses, we run MCMC for 100,000 iterations, saving posterior draws every 50 iterations to obtain 2,000 posterior draws. Of these, 500 are discarded as burn-in. The maximum number of factors J in MV and UV analyses is set to 25 and 15, respectively.

For each b, we generate the data of treated units in their post-intervention period under L = 50 different scenarios, corresponding to causal effects of increasing magnitude. Thus, we obtain L different estimates (and CIs) for all causal effects defined in Section 3. For each =1,,L, let α˜it(),β˜it(), and δ˜it() be the causal effect of intervention on μit, logit(pit) and log(qit), respectively. We use the tilde (·˜) notation to indicate that these are the effects on the mean function rather than on the outcome as in (3.2). For the binomial and count outcomes, the effects are applied on the logit-scale and log-scale, respectively, to avoid negative values. For each , we assume that for all i:Ti<T and t>Ti,α˜it()=α˜(),β˜it()=β˜(), and δ˜it()=δ˜(). The magnitude of α˜(),β˜(), and δ˜() increases with and α˜(1)=β˜(1)=δ˜(1)=0. Despite considering L scenarios for the intervention effect, we only need to run the MV and UV analyses once for each b, as the draws from the posterior of the potential untreated outcomes do not depend on postintervention data.

5.2 Results

We report results only for Ci={αi,βi,γi,δi}, representing the average response of each unit to the intervention. We compare the performance of the MV and UV approaches in terms of (i) the bias of the point estimates of the Ci; (ii) the standard error of the point estimates of the Ci; (iii) the width of the 95% CIs of Ci; (iv) the probability of detecting an intervention effect (power). For scenario =1 (i.e., no intervention effect), we define a detection as a CI that does not include zero to obtain the false positive rate. For any other scenario (i.e., positive intervention effect), we define a detection as a CI whose lower bound is larger than zero. For each estimand in Ci, we summarize each performance measure by taking the weighted average over simulated data sets and treated units. The weights are introduced to account for the varying number of treated units. More specifically, we assign a weight of 1/(BN2(b)) to each treated unit in simulated data set b, where N2(b) is the total number of treated units in that data set. To investigate the impact of Ti, we further calculate the measures over simulations and units with Ti = 8, Ti = 16, and Ti = 23.

The results for scenario =1 (no intervention effect) are presented in Table B.1 of the supplementary material available at Biostatistics online. Overall, there are no major problems with the estimates provided by UV and MV approaches: the bias of the point estimates of Ci is negligible (compared to the standard deviation), and the false positive rates are close to the nominal 5%. The exceptions are the estimates of βi, γi, and δi provided by UV approach for Ti = 8 i.e., for units with limited data in the preintervention period. For these, we find considerable bias (compared to the standard error) and inflated false positive rates. The estimates of αi are not affected because there is no confounding of these effects: the ϖit in (5.9) are not a function of μit.

In terms of efficiency, we see that the MV approach outperforms the UV approach. In particular, the standard errors of the point estimates of Ci, as well as the width of CIs, are on average smaller for the MV approach, as can be seen in Table 1 of Appendix B of the supplementary material available at Biostatistics online. The gains in efficiency due to joint outcome modeling are higher when Ti is small. This is expected, because when Ti is large there is sufficient data per treated unit on each outcome to estimate the loadings and so there is less benefit from using data on the other outcomes. Finally, it is worth noting that for all estimands and both approaches, the measures of efficiency (standard error, CI width) are better for Ti = 16 than they are for Ti = 8 and Ti = 23. The reason is that the efficiency in the estimates of Ci improves with both Ti and TTi (total number of postintervention time points). When Ti increases, there is more data to estimate the loadings and thus the potential untreated outcomes. When TTi increases, there is more data for each one of the Ci. In our simulation with fixed T, units of moderate Ti achieve the better balance between Ti and TTi.

The bias of point estimates, standard error of point estimates and width of CIs in scenarios >1 are very similar to scenario =1 and therefore not discussed3. Figure 2 shows the power achieved by the UV and MV approaches across all the different scenarios. We see that for all four estimands, the gains in efficiency due to joint outcome modeling substantially improve the probability of detecting a nonzero intervention effect. For example, β˜i=0.4 is detected with probability 44% using the UV approach, whereas it is detected with probability 65% when using the MV approach. Figure 2 in Appendix B of the supplementary material available at Biostatistics online shows the power for different values of Ti. For reasons explained above, we see that the improvements in power achieved by the MV approach compared to the UV are greater when Ti is either 8 or 16.

Fig. 2.

Fig. 2

Power of detecting an intervention effect for a randomly chosen treated unit (y axis) as a function of the magnitude of the intervention effect (x axis). (a)–(d) correspond to causal effects αi, βi, γi, and δi, respectively. The results are based on 2,500 simulated data sets.

For binomial outcomes, the uncertainty in the estimates of a treated unit’s potential untreated outcomes pit and kit(0), and thus the accuracy of the estimates of βit and γit, depends on both Ti and the values of nit in the pre-intervention period. More specifically, the lower the counts {nit}t=1Ti are, the less information to estimate the loadings there is. Similarly, for count outcomes, the uncertainty in the estimates of a treated unit’s zit(0) (t>Ti) depends on {wit}t=1Ti. Figure 3(a) presents a heatmap of the CI width for βi obtained by the MV approach for different combinations of Ti and n¯i=1Tii=1Tinit. For any fixed Ti, the width of CIs decreases with n¯i. Figure 3(b) shows the percentage decrease in CI width achieved by the MV approach compared to the UV approach for different combinations of Ti and n¯i. We see that for fixed Ti, the gains in efficiency due to joint outcome modeling are similar for the different values of n¯i. For the count outcome, we present the power achieved for moderate δ˜i (1.25) as a measure of efficiency; see Fig. 3(c). We find that power increases with w¯i for fixed Ti. Figure 3(d) shows the percentage increase in power for moderate δ˜ achieved by the MV approach compared to the UV. Again, we see that for fixed Ti, the gains do not differ much over across the w¯i.

Fig. 3.

Fig. 3

Effect of nit and wit on the efficiency of the causal estimates. (a) average (over simulated data sets and treated units) CI width of βi achieved by the MV model, as a function of n˜i=t=1Tinit/Ti and Ti. (b) average (over data sets and units) % reduction in CI width of βi as a function of n˜i and Ti achieved by the MV model compared to the UV. (c) power of detecting a moderate intervention effect from δi achieve by the MV model, for different values of w˜i=t=1Tiwit/Ti. (d) % increase in power of detecting a moderate effect on δi achieved by the MV model compared to the UV. In all panels, entries that were obtained as the average of less than 50 simulated data sets were discarded.

6 Evaluation of LTPs

In this section, we apply our multivariate FA model to the LTP data of Section 2. We set J=25 and ran three MCMC chains starting from different initial values. For each chain, we ran for 250,000 iterations, discarding the first 50,000 as burn-in, and thinning the remaining draws at every 100 iterations to obtain 2,000 draws from the posterior. Convergence was assessed visually by comparing the posterior densities and trace plots obtained from the three chains for multiple randomly selected causal estimands; these showed that all three chains have converged to the same stationary distribution and that mixing was good. Results for the case completion outcome are shown in Fig. 4. Figures D.2, D.3, and D.4 of the supplementary material available at Biostatistics online show results for the remaining outcomes.

Fig. 4.

Fig. 4

Results for outcome completion. (a) 95% CIs for pιt1 and pιt1(Tι), where ι is a unit chosen for illustration. (b) scatterplot of the posterior mean of βit1. Filled yellow dots represent βιt1. Panels (c) and (d) show posterior summaries of βi1 and γi1, respectively, where units have been ordered by increasing mean posterior βi1 in both plots. The vertical yellow lines indicate unit ι. Boxes and whiskers represent the 75% and 95% CIs, respectively.

Potential outcomes. Figure 4(a) refers to a unit, say ι, chosen for illustration. The plot shows 95% CIs for pιt1 and pιt1(Tι) in yellow and blue, respectively. For most days, the bands for pιt1(Tι) are above the bands for pιt1, indicating that the LTP had a positive effect on completion probability on these days. We also see that the 95% CIs for pιt1 are substantially wider than the CIs for pιt1(Tι); this is due to the large number of cases in the post-LTP period, allowing pιt1(Tι) to be estimated with high precision. Hence, uncertainty in βιt1 is mainly due to the uncertainty in pιt1.

Effects by unit and day. Scatterplots of the posterior means of βit1 and γit1 are shown in Fig. 4(b) and Figure D.5(a), respectively, where the βιt1 and γιt1 are shown in yellow. There is considerable heterogeneity in the estimated effects, both in terms of magnitude and sign. We observe an adverse effect during mid-September, a period during which several of the UTLAs operating LTPs reported that the number of people employed locally for contact tracing was insufficient to deal with the increasing number of cases. However, there are overall more positive than negative effects (especially toward the end of the study), suggesting that LTPs improved case completion TT on average.

There is also considerable heterogeneity of estimated effects for the other outcomes, see Section D of the supplementary material available at Biostatistics online. For timely case completion, there are both periods with positive and periods with negative effects (Fig. D.2(a) and (b) of the supplementary material available at Biostatistics online). We note that when we considered timely case completion out of nit1 (rather than kit1), the results (not shown) suggested that timeliness improves. This is expected, because LTPs lead to additional cases being completed, some of which are completed within 48 hours. For contact completion (Fig. D.3 of the supplementary material available at Biostatistics online) and number of contacts (Fig. D.4 of the supplementary material available at Biostatistics online), the estimated effects are centered around zero indicating that on average, LTPs did not affect the performance of TT for these outcomes.

A possible explanation as to why LTPs appear to improve TT performance in some units/days while negatively affecting it in some others is that the model of LTP employed varies between units (e.g., 5-day vs. 7-day working pattern or larger teams vs. smaller teams). Population factors, such as demographic characteristics of the cases, may also play a role. Finally, it is possible that staff delivering LTPs were becoming more effective as they acquired more experience.

Average unit effects. The posterior distribution of the average (over time) unit effects βi1 is summarized in Fig. 4(c), where we have sorted units by increasing mean posterior βi1. There are units that benefited from the LTPs on average (Pr(βi1<0) small), units that showed an adverse effect on average (Pr(βi1>0) small), and units that seemed unaffected (95% CIs centered at zero). This is also the case for the other three outcomes (see Section D of the supplementary material available at Biostatistics online). Figure 4(d) summarizes the posterior distribution of γi1, where units are sorted as in Fig. 4(c). Note that LTPs can have a strong effect (positive or negative) on the γi1 (or γi2,γi3) of a unit even if βi1 (or βi2,βi3) on the same unit is close to zero. This is because the latter measure depends on the total number of cases on each day: a 2% improvement in completion probability, will result in tens of additional completed cases in units with thousands of new cases, but possibly none in units with tens of new cases.

Overall effects. The estimated average effects over all units and post-intervention time points, βd, γd (d = 1, 2, 3), and δ1, are presented in Table 2 of the supplementary material available at Biostatistics online. These results suggest that case completion benefited from the LTPs as we estimate an additional 3.61 (95% CI [3.02–4.26]) completed cases per unit per day). Further, it appears that LTPs led to a drop in the number of contacts (we estimate that on average, 1.99 less contacts per unit per day were obtained due to LTPs).

Table 2.

Posterior mean along with 95% CIs for the “overall” effects βd (d = 1, 2, 3), γd (d = 1, 2, 3) and δ1. We present overall effects for the entire study period, as well as the period October 1, 2020–November 15, 2020.

Whole period
10/2020 onward
Outcome Effect Estimate 95% CI Estimate 95% CI
Completion β 1 0.013 0.006–0.021 0.044 0.038–0.051
γ 1 3.612 3.015–4.262 5.734 4.885–6.618
Timeliness β 2 –0.074 –0.080––0.068 0.014 0.007–0.020
γ 2 0.218 –0.221–0.660 1.973 1.367–2.602
Contact completion β 3 -0.014 –0.022––0.007 -0.007 –0.015–0.001
γ 3 -1.992 –3.743––0.439 -2.416 –4.746––0.411
Number of contacts δ 1 -0.552 –7.852–4.893 -0.994 –10.870–6.480

Posterior ranks. Identifying the units that benefit the least from their LTPs is important to improve the way of implementing LTPs in these units (e.g., by increasing the size of the teams). We do this using the posterior ranks. In Fig. D.6 of the supplementary material available at Biostatistics online, we summaries the posterior distribution of rιt1(β),rιt2(β),rιt3(β), and rιt1(δ). For most of the days in unit ι’s postintervention period, the mean posterior rιt1(β) is high (0.85), suggesting that unit ι was one of the units who benefited most from the LTPs in terms of case completion. The ranks for the remaining estimands as not as high.

Sensitivity analyses. In Section D of the supplementary material available at Biostatistics online, we perform a series of sensitivity to study the robustness of our findings to some hyper-parameter values and modeling choices. In particular, we repeat the analysis setting ν = 1, setting J=15 and J=35, as well as allowing the vjlUni[0,1] for all j and . The results suggest that our findings are not sensitive to these specifications.

7 Discussion

Motivated by an application concerning COVID-19, we have proposed a novel methodology for evaluating the impact of an intervention using observational time-series data. This methodology generalizes an existing causal multivariate FA method for normally distributed outcomes (Samartsidis et al., 2020) to the mixed outcome setting. This is an important addition to the literature in the field because, to our knowledge, it is the only method that can simultaneously (i) deal with outcomes of mixed type, (ii) make use of shared variability between subsets of multiple outcomes to improve the statistical efficiency of causal estimates, and (iii) provide uncertainty quantification for all causal estimands of interest.

We used the proposed methodology to estimate the impact of LTPs on the effectiveness of England’s NHS TT. The results suggest that on average, LTPs did not have a strong effect on contact completion or number of contacts elicited but improved case completion and timely case completion. However, there is big heterogeneity in the estimates of the causal effects both between and within (over time) units. It is likely that the heterogeneity is partly driven by differences in the model of LTP employed in each unit and day. Unfortunately, further investigation is not possible because this information has not been routinely collected. Nonetheless, our analyses highlight the importance of recording such data and we hope that they will encourage future collection.

Our approach leans itself to a number of extensions. First, it is likely that the potential untreated outcomes of units in spatial proximity are correlated, but our FA model does not make use of the geographical location of units. This could be addressed by assuming that for any j, Cor(λij,λιj) is a function of the distance between units i and ι. Second, we have assumed that continuous outcomes are normally distributed. This assumption might not hold in some data sets, even after transformations are applied. Hence, it is worth considering more flexible continuous distributions, e.g., the Student t and Gamma. Third, we will impose temporal structure on the factor parameters (e.g., with splines) parameters, to improve efficiency further.

The computation time required to apply our method to a typically sized policy evaluation data set is reasonable. For example, MCMC took approximately 90 min and 8 h to run on the simulated and real data, respectively, both on a 3.6 GHz Intel i7 machine with 16 GB of RAM. The main computational complexity arises from the need to invert a J×J matrix to update each of the N vectors of loadings and TD vectors of factors and is thus O((N+TD)J3). This can be prohibitive when applying to data from other fields such as biology. To overcome this issue, we will consider alternative approaches for estimation, e.g., using variational methods.

8 Software

Implementation codes are available at https://github.com/psamartsidis/cbfa.git.

Supplementary Material

kxad030_Supplementary_Data

Acknowledgements

P.S. was funded by the National Institute for Health and Care Research (NIHR) through a Programme Grants for Applied Research (PGfAR) (Grant Reference Number RP-PG-0616-20008); S.R.S. and D.D.A. were funded by Medical Research Council (MRC)-UK Research and Innovation (UKRI) (Unit Programme grant number MC_UU_00002/10 and MC_UU_00002/11 respectively) and by the UKHSA. This work was further supported by NIHR and Cambridge Biomedical Research Centre (BRC-1215-20014). The views expressed are those of the authors and not necessarily those of the NIHR or the Department of Health and Social Care. We used generators found in R packages pgdraw (Makalic and Schmidt) and GIGrvg (Leydold and Hormann, 2023) to sample from the Pólya-gamma and generalized inverse Gaussian distributions, respectively.

Footnotes

1

Cases are informed of their test result electronically (using the individual’s preferred means of communication), and may not necessarily be aware of their result at the time that TT attempt to get in touch with them.

2

In practice, one has to choose an arbitrary threshold c for |λij| to determine which columns are nonzero. This is not a problem in our application as we are not interested in making inference on J but rather account for uncertainty in this quantity when estimating the causal effects of interest.

3

For some of the estimands, this can be expected. Consider, for example, αi. Suppose that Ti=T1 and let αi, and yiT, be the simulated values of αi and yiT in scenario , respectively. For all , we will have that α^i,=yiT,y^iT(0)=N(μiT+αi,,σi)y^iT(0)(N(μiT,σi)+αi,)y^iT(0)αi,+(yiT,1y^iT(0)). The argument extends to t<T1. Hence, the bias in αi will only depend on the yit,y^it(0).

Contributor Information

Pantelis Samartsidis, MRC Biostatistics Unit, East Forvie Building, Cambridge Biomedical Campus, Cambridge, CB2 0SR, UK.

Shaun R Seaman, MRC Biostatistics Unit, East Forvie Building, Cambridge Biomedical Campus, Cambridge, CB2 0SR, UK.

Abbie Harrison, UK Health Security Agency, London, E14 4PU, UK.

Angelos Alexopoulos, MRC Biostatistics Unit, East Forvie Building, Cambridge Biomedical Campus, Cambridge, CB2 0SR, UK; Department of Economics, Athens University of Economics and Business, Athens, 104 34, Greece.

Gareth J Hughes, UK Health Security Agency, London, E14 4PU, UK.

Christopher Rawlinson, UK Health Security Agency, London, E14 4PU, UK.

Charlotte Anderson, UK Health Security Agency, London, E14 4PU, UK.

André Charlett, UK Health Security Agency, London, E14 4PU, UK.

Isabel Oliver, UK Health Security Agency, London, E14 4PU, UK.

Daniela De Angelis, MRC Biostatistics Unit, East Forvie Building, Cambridge Biomedical Campus, Cambridge, CB2 0SR, UK; UK Health Security Agency, London, E14 4PU, UK.

References

  1. Abadie A., Diamond A. and Hainmueller J. (2010). Synthetic control methods for comparative case studies: estimating the effect of California’s tobacco control program. J. Am. Stat. Assoc. 105, 493–505. [Google Scholar]
  2. Athey S., Bayati M., Doudchenko N., Imbens G. and Khosravi K. (2021). Matrix completion methods for causal panel data models. J. Am. Stat. Assoc. 116, 1716–1730. [Google Scholar]
  3. Athey S. and Imbens G. W. (2021). Design-based analysis in difference-in-differences settings with staggered adoption. J. Econ. 226, 62–79. [Google Scholar]
  4. Avalos-Pacheco A., Rossell D. and Savage R. (2022). Heterogeneous large datasets integration using Bayesian factor regression. Bayesian Analysis 17, 33–66. [Google Scholar]
  5. Barnard S., Chiavenna C., Fox S., Charlett A., Waller Z., Andrews N., Goldblatt P., Burton P. and De Angelis D. (2022). Methods for modeling excess mortality across England during the COVID-19 pandemic. Stat. Methods Med. Res. 31, 1790–1802. [DOI] [PMC free article] [PubMed] [Google Scholar]
  6. Ben-Michael E., Arbour D., Feller A., Franks A. and Raphael S. (2023). Estimating the effects of a California gun control program with multitask Gaussian processes. Ann. Appl. Stat. 17, 985–1016. [Google Scholar]
  7. Ben-Michael E., Feller A. and Rothstein J. (2021). The augmented synthetic control method. Assoc. 116, 1789–1803. [Google Scholar]
  8. Brodersen K. H., Gallusser F., Koehler J., Remy N. and Scott S. L. (2015). Inferring causal impact using Bayesian structural time-series models. Ann. Appl. Stat. 9, 247–274. [Google Scholar]
  9. Bruhn C. A. W., Hetterich S., Schuck-Paim C., Kürüm E., Taylor R. J., Lustig R., Shapiro E. D., Warren J. L., Simonsen L. and Weinberger D. M. (2017). Estimating the population-level impact of vaccines using synthetic controls. Proc. Nat. Acad. Sci. 114, 1524–1529. [DOI] [PMC free article] [PubMed] [Google Scholar]
  10. Clarke G. M., Steventon A. and O’Neill S. (2023). A comparison of synthetic control approaches for the evaluation of policy interventions using observational data: evaluating the impact of redesigning urgent and emergency care in Northumberland. Health Serv. Res. 58, 445–457. [DOI] [PMC free article] [PubMed] [Google Scholar]
  11. De Vito R., Bellio R., Trippa L. and Parmigiani G. (2021). Bayesian multistudy factor analysis for high-throughput biological data. Ann. Appl. Stat. 15, 1723 –1741. [Google Scholar]
  12. Ferman B. (2021). On the properties of the synthetic control estimator with many periods and many controls. J. Am. Stat. Assoc .116, 1764–1772. [Google Scholar]
  13. Ferman B., Pinto C. and Possebom V. (2020). Cherry picking with synthetic controls. J. Policy Analy. Manag .39, 510–532. [Google Scholar]
  14. Fetzer T. and Graeber T. (2021). Measuring the scientific effectiveness of contact tracing: evidence from a natural experiment. Proc. Nat. Acad. Sci. 118, e2100814118. [DOI] [PMC free article] [PubMed] [Google Scholar]
  15. Gao C., McDowell I. C., Zhao S., Brown C. D. and Engelhardt B. E. (2016). Context specific and differential gene co-expression networks via Bayesian biclustering. PLoS Comput. Biol. 12, 1–39. [DOI] [PMC free article] [PubMed] [Google Scholar]
  16. Girolami M. and Calderhead B. (2011). Riemann manifold Langevin and Hamiltonian Monte Carlo methods. J. R. Stat. Soc. B 73, 123–214. [Google Scholar]
  17. Gobillon L. and Magnac T. (2016). Regional policy evaluation: Interactive fixed effects and synthetic controls. Rev. Econ. Stat. 98, 535–551. [Google Scholar]
  18. Grabski I. N., De Vito R., Trippa L. and Parmigiani G. (2023). Bayesian combinatorial multi-study factor analysis. Ann. Appl. Stat. 17, 2212–2235. [DOI] [PMC free article] [PubMed] [Google Scholar]
  19. Holland P. W. (1986). Statistics and causal inference. J. Am. Stat. Assoc. 81, 945–960. [Google Scholar]
  20. Hsiao C., Steve Ching H and Ki Wan S. (2012). A panel data approach for program evaluation: measuring the benefits of political and economic integration of Hong Kong with mainland China. J. Appl. Econ. 27, 705–740. [Google Scholar]
  21. Kim D. and Oka T. (2014). Divorce law reforms and divorce rates in the usa: An interactive fixed-effects approach. J. Appl. Econ. 29, 231–245. [Google Scholar]
  22. Leydold J. and Hormann W. (2023). GIGrvg: random variate generator for the GIG distribution. R package version 0.8. [Google Scholar]
  23. Livingstone S. and Zanella G. (2022). The Barker proposal: Combining robustness and efficiency in gradient-based MCMC. J. R. Stat. Soc. B 84, 496–523. [DOI] [PMC free article] [PubMed] [Google Scholar]
  24. Makalic E. and Schmidt D. High-dimensional Bayesian regularised regression with the BayesReg package. arXiv, arXiv:1611.06649v3, preprint: not peer reviewed.
  25. Nethery R. C., Katz-Christy N., Kioumourtzoglou M.-A., Parks R. M., Schumacher A. and Anderson G. B. (2023). Integrated causal-predictive machine learning models for tropical cyclone epidemiology. Biostatistics 24, 449–464. [DOI] [PMC free article] [PubMed] [Google Scholar]
  26. Pang X., Liu L. and Xu Y. (2021). A Bayesian alternative to synthetic control for comparative case studies. Politic. Anal. 30, 269–288. [Google Scholar]
  27. Polson N. G., Scott J. G. and Windle J. (2013). Bayesian inference for logistic models using Pólya–gamma latent variables. J. Am. Stat. Assoc .108, 1339–1349. [Google Scholar]
  28. Robbins M. W., Saunders J. and Kilmer B. (2017). A framework for synthetic control methods with high-dimensional, micro-level data: evaluating a neighborhood-specific crime intervention. J. Am. Stat. Assoc .112, 109–126. [Google Scholar]
  29. Roy A., Lavine I., Herring A. H. and Dunson D. B. (2021). Perturbed factor analysis: accounting for group differences in exposure profiles. Ann. Appl. Stat. 15, 1386–1404. [DOI] [PMC free article] [PubMed] [Google Scholar]
  30. Samartsidis P., Seaman S., Montagna S., Charlett A., Hickman M. and De Angelis D. (2020). A Bayesian multivariate factor analysis model for evaluating an intervention by using observational time series data on multiple outcomes. J. R. Stat. Soc. A 183, 1437–1459. [DOI] [PMC free article] [PubMed] [Google Scholar]
  31. Samartsidis P., Seaman S. R., Presanis A. M., Hickman M. and De Angelis D. (2019). Assessing the causal effect of binary interventions from observational panel data with few treated units. Stat. Sci. 34, 486–503. [Google Scholar]
  32. Xu Y. (2017). Generalized synthetic control method: causal inference with interactive fixed effects models. Polit. Anal. 25, 57–76. [Google Scholar]
  33. Xu Y. (2023). Causal inference with time-series cross-sectional data: a reflection 10.2139/ssrn.3979613. [DOI]
  34. Zhao S., Gao C., Mukherjee S. and Engelhardt B. E. (2016). Bayesian group factor analysis with structured sparsity. J. Mach. Learn. Res. 17, 1–47. [Google Scholar]
  35. Zhou M. and Carin L. (2015). Negative binomial process count and mixture modeling. IEEE Trans. Patt. Anal. Mach. Intell. 37, 307–320. [DOI] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

kxad030_Supplementary_Data

Articles from Biostatistics (Oxford, England) are provided here courtesy of Oxford University Press

RESOURCES