Summary:
Meta-regression is widely used in systematic reviews to investigate sources of heterogeneity and the association of study-level covariates with treatment effectiveness. Existing meta-regression approaches are successful in adjusting for baseline covariates, which include real study-level covariates (e.g., publication year) that are invariant within a study and aggregated baseline covariates (e.g. mean age) that differ for each patient but are measured before randomization within a study. However, these methods have several limitations in adjusting for post-randomization variables. Although post-randomization variables share a handful of similarities with baseline covariates, they differ in several aspects. First, baseline covariates can be aggregated at the study level presumably because they are assumed to be balanced by the randomization, while post-randomization variables are not balanced across arms within a study and are commonly aggregated at the arm level. Second, post-randomization variables may interact dynamically with the primary outcome. Third, unlike baseline covariates, post-randomization variables are themselves often important outcomes under investigation. In light of these differences, we propose a Bayesian joint meta-regression approach adjusting for post-randomization variables. The proposed method simultaneously estimates the treatment effect on the primary outcome and on the post-randomization variables. It takes into consideration both between- and within-study variability in post-randomization variables. Studies with missing data in either the primary outcome or the post-randomization variables are included in the joint model to improve estimation. Our method is evaluated by simulations and a real meta-analysis of major depression disorder treatments.
Keywords: Bayesian method, Joint modeling, Meta-regression, Missing data, Post-randomization variable
1. Introduction
Meta-regression has become a widely used meta-analytic technique to explore the sources of heterogeneity in treatment effect estimates and to investigate the relationship between the effect size and one or more study-level covariates. A covariate included in a meta-regression can be a real study-level characteristic that is the same for all subjects in a study (e.g., publication year) or an individual-level characteristic that varies between subjects but is aggregated as a study-level summary statistic (e.g., mean age). Real study-level covariates are typically related to the study design, such as length of follow-up, study design (e.g., randomized vs. observational), the geographical region of study, etc. Aggregated covariates can be either baseline variables such as mean age or proportion of male or post-randomization variables such as treatment discontinuation rate, compliance measured through average pill counts and other biochemical markers, and cross-over to an alternative therapy.
While including a real study-level covariate is straightforward, adjusting for a covariate that also varies at the patient level raises several concerns. First, the conclusion from the meta-regression is subject to aggregation bias: the estimated coefficient is attenuated by measurement error in the covariate. When the variation in a covariate between studies is small compared to variation within studies, it is difficult to identify the underlying between-study trends in a covariate. Researchers have discussed the impacts of between- and within-study heterogeneity on statistical inference and power (Thompson and Higgins, 2002; Lambert et al., 2002; Schmid et al., 2004). Second, as in other meta-analytic techniques, some degree of heterogeneity between studies in the effect size is expected, because the studies are conducted at different locations, by different researchers, or at different times. A fixed effects model (Greenland, 1987) can be used when the researchers believe that between-study variability in the effect sizes is due to differences in the study-level covariates only. Random effects models have been proposed to account for the residual heterogeneity, i.e., the part of variation in the effect sizes that can not be explained by study-level covariates (Berkey et al., 1995; Thompson and Sharp, 1999).
Finally, one issue in meta-regression that has received little attention is missing covariate data. Here, we need to carefully distinguish two types of missing data in the context of meta-analysis: (1) missing data at the individual patient level and (2) missing data at the study level. The aforementioned missingness refers to missing covariate data at the study level. Meta-analytic methods generally assume that individual level missingness has already been handled with appropriate methods within each study and the reported aggregated data are unbiased or consistent estimates of the true study-specific treatment effect. Yuan and Little (2009) showed that the widely used DerSimonian and Laird estimate (DerSimonian and Laird, 1986) can yield biased estimates when individual-level missingness depends on the size of the underlying study-specific treatment effect. Ignorable missingness at the individual level may have a non-ignorable impact in a meta-analysis. In this paper, the missingness generally refers to missing aggregated values at the study level. Meta-regression is primarily based on published reports. A study may fail to report full information on relevant covariates, especially when the covariates are not significant in that study. Although a meta-analyst may contact the researchers to obtain unreported data, this is not always successful. Also, in an early investigation of a treatment, important confounding factors may not yet be well recognized, so data on these covariates may not even be collected in early studies. Conclusions based on a subset of studies with complete covariate information can be misleading.
Although the scientific value and limitations of meta-regression adjusting for real study-level covariates or aggregated baseline covariates have been thoughtfully investigated (Thompson, 1993; Schmid et al., 2004), little attention has been devoted to post-randomization variables, i.e., variables measured after the randomization within each study. Post-randomization variables share a handful of similarities with baseline covariates, so adjusting for a post-randomization variable in a meta-regression also faces challenges and limitations analogous to those in traditional meta-regression adjusting for aggregated baseline covariates. In addition, post-randomization variables often serve as predictors of the primary outcome and as responses to the treatment in their own right. Because of this, they differ from baseline covariates in important ways. This article briefly reviews some existing meta-regression models and discusses their limitations in handling post-randomization covariates, then proposes a joint meta-regression approach to adjusting for post-randomization variables. The proposed method not only hurdles the aforementioned obstacles in traditional meta-regression methods, but also takes account for the distinctive features of post-randomization variables. Throughout, we focus on meta-analyses of two-arm, parallel group trials.
2. Meta-regression methods accounting for post-randomization variables
This section is organized as follows. Section 2.1 reviews commonly used meta-regression approaches and discusses the limitations of existing methods. Section 2.2 proposes a novel joint meta-regression model adjusting for post-randomization variables. A Bayesian estimation approach is given in Section 2.3.
2.1. Existing meta-regression approaches
Like multiple regression, meta-regression can be used to assess the relationship between the effect size and covariates at the study level. Fixed and random effects models have been proposed (Greenland, 1987; Berkey et al., 1995). Here, we review a commonly used random effects meta-regression approach. The model can be specified as:
| (1) |
where yi is the reported relative effect size of study i, Δi is the true underlying effect size of study i, ζi is a known nuisance parameter available in each study (typically variance), Xi is study i’s vector of study-level confounding covariates, ωi is the random effect, which follows a normal distribution with mean zero and standard deviation σ, and β0 and β are the fixed effects of interest. For example, if the primary outcome is binary, yi can be the log odds ratio of treatment to control reported from study i. The intercept β0 is interpreted as the overall log odds ratio of treatment efficacy when Xi = 0, and β corresponds to the effects of study-level covariates. When σ2 = 0, this is the fixed effects meta-regression model.
Although a broad consensus considers it acceptable to use such a meta-regression for real study-level covariates and aggregated baseline covariates, existing meta-regression approaches have several drawbacks in adjusting for post-randomization variables, such as compliance measured through average pill counts and other biochemical markers, the proportion of missing data during patient follow-up, cross-over to an alternative therapy, or premature withdrawal from study treatment. First, existing methods are restricted to studies with balanced covariates in which no systematic difference between arms is expected within a study. As in Eq (1), such a meta-analysis often investigates the impact of study-level covariates on a study-specific relative effect size. Post-randomization variables, which are measured following the randomization, generally differ between treatment arms within a study and are commonly aggregated at the arm level. Proper consideration of these post-randomization variables is needed. Second, unlike baseline covariates, post-randomization variables can themselves be outcomes of interest. For example, in a meta-analysis of cancer trials where the primary goal is to evaluate the treatment effect in a patient reported pain outcome, opioid usage in the trials might be a confounder of the treatment effect on the primary outcome and a measure of acceptability of the treatment at the same time. Therefore, post-randomization variables are generally considered random variables instead of fixed values. For studies with relatively small sample sizes, within-study variation in post-randomization variables may be non-ignorable compared with between-study variation. Third, because these variables are observed post-randomization, they may interact dynamically with the primary outcome measure over the course of patient follow-up at the study level. For example, an unsatisfactory treatment may induce poor compliance, which further attenuates the response to treatment. Even if each individual trial uses the same strategy to construct the estimand, the primary outcome can still be confounded by intercurrent events at the aggregated/study level. Finally, traditional meta-regression methods cannot include studies with missing covariate information. Implicitly, traditional meta-regression approaches assume the covariate data is missing completely at random (MCAR), i.e., failure to observe the value of a confounding covariate is independent of both the observed and unobserved data (Little and Rubin, 2002). However, the decision to report the value of a covariate is often related to the observed effect size. As we discussed in Section 1, any analysis using only studies for which complete covariate data is available is potentially biased.
2.2. Joint meta-regression approach using Bayesian hierarchical modeling
This subsection proposes a joint meta-regression approach. The goals of the proposed method are: (1) to estimate the overall effect of a treatment relative to a control for the main outcome and the post-randomization variables; (2) to estimate the overall absolute treatment effects in both treatment groups; and (3) to quantify the relationship between the main effect size and the post-randomization variables. The major difference between the proposed method and existing meta-regression methods is that it treats post-randomization covariates as random variables and jointly models them with the outcome of primary interest. The proposed method also allows the main outcome or the post-randomization variables to be missing at random (MAR) at the study level and naturally accounts for uncertainty due to missing data. Here, data is called MAR if, given the observed data, the missing covariate or outcome data does not depend on unobserved data (Little and Rubin, 2002).
We consider a meta-analytic dataset with a collection of I clinical trials to compare an experimental treatment to an active control or a placebo. For any type (i.e., binary, categorical, or continuous) of aggregate data (i.e., summarized over individuals), assume that the observations of a primary outcome variable from each treatment arm of each trial arise independently from a parametric model: Yik ~ fy(yik∣θik, ϒik); i = 1, ……, I, k = 1, 0, where yik is the observed aggregate primary outcome for arm k in trial i, fy(·) is a probability density having an unknown location parameter θik and a known nuisance (typically variance) parameter ϒik. For example, when the measurement Yik is continuous, fy(·) is often assumed to be the normal density. A binary Yik is often assumed to follow a binomial distribution Bin(nik, θik), where nik is the sample size for treatment group k in the ith study and the unknown location parameter θik is the unknown event probability pik, i.e., θik = pik. We define to be missingness indicators for the primary outcome variable at the study level (not at the arm level), taking value 1 when yi = (yi1, yi0) is available from study i and 0 otherwise.
As mentioned above, this method treats post-randomization variables as random. Let Wik be a random aggregated post-randomization variable for the kth treatment arm of the ith trial. We assume that Wik arises independently from a parametric model: Wik ~ fw(wik∣qik, ψik), i = 1, …, I; k = 0, 1, where fw(·) is a probability density having an unknown location parameter qik and a known nuisance parameter (typically variance) ψik. We define as the missingness indicators for the post-randomization variable at the study level (not at the arm level), taking value 1 when wi = (wi1, wi0) is available from study i and 0 otherwise.
Let and be row vectors containing the study-level baseline covariates (e.g., average age and location) associated with the post-randomization variables and the primary outcome for the ith trial, respectively. Similar to traditional meta-regression approaches, and are assumed to be fixed because they are not affected by the random treatment allocation. We assume and are available for all i. To allow the study-level baseline covariates and the latent location parameter qik of the post-randomization variables to have differential effects on the primary outcome, i.e., to let the and Wik affect absolute treatment effects on the primary outcome, we consider a very general joint model:
| (2) |
where hθ(·), hq(·) and g(·) are link functions associated with the primary outcome and the post-randomization variables, respectively. The random effects ωi = (ωi1, ωi0) and νi = (νi1, νi0) are mutually independent and follow bivariate normal distributions with covariance matrices Σθ and Σq, respectively. The off-diagonal elements and diagonal elements of Σq capture the potential correlation between arms and heterogeneity across studies respectively in the post-randomization variable. The between-arm correlation and between-study variability in the primary outcome can be partially explained by the post-randomization variable, while the diagonal and off-diagonal elements of Σθ capture the residual heterogeneity across studies and correlation between arms, respectively. The adjusted overall absolute effect (arm-based effect) on the primary outcome in arm k is μ0k when the covariate terms are zero, while μ1k measures the strength of the relationship between primary outcome and the post-randomization variable in arm k. If the post-randomization variable is assumed to have a non-differential effect on the primary outcomes in the two arms, μ1k (k = 0, 1) in Eq (2) can be replaced by a single parameter μ1. If μ1k = 0, treatment efficacy is exchangeable without conditioning on the post-randomization variable and standard meta-analysis models provide consistent estimates as well. The parameters and , which account for confounding effects, can also be simplified to φ under analogous assumptions.
Implicitly, our modeling strategy assumes the primary effect size of interest underlying the data with g(qik) = 0 is similar across trials, so that systematic deviation from the overall effect size is due, e.g., to the fraction of participants who discontinue study treatment. By jointly modeling the post-randomization variable and the main outcome, the proposed method takes into consideration the randomness in the outcome of post-randomization variable.
2.3. Estimation
Let yi = (yi1, yi0), wi = (wi1, wi0), μ0 = (μ01, μ00), μ1 = (μ11, μ10), α0 = (α01, α00), and . Unlike traditional meta-regression, the underlying assumption for the joint modeling approach is that missing data are missing at random at study level, i.e., given the observed data, failure to observe the primary outcome or the post-randomization variable in a study is independent of the unobserved data. Given this assumption, the likelihood contribution conditional on the model parameters from study i with observed primary outcome and post-randomization variable is
where θik and qik are functions of μ0, α0, ωi, νi, , φθ, φq, and μ1, as shown in Eq (2). The parameter λm is a vector of parameters related only to the missingness pattern. As shown above, the distribution of the observed data can be factored into a marginal density for (yi, wi) and a conditional density for (, ) given (yi, wi). As a result, the missingness mechanism is ignorable and inferences about the model parameters are based solely on . This suggests there is no need to specify a model for the mechanism generating missing data, under the missing-at-random assumption. Similarly, we can show that the conditional likelihood contribution from study i with observed primary outcome but missing post-randomization variable is . Detailed derivation is available in Web Appendix C. Finally, the conditional likelihood contribution for a study with missing outcome data but observed post-randomization variable is li3 ∝ ∏k fw(wik∣qik, ψik). The total conditional likelihood given is then . The marginal likelihood integrated over the multi-dimensional random effects is: . To account for the complexity of the model, we adopt a full Bayesian approach using Hamiltonian Monte Carlo (HMC) algorithms implemented using the STAN software in the rstan package in R (Stan Development Team, 2016).
To complete the model specification, we assign vague normal priors to the fixed effects α0k, μ0k and the coefficient μ1k. The covariance matrices Σθ and Σq can be decomposed as Σθ = PΩθP and Σq = QΩqQ respectively, where P and Q are diagonal matrices with diagonal element {σk, k = 1, 0} and {τk, k = 1, 0} respectively. Ωθ and Ωq are correlation matrices, which capture the correlation between the two arms. We place LKJ priors (Lewandowski et al., 2009) on the correlation matrices and place half-Cauchy priors on the standard deviation parameters τk and σk (Gelman et al., 2006). Although we use unstructured covariance matrices here, these can be simplified if relevant prior knowledge is available or through a model selection procedure. For example, Σθ and Σq may be reduced to diagonal matrices, which implies that the arms in a study have independent effect sizes.
If both the primary outcome and the post-randomization variable are binary, i.e., Yik ~ Bin(nik, θik = pik), Wik ~ Bin(nik, qik), and hθ(·) and hq(·) are the logit link function, the adjusted relative treatment effect on the primary outcome and post-randomization variable with can be summarized by and respectively. Here, can be interpreted as the adjusted log odds ratio of the treatment’s effect on the primary outcome when g(qik) = 0 and the covariates are equal to 0, which reflects the treatment effect under ideal circumstances. More generally, we can estimate the overall conditional adjusted log odds ratio when the post-randomization variable takes a pre-specified value l and study-level baseline variables take pre-specified values , i.e., . For example, gives a conditional adjusted log odds ratio of the treatment’s effect on the primary outcome when the post-randomization variables in both groups are equal to the underlying population risk and the study-level covariate vector is equal to x. Other interesting quantities are the marginal means of the primary outcome in the two groups given a post-randomization variable value l, i.e., , k = 1, 0, where l is the pre-specified value of the post-randomization variable. This can be interpreted as the population average treatment effect on the primary endpoint in arm k for a given post-randomization value. Summary statistics for the relative treatment effect can be derived from πk(l, x). For example, the overall marginal relative risk can be derived as RRy(l) = π1(l, x)/π0(l, x). Although computation of πk(l, x) involves integration, there is a closed form solution for the probit random effects model, , and a well-established approximation
for the logit random effects model, where . We calculate these statistics in each iteration of the HMC algorithm and inference is based on the means of their posterior samples. As a graphic presentation of the results, we can plot the estimated πk(l, x) over a series of l in the observed range of wik for a given x to show how the primary effect size changes with the underlying post-randomization variable. Extrapolation beyond the available data is not recommended. The posterior samples also give a 95% pointwise credible band for the estimated curve to show the uncertainty in the estimates of πk(l, x).
If Yik is continuous and fy(·) is the normal density, the adjusted relative treatment effect on the primary outcome and post-randomization variable with can be summarized by and respectively. The overall conditional adjusted treatment effect when the post-randomization variable takes a pre-specified value l and study-level baseline variables take pre-specified values is . The interpretations are analogous to those for a binary outcome.
The observed data may contain little information to precisely estimate μ1k when the number of studies in the meta-analysis is small. In such situations, we suggest conducting a sensitivity analysis. Specifically, we can assign a series of informative normal priors with different means for μ1k and investigate how the inference shifts with the prior. Alternatively, we can fix μ1k at a series of values (instead of estimating it from the data) and examine how the estimates change.
The method’s performance was extensively examined in simulation studies, compared to regular meta-regression and the unadjusted meta-analysis. The detailed simulation setup, results and interpretations can be found in Web Appendix A. Based on Web Table 2-4, our proposed method generally provides estimates with almost zero bias and coverage probabilities close to the nominal level. It outperforms existing methods in complete data, MCAR and MAR scenarios and performs similarly in MNAR scenarios. Two important findings of the simulation results are (1) ignoring the association between the post-randomization variable and the primary outcome can lead to an estimate with large bias and low coverage probability and (2) the regular meta-regression method based only on the subset of studies with complete observations can be even more misleading than not adjusting for the post-randomization variables.
3. Case Study
This example was chosen because it shares characteristics with a wide variety of examples that involve other post-randomization variables. One issue in clinical trials is participants discontinuing their assigned treatments. Common reasons for treatment discontinuation are that the treatment is ineffective or has unacceptable side effects (Council et al., 2011). Clearly, treatment discontinuation is a post-randomization variable (intercurrent event) that occurs after treatment initiation and affects the interpretation and indeed measurement of the treatment effect. Meta-analysis of the primary outcome based on the intention-to-treat analysis may be confounded by a difference in treatment discontinuation at the study level, therefore the study-specific effect sizes may be exchangeable only after adjusting for treatment discontinuation. In cases where participants discontinue study treatment because of ineffectiveness, treatment discontinuation may also affect the effect size differentially in different arms. Also, clinical studies often aim to investigate both efficacy and acceptability of the experimental treatment compared to a placebo or active control. At the study level, treatment discontinuation itself is a measure of treatment acceptability or tolerability (Cipriani et al., 2009, 2018; Leucht et al., 2013). This section exemplifies the proposed method through a meta-analysis of randomized trials of the antidepressant paroxetine.
Cipriani et al. (2018) conducted a meta-analysis on 50 double-blind clinical trials comparing the anti-depressant paroxetine with placebo for acute treatment of adults (18 years of age or older and both sexes) with a primary diagnosis of major depressive disorder (MDD). Cipriani et al. (2018) noted that larger all-cause treatment discontinuation rate were associated with a lower response to treatment. We revisited this meta-analysis to exemplify the proposed method. The primary estimand was the treatment effect and acceptability of paroxetine compared to placebo in the population defined by the inclusion criteria specified in Cipriani et al. (2018). Here, efficacy was measured by the proportion of patients who responded to treatment and acceptability was measured by the proportion of patients who discontinued the allocated treatment for any reason. Response was defined as a reduction of at least 50% from baseline on a standardized observer-rating scale for depression. Although last observation carried forward (LOCF) was used to impute missing outcome data in the individual studies, it is not a prerequisite of the proposed method.
Table 1 summarizes the observed response and treatment discontinuation rates from the 50 studies. The response rate was missing in 8 studies and the treatment discontinuation rate was missing in 2 studies. To explore the potential bias due to dependence between the study-specific response rate and treatment discontinuation rate, Figure 1 plots the observed response rates against the observed treatment discontinuation rates for the 40 studies that provided both. This figure shows a clear negative association between the two outcomes, i.e., studies/arms reporting larger treatment discontinuation rates tended to have lower response rates. Therefore, the unadjusted response rates may not be exchangeable and an investigation of the effect of acceptability on efficacy is warranted. We performed meta-regression analyses to (1) compare the treatments for the two outcomes simultaneously, (2) estimate the overall treatment efficacy of paroxetine adjusted for acceptability, and (3) assess the association between the two outcomes.
Table 1.
Descriptive summaries of studies in the anti-depressant drug meta-analysis: Percentiles of response rate and treatment discontinuation rate, and number of studies that did not report these rates.
| Drug | Min. | 2.5% | Median | 97.5% | Max. | N missing | |
|---|---|---|---|---|---|---|---|
| Response Rate | Paroxetine | 0.14 | 0.23 | 0.49 | 0.73 | 0.82 | 8 |
| Placebo | 0.07 | 0.13 | 0.37 | 0.53 | 0.62 | 8 | |
| Treatment discontinuation Rate | Paroxetine | 0.11 | 0.11 | 0.30 | 0.53 | 0.60 | 2 |
| Placebo | 0.09 | 0.11 | 0.30 | 0.69 | 0.82 | 2 |
Figure 1.

Graphical summary of the data. The y-axis is log(odds)of the response, computed as log(yik/(nik – yik)). The dashed line is the simple linear regression of log(odds) on percent treatment discontinuation for placebo while the solid line is the regression line for paroxetine).
We implemented the proposed joint meta-regression method as described in Section 2. Figure 1 shows that the association between the treatment discontinuation and log odds of response rates are similar for the two treatments. Therefore, it is reasonable to assume that treatment discontinuation has non-differential effect between the two arms, i.e, μ11 = μ10 = μ1, giving a more parsimonious model. Logit and identity link functions were used for hθ(·), hq(·) and g(·), repectively. We placed a vague normal prior with mean 0 and variance 2.25 on the fixed effects α0k and μ0k, k = 1, 0, which corresponds to a prior distribution with 95% credible interval (CrI) approximately (0.05, 0.95) for the pik and qik. LKJ(2) is used as the prior for the correlation matrices Ωθ and Ωq. We assigned half Cauchy distributions with scale 0.25 for the priors of σk and τk, which corresponds to a prior distribution with 95% CrI approximately (0.01, 6.4), large enough to cover all plausible values. A vague normal prior with mean 0 and variance 100 was placed on μ1.
The posterior mean of the adjusted log odds ratio when the two arms have the same treatment discontinuation rates (i.e., under μ11 = μ10) was estimated to be 0.50 for treatment efficacy. The 95% CrI excludes zero, indicating that paroxetine is significantly better than the placebo in treating MDD if no patients discontinue treatment. The estimated log odds ratio for treatment acceptability is −0.04, with a 95% CrI including zero. The point estimate of μ1, −2.96, means that one unit decrease in the treatment discontinuation rate is association with an increase of 2.96 in the log odds, a negative association. The 95% CrI excludes 0, which indicates a significant study-level association between treatment discontinuation rates and response rates. These results imply that a lower treatment discontinuation rate is associated with a better observed treatment effect at the study level. Figure 2 shows how the treatment discontinuation rate affects the marginal mean of the response rate in the two treatment groups. In particular, the estimated marginal mean of the response rate in the paroxetine and placebo group are 0.50 and 0.38, respectively at the observed median treatment discontinuation rate 0.30. The posterior mean of was estimated to be 0.54 using the naive unadjusted analysis, with 95% CrI (0.42, 0.66). Comparing the results from the two methods, the estimate of is slightly smaller in the joint meta-regression method. Although the two methods give similar estimates for the relative effects, the absolute effect estimates of the response rates are quite different. Table 2 gives more detailed results on the estimates. It is important to recognize that even after adjusting for treatment discontinuation rate, response rates may still be confounded by other study-level or patient-level characteristics that are unknown or unavailable.
Figure 2.
Case study results: estimated marginal means of the response rate (“E(Response Rate)”) in the two groups given the treatment discontinuation rate.
Table 2.
Case study results from the proposed method adjusting for treatment discontinuation rates and the naive method without adjustment: Posterior mean (95% credible interval).
| Adjusted analysis | Unadjusted analysis | |
|---|---|---|
| 0.50 (0.39, 0.62) | 0.54 (0.42, 0.66) | |
| −0.04 (−0.18, 0.10) | −0.05 (−0.19, 0.09) | |
| μ 01 | 0.91 (0.60, 1.21) | −0.05 (−0.23, 0.12) |
| μ 00 | 0.40 (0.10, 0.71) | −0.59 (−0.76, −0.43) |
| α 01 | −0.81 (−0.99, −0.63) | −0.81 (−0.98, −0.63) |
| α 00 | −0.76 (−0.97, −0.56) | −0.76 (−0.96, −0.54) |
| μ 1 | −2.96 (−3.85, −2.08) | − |
We conducted sensitivity analyses to assess the impact on the estimates of the prior distributions and MAR assumption. Detailed descriptions and results can be found in Web Appendix B. To understand the effect of the post-randomization variable on the primary outcome in each arm, one can start with a model assuming differential effects of the post-randomization variable in each arm. Posterior samples of μ10 and μ11 can then be used to formally test whether μ10 = μ11.
4. Discussion and Future Work
Existing meta-regression methods are insufficient to model post-randomization variables that can differ substantially between treatment arms. This paper proposed a joint meta-regression approach to fill this gap. The proposed method jointly models the primary outcome and the post-randomization variables in each study arm so that both within- and between-study variation in both the primary outcome and the post-randomization variables are taken into account. It also allows either the primary outcome or the post-randomization variable to be missing at random. A Bayesian hierarchical approach naturally accounts for uncertainty arising from the missing data. The proposed model is readily applicable to a wide range of post-randomization variables, such as compliance measured through average pill counts and other biochemical markers, the proportion of missing data during patient follow-up, cross-over to an alternative therapy, or premature withdrawal from study treatment. The method can also serve as a mediation analysis in cases where the post-randomization variables may confound estimation of primary treatment effect but may not themselves be of direct clinical interest (e.g., intercurrent events).
Although we focused on meta-analysis of randomized trials, the proposed method is also relevant to meta-analysis of observational studies. Unlike clinical trials, observational studies do not have the benefit of randomization. Common reasons for being unable to do a meta-analysis on observational studies are that: (1) not all studies report the same set of confounding variables; (2) many confounding variables are not real study-level covariates and are subject to measurement error; and (3) confounding variables may modify the treatment effect both within and between studies. It is easy to see that confounding variables in a meta-analysis of observational studies share similarities with post-randomization variables in meta-analysis of randomized controlled trials. Therefore, this paper’s method can also be used to adjust for confounding variables in meta-analysis of observational studies. For the same reasons, the proposed method may also be applicable to meta-analysis in genome-wide association studies. We leave this to future work.
The proposed meta-regression method models absolute effects in the treatment and control arms. Developments in NMA have stimulated a debate about modeling relative treatment effects versus modeling the absolute effects in each treatment arm of each study (Hong et al., 2016; Dias and Ades, 2016). Recently, modeling the absolute effects has been found to be mathematically tidier with performance similar to methods modeling relative treatment effects, and both models are suitable for the analysis of NMA data with certain requirements on the data (Senn, 2010; White et al., 2019). Although that debate pertains mainly to NMA and this paper focuses on meta-regression evaluating only two treatments with outcomes missing at study level (not at arm level), readers may still wonder why this paper modeled absolute effects instead of modeling the relative effects. In a meta-regression adjusting for post-randomization variables, one obvious benefit of modeling the absolute effect is that it allows values of post-randomization variables that are imbalanced between arms, which is the key attribute that differentiates post-randomization variables from baseline variables. An analogous method that models relative effects may be possible if post-randomization variables have non-differential effects on the treatment and control arms; this warrants future investigation.
The proposed method has several limitations. Meta-regression has been widely used to determine whether the treatment effect depends on the control group risk. Control group risk is an effective measure of baseline differences between studies. It is possible that differences between studies in a post-randomization variable are also associated with underlying risk. Although the proposed method can be extended to include some study-level baseline covariates, we did not investigate the relationships among the primary outcome, control group risk and post-randomization variables. When considering the underlying risk in the control group, we need to account for regression dilution bias (Thompson et al., 1997), i.e., the artificial negative association arising from having correlated measurement error in the observed control group risk and the observed treatment effect. Several functional and structural approaches have been proposed to extend traditional meta-analysis methods to account for underlying risk (Sharp et al., 1996; Thompson et al., 1997; Sharp and Thompson, 2000). Adjusting for both post-randomization variables and underlying risk can be complicated. The proposed methods may be extended to adjust for underlying risk in line with the methods presented in Thompson et al. (1997) and Van Houwelingen et al. (2002). Applying them to the joint meta-regression method requires further investigation.
The results from the proposed method should be interpreted with caution. Conclusions derived from a meta-regression analysis are observational, even when all included studies are randomized trials. The proposed method — and all other existing meta-regression methods — cannot establish causality although they are helpful in generating new hypotheses and motivating new clinical trials. Essentially, meta-regression is an epidemiologic investigation of the relationship between the treatment effect size and the covariate, so its results are subject to the same types of bias as observational epidemiologic analyses (Thompson and Higgins, 2002; Schmid et al., 2004). Also, aggregate (ecologic) data are known to be insufficient to characterize individual-level association. Meta-regression provides ecological inference, and relationships observed for groups do not necessarily hold for individuals. Therefore, results from using meta-regression methods (including the proposed method) without individual patient data should be interpreted with caution to avoid the ecological fallacy (Greenland and Morgenstern, 1989; Berlin et al., 2002).
Supplementary Material
Acknowledgement
This research was supported in part by NIH funding: T32HL129956, 1R01AI116794, 1R01AI130460, 1R01LM012607, P50MH113840, U01DK106786, R01LM009012, R21LM012744, R01LM012982, UL1TR002494. The authors greatly appreciate the thoughtful comments and suggestions by the Co-Editor, the Associate Editor, and the anonymous Referee.
Footnotes
Supporting Information
Web Appendices, the simulation studies, and Tables, Figures and Code referenced in Sections 2.3 and 3 are available with this paper at the Biometrics website on Wiley Online Library.
Data Availability Statement
The data that support the findings in this paper are available from the corresponding author upon reasonable request.
References
- Berkey CS, Hoaglin DC, Mosteller F, and Colditz GA (1995). A random-effects regression model for meta-analysis. Statistics in medicine 14, 395–411. [DOI] [PubMed] [Google Scholar]
- Berlin JA, Santanna J, Schmid CH, Szczech LA, and Feldman HI (2002). Individual patient-versus group-level data meta-regressions for the investigation of treatment effect modifiers: ecological bias rears its ugly head. Statistics in medicine 21, 371–387. [DOI] [PubMed] [Google Scholar]
- Cipriani A, Furukawa TA, Salanti G, Chaimani A, Atkinson LZ, Ogawa Y, Leucht S, Ruhe HG, Turner EH, Higgins JP, et al. (2018). Comparative efficacy and acceptability of 21 antidepressant drugs for the acute treatment of adults with major depressive disorder: a systematic review and network meta-analysis. Focus 16, 420–429. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Cipriani A, Furukawa TA, Salanti G, Geddes JR, Higgins JP, Churchill R, Watanabe N, Nakagawa A, Omori IM, McGuire H, et al. (2009). Comparative efficacy and acceptability of 12 new-generation antidepressants: a multiple-treatments meta-analysis. The lancet 373, 746–758. [DOI] [PubMed] [Google Scholar]
- Council NR et al. (2011). The prevention and treatment of missing data in clinical trials. National Academies Press. [PubMed] [Google Scholar]
- DerSimonian R and Laird N (1986). Meta-analysis in clinical trials. Controlled clinical trials 7, 177–188. [DOI] [PubMed] [Google Scholar]
- Dias S and Ades AE (2016). Absolute or relative effects? arm-based synthesis of trial data. Research synthesis methods 7, 23–28. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Gelman A et al. (2006). Prior distributions for variance parameters in hierarchical models (comment on article by browne and draper). Bayesian analysis 1, 515–534. [Google Scholar]
- Greenland S (1987). Quantitative methods in the review of epidemiologic literature. Epidemiologic reviews 9, 1–30. [DOI] [PubMed] [Google Scholar]
- Greenland S and Morgenstern H (1989). Ecological bias, confounding, and effect modification. International journal of epidemiology 18, 269–274. [DOI] [PubMed] [Google Scholar]
- Hong H, Chu H, Zhang J, and Carlin BP (2016). A bayesian missing data framework for generalized multiple outcome mixed treatment comparisons. Research synthesis methods 7, 6–22. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Lambert PC, Sutton AJ, Abrams KR, and Jones DR (2002). A comparison of summary patient-level covariates in meta-regression with individual patient data meta-analysis. Journal of clinical epidemiology 55, 86–94. [DOI] [PubMed] [Google Scholar]
- Leucht S, Cipriani A, Spineli L, Mavridis D, Örey D, Richter F, Samara M, Barbui C, Engel RR, Geddes JR, et al. (2013). Comparative efficacy and tolerability of 15 antipsychotic drugs in schizophrenia: a multiple-treatments meta-analysis. The Lancet 382, 951–962. [DOI] [PubMed] [Google Scholar]
- Lewandowski D, Kurowicka D, and Joe H (2009). Generating random correlation matrices based on vines and extended onion method. Journal of multivariate analysis 100, 1989–2001. [Google Scholar]
- Little RJA and Rubin DB (2002). Statistical Analysis with Missing Data, 2nd Edition. John Wiley & Sons, New Jersey. [Google Scholar]
- Schmid CH, Stark PC, Berlin JA, Landais P, and Lau J (2004). Meta-regression detected associations between heterogeneous treatment effects and study-level, but not patient-level, factors. Journal of clinical epidemiology 57, 683–697. [DOI] [PubMed] [Google Scholar]
- Senn S (2010). Hans van houwelingen and the art of summing up. Biometrical Journal 52, 85–94. [DOI] [PubMed] [Google Scholar]
- Sharp SJ and Thompson SG (2000). Analysing the relationship between treatment effect and underlying risk in meta-analysis: comparison and development of approaches. Statistics in Medicine 19, 3251–3274. [DOI] [PubMed] [Google Scholar]
- Sharp SJ, Thompson SG, and Altman DG (1996). The relation between treatment benefit and underlying risk in meta-analysis. BMJ: British Medical Journal 313, 735. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Stan Development Team (2016). RStan: the R interface to Stan. R package version 2.14.1. [Google Scholar]
- Thompson SG (1993). Controversies in meta-analysis: the case of the trials of serum cholesterol reduction. Statistical methods in medical research 2, 173–192. [DOI] [PubMed] [Google Scholar]
- Thompson SG and Higgins J (2002). How should meta-regression analyses be undertaken and interpreted? Statistics in medicine 21, 1559–1573. [DOI] [PubMed] [Google Scholar]
- Thompson SG and Sharp SJ (1999). Explaining heterogeneity in meta-analysis: a comparison of methods. Statistics in medicine 18, 2693–2708. [DOI] [PubMed] [Google Scholar]
- Thompson SG, Smith TC, and Sharp SJ (1997). Investigating underlying risk as a source of heterogeneity in meta-analysis. Statistics in medicine 16, 2741–2758. [DOI] [PubMed] [Google Scholar]
- Van Houwelingen HC, Arends LR, and Stijnen T (2002). Advanced methods in meta-analysis: multivariate approach and meta-regression. Statistics in medicine 21, 589–624. [DOI] [PubMed] [Google Scholar]
- White IR, Turner RM, Karahalios A, and Salanti G (2019). A comparison of arm-based and contrast-based models for network meta-analysis. Statistics in medicine 38, 5197–5213. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Yuan Y and Little RJ (2009). Meta-analysis of studies with missing data. Biometrics 65, 487–496. [DOI] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Supplementary Materials
Data Availability Statement
The data that support the findings in this paper are available from the corresponding author upon reasonable request.

