Skip to main content
Wiley Open Access Collection logoLink to Wiley Open Access Collection
. 2025 Feb 7;67(1):e70034. doi: 10.1002/bimj.70034

A Bias‐Corrected Bayesian Nonparametric Model for Combining Studies With Varying Quality in Meta‐Analysis

Pablo Emilio Verde 1,, Gary L Rosner 2
PMCID: PMC11803498  PMID: 39917836

ABSTRACT

Bayesian nonparametric (BNP) approaches for meta‐analysis have been developed to relax distributional assumptions and handle the heterogeneity of random effects distributions. These models account for possible clustering and multimodality of the random effects distribution. However, when we combine studies of varying quality, the resulting posterior is not only a combination of the results of interest but also factors threatening the integrity of the studies' results. We refer to these factors as the studies' internal validity biases (e.g., reporting bias, data quality, and patient selection bias). In this paper, we introduce a new meta‐analysis model called the bias‐corrected Bayesian nonparametric (BC‐BNP) model, which aims to automatically correct for internal validity bias in meta‐analysis by only using the reported effects and their standard errors. The BC‐BNP model is based on a mixture of a parametric random effects distribution, which represents the model of interest, and a BNP model for the bias component. This model relaxes the parametric assumptions of the bias distribution of the model introduced by Verde. Using simulated data sets, we evaluate the BC‐BNP model and illustrate its applications with two real case studies. Our results show several potential advantages of the BC‐BNP model: (1) It can detect bias when present while producing results similar to a simple normal–normal random effects model when bias is absent. (2) Relaxing the parametric assumptions of the bias component does not affect the model of interest and yields consistent results with the model of Verde. (3) In some applications, a BNP model of bias offers a better understanding of the studies' biases by clustering studies with similar biases. We implemented the BC‐BNP model in the R package jarbes, facilitating its practical application.

Keywords: Bayesian nonparametric, bias correction, comparative effectiveness methods, conflict of evidence, cross‐evidence synthesis, hierarchical models, meta‐analysis

1. Introduction

Meta‐analysis methods help researchers answer questions that require the combination of statistical results across several studies. Very often the only available studies are of different types and with varying quality. Therefore, when we combine disparate evidence at face value we are not only combining results of interest but also biases that might threaten the quality of the results. As a consequence, the resulting meta‐analysis could be misleading.

Factors threatening the integrity and quality of studies are called internal validity biases. For example, patient selection bias, dilution bias, reporting bias, data quality, and so forth are all forms of internal validity bias. The problem of these types of biases is that they are not directly observable, making their correction in meta‐analysis a challenging problem.

Table 1 presents an example of a meta‐analysis combining different types of studies, where researchers collected evidence across 18 observational studies (OS) on COVID‐19‐infected patients (de Almeida‐Pititto et al. 2020). This meta‐analysis was performed during the COVID‐19 pandemic when researchers were urgently trying to assess evidence linking baseline risk factors, such as in this case, hypertension, with mechanical respiratory assistance within 28 days. We will present more details about this meta‐analysis in Section 4, but by looking at the studies' summaries and the different design types, we have what we call the “Anna Karenina” principle in evidence synthesis: good quality studies are alike, but biased ones are biased in their own way. However, these studies were the only available evidence at that time, and researchers faced an imperfect body of evidence to answer their question.

TABLE 1.

Example of a meta‐analysis combining disparate studies, where researchers collected evidence across 18 observational studies on COVID‐19‐infected patients. The odds ratio (OR) measures the association between hypertension and mechanical respiratory assistance within 28 days after hospitalization. Values of OR greater than one indicate higher risk of hypertension.

Author Design OR seOR
N
Guo et al. 2020 Case series 6.48 1.43 187
Li J et al. 2020 Case series 2.59 1.14 1178
Mao et al. 2020 Case series 3.17 1.39 214
Wang Z et al. 2020 Case series 6.63 2.07 69
Zhang JJ et al. 2020 Case series 1.88 1.45 140
Li X et al. 2020 Cross‐sectional 2.20 1.21 548
Wan S et al. 2020 Cross‐sectional 1.12 1.83 135
Xiang et al. 2020 Cross‐sectional 12.60 2.49 49
Chen et al. 2020 Retrospective cohort 3.56 1.57 150
Deng et al. 2020 Retrospective cohort 1.51 1.52 112
Feng et al. 2020 Retrospective cohort 5.25 1.25 476
Guan W et al. 2020 Retrospective cohort 2.02 1.22 1099
Huang et al. 2020 Retrospective cohort 1.18 2.38 41
Liu W et al. 2020 Retrospective cohort 2.49 2.27 78
Simone et al. 2020 Retrospective cohort 2.85 1.50 124
Wang D et al. 2020 Retrospective cohort 4.96 1.51 138
Wu C et al. 2020 Retrospective cohort 2.35 1.43 201
Zhang G et al. 2020 Retrospective cohort 4.37 1.40 221

Note: Source of the data de Almeida‐Pititto et al. 2020.

In this meta‐analysis, researchers estimated an alarming odds ratio (OR) of 2.96 with a 95% confidence interval of [2.33, 3.76], meaning that patients admitted to hospitals with hypertension were almost three times more likely to receive mechanical respiratory assistance than patients without hypertension. The solid line in Figure 1 corresponds to the posterior pooled OR by applying a noninformative Bayesian normal random effects meta‐analysis model, which confirms this result. However, the disparity of the evidence in Table 1 comes as a warning against these results.

FIGURE 1.

FIGURE 1

Results of the meta‐analysis comparing a random effects (RE) model (solid line) and the bias‐corrected Bayesian nonparametric (BC‐BNP) model (dashed line). Posterior distributions of the pooled odds ratio (OR) of needed for mechanical assistance for patients with hypertension. From left to right, the vertical dashed lines correspond to having no effect, OR = 1 (black), and the pooled estimate of the normal random effects, OR = 2.96 (blue). The short vertical lines at the bottom correspond to the estimated ORs from the studies.

In this paper, we present a new meta‐analysis model, called the BC‐BNP model, where BC‐BNP stands for bias‐corrected Bayesian nonparametric. The aim, or at least the hope, is that the model automatically corrects meta‐analysis results affected by internal validity bias by only using the reported effects and their standard errors. This model relaxes the parametric assumptions of the bias distribution of the model introduced by Verde (2021). The general idea behind these models is that it is possible, at least in theory, to decompose between‐study heterogeneity into two components: a component of diversity and a component of internal validity bias (Higgins, Thompson, and Spiegelhalter 2009). We can schematically present this idea as follows:

HeterogeneityDiversity+Bias.

The dashed curve presented in Figure 1 corresponds to the posterior distribution of the pooled OR for the Table 1 studies calculated by using the BC‐BNP model. The posterior mean is 1.98 with a 95% posterior interval of [0.52, 3.35]. The effect of hypertension remains important, but it is moderated after bias correction and is subject to greater uncertainty. This result reduces certainty about the evidence collected in Table 1 and to the results of a meta‐analysis without bias correction.

Bayesian nonparametric (BNP) inference represents a broad and highly‐active research area in Bayesian hierarchical modeling and computational statistics. A gentle introduction of BNP is given by Rosner, Laud, and Johnson (2021) and Gelman et al. (2013). These models are “nonparametric” in the sense that they assign a stochastic process (e.g., Dirichlet process [DP], Pólya tree [PT] process), with a theoretically infinite number of parameters, to the statistical model. In this way, BNP avoids the more restrictive assumption of “parametric models,” where such a model assumes that the data can be described by a few finite numbers of parameters. From a pragmatic point of view, BNP modeling allows to acknowledge uncertainty about an assumed model, where a base distribution is assumed and realizations of the process allow variation around this distribution.

In meta‐analysis, BNP models have been developed to increase the flexibility of the random effects models. Ohlssen, Sharples, and Spiegelhalter (2007) provided an excellent practical implementation of the DP and Dirichlet process mixtures (DPMs) in meta‐analysis. The authors present the stick‐and‐break representation of the DP and DPM and their implementation in BUGS (Spiegelhalter et al. 2007). They apply these methods to a meta‐analysis of OS reporting hospitals' performances based on mortality rates (death in hospital within 30 days of emergency admission for myocardial infarction) from the United Kingdom.

Burr and Doss (2005) introduced a BNP model for random effects meta‐analysis based on DPMs. The motivation of the authors was based on a meta‐analysis of case‐control studies that investigated if there was an association between the presence of a certain genetic trait and an increased risk of coronary heart disease. The available evidence was contradictory, dispersed and motivated the development of a flexible random effects distribution. Although the DPM adds flexibility to the random effects, this happens at the cost of eliminating the interpretation of a location parameter as a pooled effect across studies. As a remedy, Burr and Doss (2005) considered a conditional DPM (CDPM) model in which the random effects distribution is constrained to have the median as the pooled effect size. The R package bspmma implements the Burr and Doss model (Burr 2012).

Branscum and Hanson (2008) developed a meta‐analysis random effects model based on mixture of Pólya Trees (MPTs), which is a family of random partitions of the sample space. PTs lend themselves naturally to fixing percentiles. For example, the first partition can be located at the median of the random effects distribution. In this way, an MPT approach allows for the desired flexibility in the random effects distribution while retaining the simplicity of the normal–normal meta‐analysis model in terms of evaluating a single location parameter. Using simulation experiments with relatively large number of studies (N=100), the authors showed that the MPT approach works well when normality holds as well as when it does not because the approach anticipates the possibility of misspecification of normality. To illustrate the MPT approach, the authors presented a meta‐analysis of prospective and retrospective OS investigating the relationship between alcohol consumption and breast cancer.

Dunson, Xue, and Carin (2008) presented the matrix stick‐breaking process (MSBP) for flexible meta‐analysis. This model is suitable in multiparameters meta‐analysis when each study reports multiple regression coefficients, but not necessarily the same ones in each study.

Karabatsos, Talbott, and Walker (2015) presented a BNP model for meta‐analysis regression, which is a special case of the adaptive‐modal BNP regression (Karabatsos and Walker 2012). This model increases the flexibility of the random effects distribution and allows nonlinear relationships in the metaregression component.

Poli et al. (2023) recently developed a multivariate PT model for meta‐analysis of S studies with time‐to‐event endpoints across n possible cohorts (i.e., n<S). In this model, the independent beta priors for the splitting probabilities in the PT construction are replaced by a Gaussian process prior. The Gaussian process links study‐specific covariates (e.g., tumor type, treatment agent, and biomarker status) to the PT partition in the logistic scale. In this way, the random partition introduces correlation between any pair of studies with similar baseline characteristics. This method is motivated by an extensive meta‐analysis of S=174 phase I/II studies on assessing the effect of biomarkers on clinical outcomes in patients with solid tumors. The authors evaluate the model using a simulation study mimicking the empirical characteristics of the motivated meta‐analysis.

These BNP meta‐analysis models have increased the flexibility of random effects distributions allowing nonlinear relationships in metaregression and complex dependencies in multivariate meta‐analysis. Moreover, they have been successfully applied to combine studies of different designs. However, these models have not been specifically tailored to adjust for multiple biases in meta‐analysis.

In Verde and Ohmann (2015), we reviewed over two decades of methods and applications of combining disparate evidence and adjusting for bias in meta‐analysis. We classified statistical approaches in four main groups: The confidence profile method (CPM, Eddy, Hasselblad, and Shachter 1992), cross‐design synthesis (CDS, Droitcour, Silberman, and Chelimsky 1993), direct likelihood bias modeling, and Bayesian hierarchical modeling. Within the Bayesian parametric hierarchical models, there are approaches to combine aggregated evidence with potential risk of bias, for example, the work of Prevost, Abrams, and Jones (2000), Welton et al. (2009), and Dias et al. (2010). These approaches differ in the way that the bias component is estimated and incorporated into the meta‐analysis model.

We organize this paper as following: In Section 2, we present methodological details of the BC‐BNP model. Here, we describe a flexible model that allows inference of the population effect in a meta‐analysis subject to studies having different degrees of internal validity bias. In Section 3, we perform a simulation study to evaluate the BC‐BNP under different scenarios, and compare it with three alternative meta‐analysis models. In Section 4, we show the BC‐BNP model in action by applying this methodology to two meta‐analyses; one meta‐analysis combines studies with different designs, and the other RCTs with varying quality. We conclude the paper in Section 5 by highlighting advantages, limitations, and pointing out possible extensions of the BC‐BNP model.

2. BC‐BNP Meta‐Analysis Model

In this section, we present methodological details of the BC‐BNP meta‐analysis model. This model is based upon a mixture of a parametric random effects distribution, which represents the model of interest, and a BNP model for the bias component. The resulting random effects model is a combination of a model of interest and a process of bias.

The BC‐BNP model has potentially two advantages. First, the parameters of the model of interest have a simple interpretation. Second, the bias component does not have the parametric restrictions presented in Verde (2021). In this way, we expect that it is possible to isolate the model of interest and make fewer model assumptions on the bias component. In addition, the BNP model adds flexibility to the identification of biased studies by clustering their results.

2.1. Modeling Biased Evidence

Suppose that a meta‐analysis of N studies has reported effect estimates y1,y2,,yN with their corresponding standard errors SE1,SE2,,SEN. In this situation, we assume that the sample size within studies, say ni, is large enough to ignore uncertainty on the estimated standard errors SEi.

The crucial step in modeling biased evidence is to recognize that the reported effect estimate yi is a potentially biased version of the effect that could be observed in an “ideal study.” Therefore, the likelihood contribution of yi corresponds to a biased study effect θiB that we model with a normal likelihood:

yi|θiBN(θiB,SEi2),i=1,,N. (1)

2.2. Modeling Heterogeneity as Diversity and Bias

Following Verde (2021), we represent the biased study effect θiB as the combination of “three hidden variables”: θi, βi, and Ii, where

θiB=(1Ii)θi+Ii(θi+βi), (2)

which is equivalent to θiB=θi+Iiβi.

The random effect θi represents the bias‐corrected study effect, βi models the amount of the internal validity bias of the study, and the indicator variable Ii labels if study i is biased or not, that is,

Ii=1,if studyiis biased,0,otherwise,

where Pr(Ii=1) is the probability that the study i is biased. We assume a priori that each study has the same probability to be biased and we denote the class probability for the biased studies by πB. This parameter expresses the uncertainty about the proportion of biased studies in the meta‐analysis.

We model the random effect θi as

θiNμθ,τθ2, (3)

where the parameters μθ and τθ2 represent the mean effect and the variability due to “diversity” across bias‐corrected studies, respectively. The random effect model (3) is “the model of interest” that we aim to isolate in the meta‐analysis.

In Verde (2021), we used a parametric model for the bias effect βi. The advantage of a parametric model is the simple interpretation of the parameters. However, it is not immediately clear if a bias is discrete or continuous, or which probability distribution could be appropriate for βi. In this paper, we relax the distribution assumption of βi by using a BNP prior.

More generally, let Gβ be the probability distribution of βi. We propose to handle uncertainty around the bias effect by giving a BNP prior distribution to Gβ. In this approach, we assume that Gβ is random, with a prior on the space of probability distributions. Specifically, we model the BNP prior of Gβ with a DP, having a base distribution N(μβ,τβ2), and concentration parameter α:

βi|GβGβ,Gβ|μβ,τβ2,αDPNμβ,τβ2,α. (4)

The DP prior in (4) is a discrete and infinite‐dimensional parameter prior, where μβ and τβ2 represent the location and scale of the base distribution. The prior expectation of the process corresponds to the base distribution

EGβ|μβ,τβ2=Nμβ,τβ2 (5)

and the prior variance depends on the base distribution and α:

varGβ|μβ,τβ2=Nμβ,τβ21Nμβ,τβ2α+1. (6)

The variance of the process (6) shows that the concentration parameter α influences the proximity of the process to the base distribution. In addition, α determines the number of clusters in the process, which is elaborated later in Equation (9).

The DP prior in (4) allows to represent the random distribution function Gβ with the “stick‐breaking” construction (Sethuraman 1994). In this case, Gβ can be expressed as

Gβ=k=1wkδβk, (7)

where the stochastic weights w1,w2, and the support points β1,β2, are a priori two independent sequences of random variables, with wk>0 and k=1wk=1, and δβk is the Dirac delta function, which places a measure 1 on the location βk. In the representation (7), the collection of β1,β2 and w1,w2 constitutes a realization of Gβ.

The support points β1,β2, are independently sampled from the base distribution:

βkN(μβ,τβ2). (8)

In the presence of a positive bias, where some studies exaggerate results, we expect that μβ>0 and the ordinate values β1,β2, will be sampled to the right from the model of interest (3).

The random weights wk are sampled with decreasing expectations as follows:

wk=qkj<k(1qj)withqkBeta(1,α). (9)

The stick‐breaking weights w1,w2,, determine the amount of clustering of the process, which is driven by the concentration parameter α.

A simple way to understand the influence of the parameter α in the number of clusters is by observing that w1Beta(1,α), then

E(w1)=11+α.

In the limit when α0, the first weight approaches w11, and the random distribution Gβ has a single support point. In this situation, Gβ is quite different from the base distribution N(μβ,τβ2). When α increases, no finite collection of weights dominates and each random draw of Gβ becomes arbitrarily close to N(μβ,τβ2), that is, the process becomes concentrated at the base distribution. In Section 2.3, we will see that the value of α can be bounded to a finite value that reflects the sample size of the meta‐analysis and the prior expected number of biased studies.

We can summarize the model construction of θiB as follows:

θiBN(μθ,τθ2),with probability(1πB),N(μθ,τθ2)+k=1wkδβk,with probabilityπB. (10)

In (10), the biased study effect θiB follows a convex combination of two stochastic processes. With probability (1πB) we obtain a constant process, where all sequences of the process are the same, and they correspond to model of interest (3). With probability πB, we obtain a random process where each sequence of the process is different and each βi is draw from a common distribution that we model with a DP.

This model construction aims to separate diversity and bias from the total heterogeneity in the meta‐analysis. Although the model is a BNP model for random effects, it has the advantage that the parameters μθ and τθ have a familiar interpretation.

2.3. Specification of Priors of Hyperparameters

Prior distributions are a fundamental part of the model, in particular for meta‐analysis models intending to correct for internal validity bias. The BC‐BNP model has two sets of hyperparameters μθ, τθ that corresponds to the model of interest (3), and μβ, τβ, πB, and α that drive the process modeling the bias (4). We assume independent hyperpriors on these parameters, and we provide a default setting for the simulation experiment of Section 3 and the applications of Section 4. Although, this default setting is specific for these applications, it can be used as guideline for similar meta‐analysis. In addition, we give a simple approximation to the random probability distribution Gβ to make it more computationally tractable.

2.3.1. Priors for the Location Parameters μθ and μβ

For location parameter μθ, we use μθN(0,σμ2), and we set the default value σμ=106 to represent a local flat prior for μθ.

For the bias hyperparameter μβ, we use a uniform distribution: μβUniform(Blower,Bupper). In Section 3, and in Section 4, we set the default values Blower=15 and Bupper=15, which correspond to an unknown bias in any direction. In addition, in Section 4 we perform a sensitivity analysis by restricting the prior of μβ to positive bias, that is, we set μβUniform(0,15).

2.3.2. Prior for the Standard Deviation Parameters τθ and τβ

We parametrize the prior of τθ in terms of the precision parameter 1/τθ2, and we use a scale gamma with scale parameter S and with degrees of freedom df as a prior: 1/τθ2Scale.Gamma(S,df) (Huang and Wand 2013). We set the default values S=0.5 and df=1.

The scale gamma distribution on 1/τθ2 implicitly defines a prior on the standard deviation τθSt+(df), where t+(df) represents the t‐distribution with df degrees of freedom restricted to positive values. In the limit as the degrees of freedom df increases to infinity, the distribution then becomes a half‐normal with standard deviation S. Gelman (2006) called this prior a “weakly informative prior,” and Röver et al. (2021) discussed “weakly informative priors” in the context of the Bayesian normal–normal meta‐analysis model.

We set as default values S=0.5 and df=1, which implicitly corresponds to using a half‐Cauchy prior for τθ. This prior has P(τθ<0.5)=0.5 and P(τθ>4×0.5)=0.16, which covers plausible values of τθ in the log(OR) and mean difference scale.

In a similar way, we parameterize the prior of τβ in terms of the precision parameter 1/τβ2, and we use a scale gamma with scale parameter S=0.5 and with degrees of freedom df=1 as default prior for 1/τβ2.

2.3.3. Prior for the Probability of Bias Parameter πB

We model the probability of bias πB with a Beta distribution πBBeta(a0,a1), and we take the default values a0=0.5 and a1=1. This prior reflects that we expect about one third of biased studies in the meta‐analysis, that is, E(πB)0.33. However, this prior is asymmetric to the right with a spike at πB0. Therefore, if bias adjustment is unnecessary, posterior values of πB will be concentrated close to zero.

In addition, if we have information about the proportion of low‐quality studies that might be at risk of bias, this information can be incorporated in the analysis by eliciting the values of a0 and a1. Suppose that after evaluating the quality of the N studies, we suspect that πB is around πB and it would be fairly surprising that it could be greater than Qbias. Then, we calculate a0 and a1, such that the median of the distribution is πB and the 90% quantile is Qbias.

2.3.4. Finite Truncation of the DP and the Prior for α

Bayesian models with infinite‐dimensional parameters require special strategies for using Markov chain Monte Carlo (MCMC) techniques to approximate posterior distributions. Two popular approaches in DP modeling are: marginalization and truncation of the process. Marginalization leads to a Pólya urn schema that allows efficient MCMC algorithms. The disadvantage of marginalization is that we cannot make inference about functions of the process, for example, calculate the expected value E(Gβ).

Truncation allows to have a finite parameter approximation for the distribution of the process. A recent overview of truncated Dirichlet process (TDP) is given by Griffin (2016), and further developments using algorithms for adaptive empirical truncation are provided by Arbel, De Blasi, and Prünster (2019) and Zhang and Dassios (2023).

A TDP approximates Gβ with a maximum of K components:

k=1wkδβkk=1Kwkδβk. (11)

In practical terms, this approximation allows clustering the bias effects βi into a maximum of K clusters, where the maximum number of clusters K is empirically adapted on the application context.

A pragmatic way to choose K is by making the expected probability of the final component wK=1k=1K1wk to be small:

E(wK)=ε. (12)

After K1 random breaks, the remaining probability has expectation

E1k=1K1wk=Ek=1K1(1qk)=α1+α(K1). (13)

From (12) and (13), we have

K=1+log(ε)logα1+α, (14)

and by taking ε=0.01 we can approximate K as

K1+5α. (15)

On the other hand, the maximum number of clusters Kmax is related to the prior distribution of α, that we choose to be

αUniform(0.5,αmax). (16)

We take the lower bound of 0.5 to avoid very low values of wk that may produce numerical issues.

Therefore, by using (15), and the upper bound of the prior (16), we can approximate the maximum number of clusters by

Kmax=1+5αmax. (17)

Now, the prior expectation of πB is

E(πB)=a0a0+a1,

which gives as an expected number of biased studies in the meta‐analysis NB:

NB=Na0a0+a1. (18)

In practice, the maximum number of clusters Kmax in the bias component should not exceed the expected number of biased studies NB. Then, we have

Kmax=1+5αmax<Na0a0+a1, (19)

and solving for αmax resulted that this hyperconstant should be

αmax<15(N1)a0a1a0+a1. (20)

We follow this pragmatic approach to specify αmax and Kmax in our applications.

2.4. Implementation, Bayesian Computations, and Reporting Results

The models presented in this paper are implemented in the R package jarbes (Verde 2024). The name of the package stands for Just a rather Bayesian Evidence synthesis. This package implements a number of Bayesian meta‐analysis models within the family of hierarchical metaregression (HMR) models. The main characteristic of HMR models is the explicit modeling of the bias process in CDS and cross‐evidence synthesis, that is, when different study types, for example, RCT, OS, and different data types, aggregated data (AD) with individual participant data (IPD) are combined in a meta‐analysis. The model presented in Section 2 is an example of an HMR, where AD data from studies of different types or varying quality are combined.

The implementation of the functions in jarbes follows the same strategy as the package bamdit (Verde 2018), where each function represents a Bayesian meta‐analysis model. Once the user selects a model's function, the function automatically writes the BUGS language script needed to perform the MCMC computations. The function sends the BUGS script to JAGS (Plummer 2003), and results are sent back to R.

The JAGS software has an internal directed acyclic graph (DAG) representation of the model. The software automatically factorizes the DAG and chooses the sample algorithm for each node, the algorithm decides the sampler according to local conjugate between the parent–child distribution. After this process, a hybrid Gibbs sampler is used for MCMC simulations.

The function bcmixmeta() implements the BC‐BNP model presented in Section 4. The outcome of this function is an R object from the class “bcmixmeta,” which contains the MCMC simulations, the model parameterization, the data used in the calculations, the JAGS script, and so forth. In particular, the JAGS script created during the application of bcmixmeta() could be useful to practitioners interested in expanding the model or using other MCMC software, such as MultiBUGS, Stan, or Nimble.

Another useful functionality is a diagnostic function, which recognizes the “bcmixmeta” object and plots the posterior results of Δ=μβμθ against πB. This diagnostic plot was developed in Verde (2021) to assess if the bias correction was necessary. If the 95% posterior interval of Δ overlaps the horizontal line at E(2τθ|Data) then bias correction is not needed.

We summarize the results of the meta‐analyses by reporting the posterior distributions of the hyperparameters: μθ, τθ, μβ, τβ, and πB. At the level of the studies, we present the posterior probabilities that a study is biased, the posteriors of the bias effect θiB, and the bias‐corrected effect θi. For each posterior distribution, we present the mean, the standard deviation, and the 95% posterior interval.

2.5. Coclustering Studies

One potential advantage of the BC‐BNP model is the ability to identify subgroups of studies that tend to visit the same cluster in the MCMC iterations. This model feature could help clustering similar studies and identify those that disagree with others, and learn about disparate sources of bias across studies.

Let Li be the cluster label of study i, where L=1 indicates the cluster of the unbiased studies and L=2,,K+1 the different clusters in the bias component. In each MCMC iteration t, we monitor the coclustering between studies by calculating an N×N matrix St of the indicator variables Si,jt, where

Si,jt=1ifLi=Lj0ifLiLj. (21)

The variable Si,jt indicates whether or not two studies visit the same cluster at iteration t. Therefore, S^, the average of the matrix of St after T iterations, estimates the coclustering probability between studies.

We define the dissimilarity distance between studies as D^=1S^, and we display the coclustering results using a heatmap plot generated by the R package pheatmap (Kolde 2019). The heatmap is generated by clustering columns and rows of D^ with the default hierarchical clustering algorithm in R (method = “complete”).

3. A Simulation Study

This simulation study aims to determine whether the posterior distributions of the BC‐BNP model can detect bias under simulation conditions where bias is present and can confirm the absence of bias under data simulation condition where bias is not present. The simulation study is designed to have similar characteristics to the meta‐analyses that are presented in Section 4.

We compare the performance of the BC‐BNP model with three alternative meta‐analysis models: (1) The usual normal–normal “Bayesian random effects,” which corresponds to (3) without bias correction, (2) an “Oracle Bayesian random effects,” where biased studies were manually excluded before running the previous model, and (3) the parametric BC model (Verde 2021). We run the four models with the default hyperpriors presented in Section 2.3. Further computational details and results can be found in the R code supplement in the folder “Simulations‐Section‐3” of Supporting Information of the paper.

3.1. Simulation Model and Different Levels of Bias

To simplify the presentation, we work in the scale of the logarithm of the odds ratio (i.e., log(OR)), which would be a typical situation in practice. For a meta‐analysis of size N, we simulate the studies' results using a normal–normal random effects model:

yi|θi,SEiN(θi,SEi),andθi|μθ,τθN(μθ,τθ2)(i=1,2,N), (22)

where yi is the simulated log(OR), and θi the effect of study i. The parameters μθ and τθ represent the pooled log(OR) and the between‐studies standard deviation, respectively.

In model (22), SEi is the standard error of yi, which is equivalent to σ2/ni. The parameter σ is the within‐study standard deviation, and ni is the sample size of study i. Thus, by marginalization over the study effect θi, the model (22) is equivalent to

yiNμθ,τθ2+σ2/ni. (23)

For each simulated scenario, the sample sizes ni are generated at random from those of the real studies presented in Section 4. For a meta‐analysis of size N=20, we sample ni from the data in Section 4.2, and for N=30 we sample ni from the data in Section 4.1.

For all scenarios, we set the pooled log(OR) at μ=1, and τ=0.5, representing “a fairly high” heterogeneity in the log(OR) scale (Spiegelhalter, Abrams, and Myles 2004, 170). The standard error of the log(OR) derived from a 2×2 table is approximately given by SEi=σ/ni. We set the value of σ=4.95, which was empirically determined by calculating the median of σ from the studies, analyzed in Section 4.

Given a sample of simulated data y1,,yN, we generate B biased studies by taking a random sample without replacement of B studies out of N. The selected studies are shifted by a mean bias μβ, that is, yb+μβ (b=1,,B). We add uncertainty of the amount of bias by considering that a study could have a mild, large, or extreme bias. We introduce this feature by considering three possible values of the mean bias: μβM, μβL, and μβE, where

μβM=Q3+1.50IQR,μβL=Q3+2.25IQR,andμβE=Q3+3.0IQR. (24)

The quantities Q3 and IQR correspond to the third quartile and the inter‐quartile range of (23), respectively. For a particular simulation scenario, Q3 and IQR are calculated by taking the marginal variance of y at var(y)=τ2+σ2/n¯, where n¯=1/Ni=1Nni.

3.2. Scenarios

Table 2 presents the configuration of the simulated scenarios. We consider eight possible scenarios corresponding to two meta‐analysis sample sizes, N=20 and N=30, and four percentages of biased studies, 0%, 10%, 30%, and 50%. For each scenario, we include studies with mild, large, and extreme bias. For example, in the scenario with N=20 and 50% biased studies, we assigned three studies with mild bias, three studies with large bias, and four studies with extreme bias.

TABLE 2.

Configuration of the simulated scenarios: We considered two total sample sizes for the meta‐analysis, N=20, and N=30, and four percentages of the number of biased studies, 0%, 10%, 30%, and 50%. For each percentage, we consider different distributions of biased studies: mild, large, and extreme bias.

N
% Bias Biased studies Mild Large Extreme
20 0% 0 0 0 0
10% 2 0 1 1
30% 6 2 2 2
50% 10 3 3 4
30 0% 0 0 0 0
10% 3 1 1 1
30% 9 3 3 3
50% 15 5 5 5

For all models, we compare the resulting posterior distributions of μθ and τθ. In addition, for the BC‐BNP model, we present the forest plot comparing the biased θB and bias‐corrected effects θ, and the heatmap plot for coclustering.

3.3. Simulation Results

The results for N=20 and N=30 were fairly similar. Therefore, in this section, we present the results for N=20. The results for N=30 are summarized in the supplementary material of this section.

Each panel of Figure 2 shows the posterior distributions of μθ for the meta‐analysis models, and the different amounts of biased studies. In the nonbiased scenario, the Bayesian random effects model shows a posterior of μθ centered at the true value of μθ=1. However, this model is sensitive to the biased studies, where the inclusion of two biased studies shifted the posterior to the right. As expected, the Oracle model performs correctly in all scenarios.

FIGURE 2.

FIGURE 2

Simulation results for N=20: Comparison of four meta‐analysis models and the effect on the posterior distribution of μθ by increasing the number of biased studies B=0,2,6,10. The models are: Bayesian random effects, Oracle Bayesian random effects (i.e., biased studies excluded), BC parametric, and BC nonparametric. The dashed vertical line corresponds to μθ=1.

The lower panels of Figure 2 present the results of the parametric BC and the BC‐BNP. We see that in the nonbiased scenario, the posteriors of μθ are correctly centered at the true value of μθ=1. In the biased scenarios, the parametric and the BC‐BNP remained robust against biased studies.

The four panels of Figure 3 present the posterior of τθ for each model and different biased scenarios. Similarly to the posterior of μθ, the Bayesian random effects model performed correctly in the nonbiased scenario, and was very sensitive to biased studies. The lower panels of Figure 3 show that the posteriors of the parametric BC and the BC‐BNP performed correctly in the nonbiased and the biased scenarios.

FIGURE 3.

FIGURE 3

Simulation results for N=20: Comparison of four meta‐analysis models and the effect on the posterior distribution of τθ by increasing the number of biased studies B=0,2,6,10. The models are: Bayesian random effects, Oracle Bayesian random effects (i.e., biased studies excluded), BC parametric, and BC nonparametric. The dashed vertical line corresponds to τθ=0.5.

Taking the scenario of N=20 and 50% of biased studies, Figure 4 illustrates the resulting forest plot with the posterior medians and the 95% posterior intervals comparing the biased effect θB, and the bias‐corrected effect θ of each study. The figure displays the amount of bias correction performed by the BC‐BNP model at the study level. For example, study number 1 has P(I1=1|Data)=0.04, and bias correction is not required. In this case, the posterior medians and the 95% intervals of θ1B and θ1 overlapped.

FIGURE 4.

FIGURE 4

Simulation results for N=20: Forest plot comparing for each study the 95% posterior interval of the biased effect θB and the bias‐corrected effect θ. The dashed vertical line corresponds to μθ=1.

Different situations correspond to studies numbered 16, 19, and 20, which are generated with large, moderate, and extreme biases, respectively. These studies have the following posteriors probability of been biased: P(I16=1|Data)=0.92, P(I19=1|Data)=0.90, and P(I20=1|Data)=1.00. For these studies, Figure 4 illustrates an adaptive bias correction toward the true value μθ=1. In particular, for the extremely biased study number 20, the posteriors of θ20B and θ20 do not overlap.

For the same scenario, N=20 and 50% of biased studies, the heatmap of Figure 5 illustrates the coclustering results. At the bottom of the figure, we labeled the columns with the posterior probability of bias P(Ii|Data), and on the right, we labeled the rows with the studies' numbers and the true bias status. For example, in the top left corner, Study‐13 is detected to have “large bias” with a posterior probability of bias of 0.512. The heatmap shows that the studies are grouped into two distinct clusters corresponding to the unbiased and biased studies. However, a unique feature of the BC‐BNP model is the possibility to explore coclustering within the biased studies generated by the DP. In this example, five studies (20, 9, 17, 4, and 12) and two studies (10, 16) shared the same cluster, while the remaining three biased studies (7, 13, and 19) did not cocluster with the others.

FIGURE 5.

FIGURE 5

Simulation results for N=20: Heatmap of coclustering between studies sharing the same cluster. The label of the columns corresponds to the posterior probability of bias P(Ii|Data) and the label of the rows to the true bias status (No B = no bias, Mild B = mild bias, and Large B = large bias).

3.4. Conclusions of the Simulation Results

The simulation experiment of this section was neither extensive nor fully conclusive but it was challenging for any meta‐analysis model. The aim was to evaluate the BC‐BNP under different situations that represented the case studies of Section 4, and that can be found in practice. We can summarize the trends in the results of our simulation experiment as follows: (1) For the scenarios generated, the BC‐BNP model can detect bias at the study level when bias is present, and can detect nonbias when bias is absent. (2) The bias correction is adaptive according to the level of bias (no bias, mild, large, and extreme). (3) The posteriors of μθ and τθ are robust against biased studies. (4) It is possible to explore the coclustering of bias studies. These conclusions were the same for N=20 and N=30.

4. Applications

In this section, we present two real meta‐analyses. In each one, we apply the BC‐BNP model using two settings of the prior distributions. First, we apply the BC‐BNP model with informative priors, where the prior distribution of μβ reflects the direction of bias, and the prior distribution of πB is elicited using some quality information from the studies. Second, a sensitivity analysis is performed using the default priors of the BC‐BNP model. In addition, we compare the results of the BC‐BNP model with the parametric BC model and the Bayesian random effects model.

In the Supporting Information of the paper, we present the R script to run the analyses of this section. The numerical results are based on four parallel MCMC chains, with a length of 50,000 iterations. In each chain, we discarded the first 10,000 iterations as a burn‐in period, and we apply the remaining 40,000 iterations to approximate the posterior distributions.

4.1. Meta‐Analysis of RCTs With Varying Quality: Stem‐Cell Treatment and Cardiovascular Disease Patients

Nowbar et al. (2014) performed a meta‐analysis consisting in N=31 RCTs of heart disease patients, where a treatment based on bone marrow stem cell was assessed for efficacy. In each trial, patients were randomized to a treatment group receiving bone marrow stem cell or a control group receiving placebo treatment. The primary endpoint was the ejection fraction. The treatment effect was defined as the difference in means of the ejection fraction between groups. The data of this meta‐analysis can be found in dataframe “stemcells” in the R package jarbes (Verde 2024).

What makes this meta‐analysis particularly interesting is the large heterogeneity between studies and the context in which these 31 RCTs were published. On the one hand, there was a high interest in assessing the efficacy of this type of treatment, but on the other hand, the varying quality of the published RCTs cast doubts on their results (Francis et al. 2013), and further investigation has found evidence of scientific misconduct (Abbott 2014).

In particular, Nowbar et al. (2014) investigated if the number of detected discrepancies in the published studies were correlated with efficacy. In this context, discrepancies were defined as two or more reported facts that cannot both be true because they are logically or mathematically incompatible. The median number of discrepancies in the meta‐analysis were 7, with a range of [0, 55], and only three studies have 0 discrepancies.

To assess the heterogeneity between studies, we applied a noninformative Bayesian normal random effects model to this meta‐analysis. The results showed a posterior pooled mean difference of μθ=2.92 with 95% posterior interval of [1.43, 4.37], which pointed out efficacy. However, the posterior between‐studies standard deviation σθ=3.44, with a posterior 95% interval [2.40, 4.80] shows a very large heterogeneity. In addition, the predictive 95% posterior predictive interval of a future study effect θnew was [−4.09, 9.93], indicating no efficacy. We can suspect that the heterogeneity was a clear combination of studies' internal validity biases and diversity, making this meta‐analysis an interesting case study for the BC‐BNP model.

We apply the BC‐BNP model presented in Section 2 to this meta‐analysis, using weakly informative priors for the model of interest. For the bias component, we elicit the hyperparameters as follows:

  • We assume a positive direction of bias with a prior μβUniform(0,15).

  • For illustration, we elicit the prior of πBBeta(a0,a1) by using the number of discrepancies reported by Nowbar et al. (2014), where 18 studies have five or more discrepancies. We calculate a0 and a1 such that the median of the Beta(a0,a1) is 18/310.581 and its 90th quantile is 28/310.903, which result in a0=1.51 and a1=1.17. It is worth mentioning that in a real application, this prior should be elicited without using the data of the meta‐analysis.

  • For the concentration parameter α we used a uniform distribution αUniform(0.5,αMax), and by using (20) we have that αMax=3.29, and KMax=17.

  • For a prior‐to‐posterior sensitivity analysis we used the default priors, without bias direction μβUniform(15,15), and πBBeta(0.5,1), which resulted in αMax=1.87, and KMax=10.

The solid line in Figure 6 presents the posterior distribution of the μθ using a normal–normal random effects model, while the dashed line corresponds to the BC‐BNP with informative priors. The posterior mean of μθ using the BC‐BNP supports inferring no efficacy with a posterior mean and 95% posterior interval of 1.12 [−2,29, 2.92], respectively. The dotted line presents the resulting posterior of μθ by applying the BC‐BNP with default priors. In this case, the posterior mean was 1.51 [−0.74, 3.17]. This analysis shows that the conclusion of the lack of efficacy was not sensitive with respect to the priors of the bias component.

FIGURE 6.

FIGURE 6

Results of the BC‐BNP model applied to efficacy of stem‐cell treatment for cardiovascular disease patients. The posterior distributions of the pooled difference in means between treatments are displayed for the BC‐BNP with informative priors (dashed line), the BC‐BNP with default priors (dotted line), and the normal–normal random effects model (solid line).

In addition, the posterior of μθ using the parametric BC with default priors resulted in a posterior mean of 1.45, with a posterior 95% interval of [−1.27, 2.99], which is consistent with the BC‐BNP model.

Figure 7 shows the model diagnostic for the direction of bias. The left panel displays the joint posterior distribution between the bias correction Δ=μβμθ and the probability of bias πB resulted from the BC‐BNP analysis with informative priors. The right panel shows the resulting posteriors using the BC‐BNP with default priors. These diagnostic plots show that results remain stable after applying the two priors' settings.

FIGURE 7.

FIGURE 7

Stem‐cell treatment and cardiovascular disease patients case study. Sensitivity analysis of the priors for the bias component in the BC‐BNP model: Joint posterior distribution of the mean bias and probability of bias. The scatter plots correspond to random samples from the MCMC iterations. Left panel: Using informative prior distributions has concentrated the range of the posterior distributions. Right panel: Effect of using default priors for the bias direction and the probability of bias. The posterior of the bias remains stable, and more concentrated than the left panel.

With respect to the structure of the bias model, the posterior distributions of the number of clusters K showed lower number of clusters for default priors. For the informative priors, K has a posterior median of 6 with a 95% posterior interval [2, 10], and for default priors a median of 4 and a 95% posterior interval of [1, 8].

The forest plot in Figure 8 presents the effect of the bias correction at the study level using the BC‐BNP model with default priors. The main pattern observed in this forest plot is a shrinkage and bias correction effect for studies with larger values of θB in the direction of the posterior mean E(μθ|Data)=1.51. The bias correction is adaptive; for example, studies 1, 5, 24, and 31 have posterior probabilities of being biased of 0.24, 0.88, 0.96, and 0.96, respectively.

FIGURE 8.

FIGURE 8

Stem‐cell treatment and cardiovascular disease patients case study. BC‐BNP model with default priors, forest plot comparing for each study the posterior distributions of the biased study effect θB (biased), and the biased‐corrected θ. The vertical dashed line corresponds to E(μθ|Data)=1.51.

In study 1, where bias correction is unnecessary, both posteriors overlap. For studies 5, 24, and 31, a strong bias correction is automatically performed in the direction of the posterior mean of μθ=1.51. It is worth mentioning that in the presence of bias the posteriors of θ are wider than the posterior of θB for studies 5, 24, and 31, which is “the price” we pay for bias correction at the study level.

The heatmap of Figure 9 shows a coclustering structure that resulted in three clusters. Cluster 1 corresponds to the unbiased studies with a posterior probability of bias in the range [0–0.21], cluster 2 with a range of [0.32–0.56], and four studies coclustered at the bottom right with a range of [0.76–1].

FIGURE 9.

FIGURE 9

Heatmap of coclustering analysis of studies using default priors for the bias component. The labels of the columns correspond to the posterior probability of bias P(Ii|Data) and the labels of the rows correspond to the cluster label and the study number. In this example, the BC‐BNP model generated three clusters: Cluster 1 corresponds to a posterior probability of bias in the range [0–0.22], cluster 2 with a range of [0.32–0.56], and four studies coclustered at the bottom right with a range of [0.76–1].

Nowbar et al. (2014) reported a lack of information about sequence allocation (randomization) in the Risk of Bias evaluation was linked with exaggerated treatment effects. Among the five studies in cluster 2, three did not provide clear information about sequence allocation while none of the four studies in cluster 3 had clear information. These results suggest that coclustering may be useful in linking the lack of study quality to exaggerated results.

4.2. Generalized Evidence Synthesis of the Relationship Between Hypertension and Severity in COVID‐19 Patients

This case study, which we briefly presented in the introduction, is a meta‐analysis of OS performed by de Almeida‐Pititto et al. (2020). The aim of the authors was to evaluate the impact of diabetes, hypertension, cardiovascular disease, and the use of angiotensin converting enzyme inhibitors/angiotensin II receptor blockers (ACEI/ARB) with the primary outcomes (1) COVID‐19 severity (including need for invasive mechanical ventilation or intensive care unit admission or O2 saturation less than90%) and (2) intrahospital mortality due to confirmed COVID‐19.

We illustrate the BC‐BNP model by analyzing the relationship between hypertension and COVID‐19 severity, where the number of included studies was N=18. From these 18 studies, five were case series, three were cross‐sectional, and 10 were retrospective cohorts. According to their study design, these studies are at the lower level of clinical evidence (Guyatt et al. 1995). Therefore, the 18 studies in this meta‐analysis are prone to internal validity bias.

This type of meta‐analysis, which combines studies of different types, is called Generalized Evidence Synthesis. There are two simple approaches for this type of meta‐analysis. On the one hand, we can ignore study types and combine all studies using a random effects meta‐analysis model. This is the approach that was applied by de Almeida‐Pititto et al. (2020). On the other hand, we can perform a meta‐analysis by study types, this approach has the assumption that the variability between study types is so large that we cannot gain any information by combining studies' results. We can expect that neither ignoring study types nor assuming that results are extremely different is the best way to synthesize these data.

We apply the BC‐BNP model by using weakly informative priors for the model of interest. For the bias component, we elicit the hyperparameters as follows:

  • We truncate the bias prior to be positive ΔUniform(0,15).

  • For the probability of bias πBBeta(a0,a1), we elicit the a0 and a1 by using the distribution of the different study types. We calculate a0 and a1 such most of the studies in this meta‐analysis are at risk of internal validity bias. We arbitrarily take that the median of the Beta(a0,a1) is 15/180.83 and its 90th quantile is 17/180.944, which results in a0=8.6 and a1=1.97. This is an informative prior with higher probability values for large values of πB.

  • For the concentration parameter α, we use a uniform distribution αUniform(0.5,αMax), by using (20) αMax=2.73. In this setup, the maximum number of clusters is KMax=15.

  • We perform a sensitivity analysis of the direction of bias setting the default prior μβUniform(15,15) and the bias information of the study types by using πbiasBeta(0.5,1). This prior resulted in αMax=1 and the maximum number of clusters KMax=6.

The solid line in Figure 10 shows the posterior distribution of the OR using a normal random effects model that shows a strong association between hypertension and complications, 2.96 [2.33, 3.76], while the dashed line corresponds to the bias‐corrected posterior using the BC‐BNP with informative priors that indicates far less support for there being an association 1.98 [0.52, 3.35]. The dotted line shows the resulting posterior when we ignore bias direction and study types, the OR has a mean posterior of 2.57 [1.60, 3.86]. The posterior of the OR using the parametric BC with default priors resulted in a posterior mean of 2.60 [1.96, 3.53], which is consistent with the default BC‐BNP model.

FIGURE 10.

FIGURE 10

COVID‐19‐infected patients case study. Results of the BC‐BNP model applied to the relationship between hypertension and complications. The posterior distributions of the pooled OR are displayed for the BC‐BNP with informative priors (dashed line), the BC‐BNP with default priors (dotted line), and the normal–normal random effects model (solid line). The vertical dashed line corresponds to OR=1.

The left panel of Figure 11 shows the joint posterior distribution between the bias correction Δ and the probability of bias πbias by using informative priors. We see a high probability of bias and strong concentration of the bias correction. We can compare this result with the right panel of Figure 11, which shows the same joint posterior by using the default priors. Clearly, the use of informative priors plays an important role in this meta‐analysis.

FIGURE 11.

FIGURE 11

COVID‐19‐infected patients case study. Sensitivity analysis of the priors for the bias component in the BC‐BNP model: Joint posterior distribution of the mean bias and probability of bias. The scatter plots correspond to random samples from the MCMC iterations. Left panel: Using informative prior distributions has concentrated the range of the posterior distributions. Right panel: Effect of using default distributions for the bias direction and the probability of bias.

These results show that the assessment of the study quality, and the bias direction strongly influence the conclusions from this meta‐analysis. Therefore, given the low level of clinical evidence of these N=18 studies, we could conclude that there is great uncertainty about the strength of an association.

Regarding the structure of the bias model, the posterior distributions of the number of clusters K are as following: For the informative priors, K has a median of 4 with a 95% posterior interval [1, 8], while for the default priors, K has a median of 2 with a 95% posterior interval of [1, 5].

The forest plot of Figure 12 presents the effect of the bias correction using the BC‐BNP with informative priors. We can see that the effect of informative priors is to shift the posteriors of θ to the posterior mean E(μθ|Data)=0.59.

FIGURE 12.

FIGURE 12

COVID‐19‐infected patients case study using BC‐BNP with informative priors. Forest plot comparing for each study the posterior distributions of the biased study effect θB, and the biased‐corrected θ. The vertical dashed line corresponds to E(μθ|Data)=0.59.

We can compare the results for studies 1, 8, and 18. Under the BC‐BNP model with informative priors, these studies have posterior probabilities of being biased of 0.95, 1, and 0.90, respectively. For studies 1, 8, and 18, we observe a shift to the left in the posteriors of θ due to the bias correction effect. In order to interpret these results, we can refer to Table 1, where the estimated OR for studies 1 (Gua et al. 2020), 8 (Xiang et al. 2020), and 18 (Zhang et al. 2020), were 6.48, 12.60, and 4.37, respectively.

The heatmap of Figure 13 shows that the cocluster information using the default priors agglomerated studies in three clusters. Xiang et al. (2020) is the only study in cluster 3, while the other two studies shared cluster 2. In this example, the BC‐BNP model has detected studies with overestimated ORs and low‐quality study design. It is worth mentioning that we requested a risk of bias evaluation from de Almeida‐Pititto et al. (2020), but this evaluation was neither published nor sent to us. This case study shows that the BC‐BNP model can be a useful tool to assess risk of bias by only using information about the reported effects and their standard errors.

FIGURE 13.

FIGURE 13

Heatmap of coclustering analysis of studies using default priors for the bias component. The labels of the columns correspond to the posterior probability of bias P(Ii|Data) and the labels of the rows correspond to the cluster label and first author of the study. In this example, the BC‐BNP model generated three clusters: The cluster 1 corresponds to studies with a probability of bias in the range [0–0.287], the cluster 2 to [0.332–0.648], and a single study, Xiang et al. (2020), is assigned to cluster 3 representing an extreme reported result. Note: Source of the data de Almeida‐Pititto et al. 2020.

5. Discussion and Conclusion

Meta‐analysis deals with combining indirect evidence in statistics, where meta‐analysis of RCTs represents the highest level of clinical evidence. However, the COVID‐19 pandemic showed that urgent informed decisions sometimes have to be based on imperfect pieces of evidence, evidence that may contain conflicting results, or be prone to bias.

In this paper, we presented the BC‐BNP model, which aims to automatically correct for internal validity bias in meta‐analysis by relaxing the parametric assumptions of the bias model introduced by Verde (2021).

We evaluated the BC‐BNP model using simulated data sets and we illustrate its application with two real case studies. Our results showed several potential advantages of the BC‐BNP in practical applications. First, the BC‐BNP model can detect bias when present and yields results similar to a simple normal–normal random effects model when bias is absent. Second, relaxing the parametric assumptions of the bias component does not affect the model of interest and provides results consistent with Verde (2021) model. Third, having a BNP component of bias may offer a better insight into the nature of biases by clustering studies with similar biases.

The setup of the default hyperpriors we provided represent a neutral elicitation on the direction of bias and a relative optimistic assumption about the number of studies at risk of bias. However, external validity bias represents unobserved external factors that cannot be directly estimated from studies' results. Therefore, we recommend performing a sensitivity analysis on the direction of bias and on the number of studies at risk of bias.

The examples in Section 4 demonstrated the extent to which bias correction is possible and whether this correction remains stable after a sensitivity analysis. For the stem‐cell example, the results were consistent across different hyperpriors, indicating that there is sufficient observed information to assert the presence of bias. However, in the COVID example, the results differed between the informative and default priors. In this case, the BC‐BNP model indicates that combining results at face value maybe misleading, and a prior elicitation is crucial to arrive at useful conclusions.

The BNP component of the BC‐BNP model demonstrates its potential as an exploratory tool for clustering studies with varying levels of bias. We can expect this feature to be even more useful in more complex meta‐analysis models (e.g., metaregression, hierarchical meta‐analysis of subgroups), but it remains a topic for further research.

There are several potential extensions of the BC‐BNP model that we did not cover in this paper. One important topic is using exact likelihood contributions in the meta‐analysis model. In this paper, we simplify this point by assuming normal likelihood approximations, but we know that different endpoint types may influence meta‐analysis results. An important type of bias in meta‐analysis is the publication bias, where small‐sample studies may overreport significant results. A possible extension of the BC‐BNP could involve modifying the base distribution of the DP to include the sample sizes of the studies and correct for publication bias. Other important extensions include a hierarchical BC‐BNP model combining several studies subgroups within each publication, and external validity bias correction by combining AD and IPD in meta‐analysis (Verde 2017). Furthermore, a methodological comparison of BNP models in meta‐analysis and the extent to which these models indirectly correct for bias remains an open research topic.

This work leads to two important conclusions. First, extending the flexibility of the random effects distribution is a valuable approach. One practical aspect, demonstrated empirically in this paper, is the possibility to obtain results that are robust against biased studies. Second, ignoring internal validity bias in meta‐analysis constitutes a form of model misspecification that can lead to misleading conclusions. The approach presented in this paper offers a promising solution to this important problem in meta‐analysis.

Conflicts of Interest

The authors declare no conflicts of interest.

Open Research Badges

This article has earned an Open Data badge for making publicly available the digitally‐shareable data necessary to reproduce the reported results. The data is available in the Supporting Information section.

This article has earned an open data badge “Reproducible Research” for making publicly available the code necessary to reproduce the reported results. The results reported in this article could fully be reproduced.

Supporting information

Supporting Information

Acknowledgments

We want to express our gratitude to the editor, Prof. Monica Chiogna, for her patience throughout the paper revision process. We also thank the three anonymous reviewers whose critiques and suggestions that significantly improve the quality of this paper. We are thankful to Dr. Hornung and the reproducible research team for enhancing the quality of our R scripts and making the results of this paper trustfully. Pablo E. Verde extends his appreciation to Dr. Ondrej Pokora for the engaging discussions on stochastic processes during his visit to the Department of Mathematics and Statistics at Masaryk University in Brno, Czech Republic. This visit was financially supported by the INOLLEC lecture series. In addition, we gratefully acknowledge the financial support from the Deutsche Forschungsgemeinschaft (DFG) under grant number DFG 503988801.

Funding: This study was supported by Deutsche Forschungsgemeinschaft (DFG) (grant number DFG 503988801).

Data Availability Statement

The data that support the findings of this study are openly available in the R package jarbes at https://cran.r‐project.org/web/packages/jarbes/index.html.

References

  1. Abbott, A. 2014. “Evidence of Misconduct Found Against Cardiologist.” Nature Newsblog. https://blogs.nature.com/news/2013/07/german-cardiologists-stem-cell-papers-attacked.html.
  2. Arbel, J. , De Blasi P., and Prünster I.. 2019. “Stochastic Approximations to the Pitman–Yor Process.” Bayesian Analysis 14, no. 4: 1201–1219. [Google Scholar]
  3. Branscum, A. J. , and Hanson T. E.. 2008. “Bayesian Nonparametric Meta‐Analysis Using Polya Tree Mixture Models.” Biometrics 64, no. 3: 825–833. [DOI] [PubMed] [Google Scholar]
  4. Burr, D. 2012. “Bspmma: An R Package for Bayesian Semiparametric Models for Meta‐Analysis.” Journal of Statistical Software 50, no. 4: 1–23.25317082 [Google Scholar]
  5. Burr, D. , and Doss H.. 2005. “A Bayesian Semiparametric Model for Random‐Effects Meta‐Analysis.” Journal of the American Statistical Association 100, no. 469: 242–251. [Google Scholar]
  6. de Almeida‐Pititto, B. , Dualib P. M., Zajdenverg L., et al. 2020. “Severity and Mortality of COVID‐19 in Patients With Diabetes, Hypertension and Cardiovascular Disease: A Meta‐Analysis.” Diabetology & Metabolic Syndrome 12: 75. [DOI] [PMC free article] [PubMed] [Google Scholar]
  7. Dias, S. , Welton N. J., Marinho V. C. C., Salanti G., Higgins J. P. T., and Ades A. E.. 2010. “Estimation and Adjustment of Bias in Randomized Evidence by Using Mixed Treatment Comparison Meta‐Analysis.” Journal of the Royal Statistical Society: Series A (Statistics in Society) 173, no. 3: 613–629. [Google Scholar]
  8. Droitcour, J. , Silberman G., and Chelimsky E.. 1993. “Cross‐Design Synthesis: A New Form of Meta‐Analysis for Combining Results From Randomized Clinical Trials and Medical‐Practice Databases.” International Journal of Technology Assessment in Health Care 9, no. 3: 440–449. [DOI] [PubMed] [Google Scholar]
  9. Dunson, D. , Xue Y., and Carin L.. 2008. “The Matrix Stick‐Breaking Process.” Journal of the American Statistical Association 103, no. 481: 317–327. [Google Scholar]
  10. Eddy, D. M. , Hasselblad V., and Shachter R.. 1992. Meta‐Analysis by the Confidence Profile Method: The Statistical Synthesis of Evidence. San Diego, CA: Academic Press. [Google Scholar]
  11. Francis, D. P. , Mielewczik M., Zargaran D., and Cole G. D.. 2013. “Autologous Bone Marrow‐Derived Stem Cell Therapy in Heart Disease: Discrepancies and Contradictions.” International Journal of Cardiology 168: 3381–3403. [DOI] [PubMed] [Google Scholar]
  12. Gelman, A. 2006. “Prior Distributions for Variance Parameters in Hierarchical Models.” Bayesian Analysis 1: 1–19. [Google Scholar]
  13. Gelman, A. , Carlin J., Stern H., Dunson D., Vehtari A., and Rubin D.. 2013. Bayesian Data Analysis, Third Edition, Chapman & Hall/CRC Texts in Statistical Science, 545–573. Taylor & Francis. [Google Scholar]
  14. Griffin, J. E. 2016. “An Adaptive Truncation Method for Inference in Bayesian Nonparametric Models.” Statistics and Computing 26, no. 1: 423–441. [Google Scholar]
  15. Guyatt, G. H. , Sackett D. L., Sinclair J. C., et al. 1995. “Users' Guides to the Medical Literature: IX. A Method for Grading Health Care Recommendations.” Journal of the American Medical Association 274, no. 22: 1800–1804. [DOI] [PubMed] [Google Scholar]
  16. Higgins, J. , Thompson S., and Spiegelhalter D.. 2009. “A Re‐Evaluation of Random‐Effects Meta‐Analysis.” Journal of the Royal Statistical Society, Series A 172: 127–159. [DOI] [PMC free article] [PubMed] [Google Scholar]
  17. Huang, A. , and Wand M. P.. 2013. “Simple Marginally Noninformative Prior Distributions for Covariance Matrices.” Bayesian Analysis 8, no. 2: 439–452. [Google Scholar]
  18. Karabatsos, G. , Talbott E., and Walker S. G.. 2015. “A Bayesian Nonparametric Meta‐Analysis Model.” Research Synthesis Methods 6, no. 1: 28–44. [DOI] [PubMed] [Google Scholar]
  19. Karabatsos, G. , and Walker S. G.. 2012. “Adaptive‐Modal Bayesian Nonparametric Regression.” Electronic Journal of Statistics 6, no. no: 2038–2068. [Google Scholar]
  20. Kolde, R. 2019. Pheatmap: Pretty Heatmaps . R Package Version 1.0.12.
  21. Nowbar, A. N. , Mielewczik M., Karavassilis M., et al. 2014. “Discrepancies in Autologous Bone Marrow Stem Cell Trials and Enhancement of Ejection Fraction (DAMASCENE): Weighted Regression and Meta‐Analysis.” British Medical Journal 348: g2688. [DOI] [PMC free article] [PubMed] [Google Scholar]
  22. Ohlssen, D. I. , Sharples L. D., and Spiegelhalter D. J.. 2007. “Flexible Random‐Effects Models Using Bayesian Semi‐Parametric Models: Applications to Institutional Comparisons.” Statistics in Medicine 26, no. 9: 2088–2112. [DOI] [PubMed] [Google Scholar]
  23. Plummer, M. 2003. “JAGS: A Program for Analysis of Bayesian Graphical Models Using Gibbs Sampling JAGS: Just Another Gibbs Sampler.” In Proceedings of Distributed Statistical Computing.
  24. Poli, G. , Fountzilas E., Tsimeridou A.‐M., and Müller P.. 2023. “A Multivariate Polya Tree Model for Meta‐Analysis With Event Time Distributions.” Preprint. 10.48550/arXiv:2312.06018. [DOI] [PubMed]
  25. Prevost, T. , Abrams K., and Jones D.. 2000. “Hierarchical Models in Generalized Synthesis of Evidence: An Example Based on Studies of Breast Cancer Screening.” Statistics in Medicine 19, no. 24: 3359–3376. [DOI] [PubMed] [Google Scholar]
  26. Rosner, G. , Laud P., and Johnson W.. 2021. Bayesian Thinking in Biostatistics, Chapman & Hall/CRC Texts in Statistical Science, 65–69. CRC Press. [Google Scholar]
  27. Röver, C. , Bender R., Dias S., et al. 2021. “On Weakly Informative Prior Distributions for the Heterogeneity Parameter in Bayesian Random‐Effects Meta‐Analysis.” Research Synthesis Methods 12, no. 4: 448–474. [DOI] [PubMed] [Google Scholar]
  28. Sethuraman, J. 1994. “A Constructive Definition of Dirichlet Priors.” Statistica Sinica 4, no. 2: 639–650. [Google Scholar]
  29. Spiegelhalter, D. , Abrams K. R., and Myles J. P.. 2004. Bayesian Approaches to Clinical Trials and Health‐Care Evaluation. Chichester: Wiley. [Google Scholar]
  30. Spiegelhalter, D. , Thomas A., Best N., and Lunn D.. 2007. WinBUGS User Manual Version 1.4.3 . Cambridge: Medical Research Council Biostatistics Unit. [Google Scholar]
  31. Verde, P. E. 2017. “Two Examples of Bayesian Evidence Synthesis With the Hierarchical Meta‐Regression Approach.” In Bayesian Inference, edited by Tejedor J. P., 189–206. InTech. 10.5772/intechopen.70231. [DOI] [Google Scholar]
  32. Verde, P. E. 2018. “bamdit: An R Package for Bayesian Meta‐Analysis of Diagnostic Test Data.” Journal of Statistical Software 86, no. 10: 1–32. [Google Scholar]
  33. Verde, P. E. 2021. “A Bias‐Corrected Meta‐Analysis Model for Combining, Studies of Different Types and Quality.” Biometrical Journal 63, no. 2: 406–422. [DOI] [PubMed] [Google Scholar]
  34. Verde, P. E. 2024. “jarbes: An R Package for Bayesian Evidence Synthesis.” Version 2.2.3.
  35. Verde, P. E. , and Ohmann C.. 2015. “Combining Randomized and Non‐Randomized Evidence in Clinical Research: A Review of Methods and Applications.” Research Synthesis Methods 6, no. 1: 45–62. [DOI] [PubMed] [Google Scholar]
  36. Welton, N. J. , Ades A. E., Carlin J. B., Altman D. G., and Sterne J. A. C.. 2009. “Models for Potentially Biased Evidence in Meta‐Analysis Using Empirically Based Priors.” Journal of the Royal Statistical Society: Series A (Statistics in Society) 172, no. 1: 119–136. [Google Scholar]
  37. Zhang, J. , and Dassios A.. 2023. “Truncated Poisson–Dirichlet Approximation for Dirichlet Process Hierarchical Models.” Statistics and Computing 30, no. 3: 743–759. [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Supporting Information

Data Availability Statement

The data that support the findings of this study are openly available in the R package jarbes at https://cran.r‐project.org/web/packages/jarbes/index.html.


Articles from Biometrical Journal. Biometrische Zeitschrift are provided here courtesy of Wiley

RESOURCES