Skip to main content
UKPMC Funders Author Manuscripts logoLink to UKPMC Funders Author Manuscripts
. Author manuscript; available in PMC: 2024 Dec 20.
Published in final edited form as: Biom J. 2024 Apr 18;66(3):bimj.202200316. doi: 10.1002/bimj.202200316

Sharing information across patient subgroups to draw conclusions from sparse treatment networks

Theodoros Evrenoglou 1,, Silvia Metelli 1, Johannes-Schneider Thomas 2, Spyridon Siafis 2, Rebecca M Turner 3, Stefan Leucht 2, Anna Chaimani 1
PMCID: PMC7617248  EMSID: EMS201286  PMID: 38637311

Abstract

Network meta-analysis (NMA) usually provides estimates of the relative effects with the highest possible precision. However, sparse networks with few available studies and limited direct evidence can arise, threatening the robustness and reliability of NMA estimates. In these cases, the limited amount of available information can hamper the formal evaluation of the underlying NMA assumptions of transitivity and consistency. In addition, NMA estimates from sparse networks are expected to be imprecise and possibly biased as they rely on large-sample approximations that are invalid in the absence of sufficient data. We propose a Bayesian framework that allows sharing of information between two networks that pertain to different population subgroups. Specifically, we use the results from a subgroup with a lot of direct evidence (a dense network) to construct informative priors for the relative effects in the target subgroup (a sparse network). This is a two-stage approach where at the first stage, we extrapolate the results of the dense network to those expected from the sparse network. This takes place by using a modified hierarchical NMA model where we add a location parameter that shifts the distribution of the relative effects to make them applicable to the target population. At the second stage, these extrapolated results are used as prior information for the sparse network. We illustrate our approach through a motivating example of psychiatric patients. Our approach results in more precise and robust estimates of the relative effects and can adequately inform clinical practice in presence of sparse networks.

Keywords: informative priors, mixed treatment comparisons, sharing information, sparse data

1. Introduction

Network meta-analysis (NMA) has become an essential tool for the comparative effectiveness research of healthcare interventions because it allows integrating all available information on a specific disease and usually provides estimates of relative effects with the highest possible precision (Chaimani et al., 2019; Nikolakopoulou et al., 2018). However, networks of interventions with limited data that fail to provide useful conclusions may arise under certain situations. A recent empirical study found 92 (7.4%) published NMAs that included more treatments than the number of studies in a sample of 1236 NMAs of randomized controlled trials (RCTs) with at least four interventions (Veroniki et al., 2021). These numbers suggest that the phenomenon of “sparse” networks, namely, networks with limited direct evidence in the form of few direct comparisons and/or few—1 or 2—studies per comparison, is not very rare in the literature. For example, this is often the case for “sensitive” subgroups of the population that cannot be easily included in trials, such as children, elder patients, or individuals with multimorbidity.

Results from such networks of interventions are accompanied with substantial uncertainty not only in the estimation of the relative treatment effects but also in the plausibility of the underlying assumptions because their formal evaluation is impossible (Dias et al., 2010). In addition, the large-sample approximations on which NMA models typically rely upon fail in the presence of only a handful of studies per comparison. Hence, lack of robustness and potentially limited reliability are common issues that might be encountered when analyzing sparse treatment networks (Yoon et al., 2021). To avoid these issues, it might seem reasonable to wait until more studies become available for the outcome(s) of interest. However, the pace of trial production in these contexts is usually very slow and the chance to obtain new study results shortly is low.

The use of external evidence in the form of informative prior distributions has been suggested previously in meta-analysis, either to achieve more accurate estimation of the heterogeneity parameter (Röver et al., 2019; Turner et al., 2019) or for the incorporation of non-randomized evidence (Efthimiou et al., 2017). Here, we introduce a new framework that allows sharing information between two networks of treatments that pertain to different subgroups of the population. Our framework refers to the case when only aggregate data are available; in the presence of individual participant data, other approaches may be used, such as population adjustment methods (Phillippo et al., 2020). Specifically, we use the results from a subgroup with a lot of direct evidence (i.e., many direct comparisons and many studies per comparison) forming a “dense” network to construct informative priors for the relative effects in the target subgroup forming a sparse network.

Information sharing across population subgroups is also common outside the context of NMA. For example, similar approaches are used in the so-called “basket trials,” where information is borrowed from large baskets to facilitate the estimation of treatment effects in smaller baskets (Parmar et al., 2017; Renfro & Sargent, 2017; Turner et al., 2022) involving subpopulations that are challenging to include in clinical trials. The idea of extrapolating information from a population with a lot of data to another with limited evidence has been discussed previously in an FDA report (U.S. Department of Health & Human Services Food & Drug Administration, 2016) that used adult data to draw conclusions for children for the case of individual studies. Extending these ideas to the meta-analytic setting would allow us to obtain more robust conclusions as individual trials are often underpowered (Salanti et al., 2008).

In this manuscript, we propose a two-stage approach where at the first stage, we synthesize the data from the dense network using a hierarchical NMA model. This includes a location parameter that shifts the distribution of the relative effects to make them applicable to the population of the sparse network. We also add a scale parameter that allows us to further downweight the external data and to reduce their influence. At the second stage, the results from the first stage are used as prior information in the analysis of the sparse network. We inform the location and scale parameters either through the data or using expert opinion. We illustrate our approach through an example examining the relative effects of antipsychotic treatments in two subgroups of patients: a target subgroup of children adolescents (CAs) that forms a rather sparse network and a subgroup of chronic adult patients with acute exacerbation of schizophrenia, referred here as “general patients” (GP), which forms a dense network.

2. Motivating Dataset

Our work is motivated by a recent article investigating the differences in the treatment effects among several subgroups of schizophrenia patients (Leucht et al., 2022). The aim of the article was to compare the subgroup of GP, defined as chronic patients with an acute exacerbation of positive symptoms, with more specific subgroups of patients: CA, first-episode patients, treatment-resistant patients, patients with negative symptoms, patients under substance abuse, and elder patients. In contrast to the dataset for GP, all these specific subgroups were informed by very limited direct evidence, and therefore, the original authors, who acknowledged the problems of analyzing sparse networks, only used pairwise meta-analysis. To illustrate our approach, we use the data from two of the above subgroups: CA and GP. The former is considered the target subgroup that forms a sparse network and the latter is the only available subgroup with abundant direct evidence (forming a dense network) that can be safely used to elicit informative priors for CA.

The CA network comprises 19 RCTs published between 1973 and 2017 comparing 14 antipsychotics and placebo, and providing direct information for 21 out of 105 possible comparisons (Figure 1a). Out of these 21 comparisons, two include two RCTs and the rest include one RCT only. In addition, many of these RCTs are small, with a median study sample size of 113 patients. The pace of new RCT production for this subgroup is very slow (on average, one RCT every 2.3 years), which confirms the difficulty of conducting RCTs for this group of patients. On the other hand, the network for GPs contains 255 RCTs comparing 33 antipsychotics and placebo, and forming 116 direct comparisons (Figure 1b). We also used the community detection algorithm proposed by Law et al. (2019) to examine how sparse or dense the two networks are (Online Appendix 1). The conclusions from this approach are in agreement with what can be obtained by descriptive characteristics and visual inspection of the network diagrams. Specifically, almost all treatments except one (Molindone) in the GP network appear to be very close to each other and thus well identified (Appendix 1, Figure 1). In contrast, six treatments in the CA network appear isolated that reveals that at least 40% of this network is not well identified.

Figure 1. Sparse network for the population of CA (Panel A) and informative network for GP (Panel B).

Figure 1

The size of the edges (lines in the networks) is proportional to the number of studies that compare the respective treatments, whereas the size of the nodes is proportional to number of patients who received each treatment.

The two networks are distinct, which means that they do not have any RCT in common. However, some drugs are included in both networks; this is a necessary condition to allow sharing information between them. The outcome of interest is the reduction of the overall schizophrenia symptoms that is measured in a continuous scale. Following the original article, we used the standardized mean difference (SMD) as effect measure because the different RCTs have used different scales.

3. Methods

3.1. Notation and general setting

Suppose that we have a set of N = n1 + n2 studies. Studies {1, 2, …, n1} inform the target population subgroup P1 and form a sparse network, while the rest {n1 + 1, n1 + 2,…, N} inform a different population subgroup P2 and form a dense network. Suppose also that T1 treatments have been evaluated for population P1 and T2 treatments for population P2. We denote with Ta = T1T2 the set of all available treatments and with Tc = T1T2 the set of common treatments included in both networks; the latter is the group of treatments for which information can be extrapolated from P2 to the target subgroup P1. Each study i provides the observed mean (change score/endpoint) yitk for the treatment tkTa of arm k with k = 1, …, Ki and Ki the total number of arms in i. For simplicity, in what follows we write the yitk as yik, its variance sdik2 , and the respective “true” mean is denoted as θik. We consider treatment j = 1 as the common network reference for both P1 and P2 and the relative effects µ1j (j > 1, jTa) as the basic parameters. For every study, we arbitrarily choose the treatment of arm k = 1 (t1) as the study-specific baseline treatment.

3.2. Models for sharing information between two population subgroups

Different models stem from different assumptions about the relation between the two subgroups P1 and P2. We start with the description of the standard hierarchical NMA model that in the present setting synthesizes the two subgroups in a naïve way, namely, as if they were equivalent. Then, we move to more plausible NMA models in which we incorporate a location parameter acknowledging the difference between P1 and P2 and a scale parameter aiming to increase the uncertainty of the studies in subgroup P2.

3.2.1. Naïve synthesis

In this approach, the two subgroups are combined together as if all participants were coming from the same population P. In other words, this approach makes the strong assumption that P1 and P2 are equivalent, namely, P1P2P. Hence, for every study i = 1, … N and treatment tkTα, the observed means follow a normal distribution

yikN(θik,sdik2). (1)

Then, the underlying SMD between every treatment t2,t3,,tKi versus the baseline treatment t1 is

δi,1k=θikθi1sdipooled,(k>1), (2)

with sdipooled=Σk=1Kiniksdik2nikKi the between-arm standard deviation. Under the random-effects assumption, the underlying relative effects are assumed to follow a multivariate normal distribution

δiNKi1(μ,Σ), (3)

where μ=(μ12,,μ1Ki) is the vector of the summary relative effects and Σ the between-study variance–covariance matrix with entries τ2 (the common heterogeneity variance across comparisons) in the diagonal and τ22 in the off-diagonal. Finally, the transitivity assumption for every pair of treatments j, lTa (j, l > 1) implies that µjl = µ1lµ1j.

Non-informative prior distributions are usually employed for every µ1j and τ, such as N(0, 10000) and the half-normal HN(1), respectively.

Apart from the strong assumption about the similarity of the two subgroups, a further problem in the present setting is that the final results for the joint population P will be dominated by the dense network (P2) rather than from the target subgroup P1. The same would apply if the results of the dense network were used directly as prior information for P1. To avoid this issue, in the proposed approach, we extrapolate the results from P2 to P1, acknowledging differences between the two subgroups, before forming informative prior distributions for P1.

3.2.2. Using the external subgroup P2 to construct informative priors for the target subgroup P1

This is a two-stage approach where at the first stage, we extrapolate the results from the dense network of subgroup P2 to P1, and at the second stage, we use predictions from this extrapolation to form prior distributions for the analysis of P1.

First stage: Extrapolating external results to the target population subgroup

In contrast to the naïve synthesis, the main assumption here is that the two subgroups have some underlying differences in the treatment effects that can be considered as a different location of their outcome distributions. To incorporate some uncertainty about this assumption, in Equation (1), we introduce a scale parameter wi ∈ (0, 1] that inflates the variance of each yik in subgroup P2 for each i = n1 + 1, n1 + 2,…, N (Baker & Jackson, 2013). This modifies Equation (1) into

yikN(θik,sdik2wi). (4)

Equation (4) implies that the parameters wi only affect the variance of the the mean scores yik and not the true study-specific SMDs that are still the same as in Equation (2). Assuming that β1j=μ1jP2μ1jP1 for each {1, j}, jTc, the location parameters β that aim to shift the original distribution of the SMDs in P2 toward the distribution of P1 are then added in Equation (3):

δiNKi1(μP2β,Σ), (5)

with β=(β12,,β1Ki) being the vector of the “true” differences between the comparison-specific SMDs of the two sub-groups. The extrapolated means μ=μP2β for population P2 in Equation (5) are then considered similar to those expected for population P1. Note that Equation (5) could be equivalently written as δiNKi1(μP2+β,Σ), if we had assumed that β1j=μ1jP1μ1jP2.

By fitting this modified NMA model, we obtain the extrapolated SMD estimates μ^jl for every pair of treatments j, lTc. Subsequently, the predictive distributions (Higgins et al., 2009) of μjlnew, namely, N(μjl,(τP2)2) are used at the second stage as prior distributions for P1. Note that the parameter (τP2)2 represents the heterogeneity across the studies in P2.

Second stage: NMA of the target subgroup using informative prior distributions for relative effects

Here, the model of Section 3.2.1 is used for i = 1, …, n1. The key difference is that for the µ1l s, we use as prior distribution μ1lP1N(μ^1lnew,var(μ^1lnew)) where μ^1lnew and var(μ^1lnew) are the posterior mean and variance of μ1lnew , respectively. The use of informative priors is expected to improve the precision in the final NMA estimates without dominating the analysis. This is because the priors are constructed at the first stage by extrapolating the results of the dense network to those expected for the target sparse network.

3.2.3. Informing the location and the scale parameters

Constructing prior distributions for βs using the data

To obtain an estimate of the difference in the outcome distributions between the two subgroups, we first use the estimates u1jP1 and u1jP2 obtained from separate pairwise meta-analyses for each population and each pair of treatments {1, j}, jTc. To allow the estimation of heterogeneity for every comparison in P1, here we assume a common comparison-specific heterogeneity across the two subgroups (σ1jP1)2=(σ1jP2)2=σ1j2. Using the difference of these estimates d1j=u1jP2u1jP1, we construct prior distributions for the parameters β1j, namely,

β1jN(d1j,var(d1j)). (6)

For treatment comparisons not evaluated in P2, we use a non-informative normal prior N(0, 10000).

Constructing prior distributions for βs using expert opinion

Elicitation of expert opinion can be undertaken with several methods such as face-to-face interviews, software tools, or questionnaires. The pooled change score of each treatment jTc in P2 can be used as a reference for the experts who need to provide an “estimate” for the change score in P1. Suppose that xhj is the change score and sdhj the standard deviation provided by the expert h = 1, 2, …, H for treatment jTc. Let also γh denote the overall confidence of each expert to provided opinions. To combine the estimates provided by different experts, we can then use the following meta-analysis model:

xhjN(chj,sdhj2γh), (7)
dhjN(ξjP1,σ2), (8)

with dhj=chjmed(sdipoled), ∀i, h, j the standardized values of chj. Given that the pooled standard deviation cannot be elicited by experts, the median across the studies in the data is used. Note that standardization is not necessary when all studies in the data use the same scale. Then, the pooled SMDs can be obtained as u1jP1=ξjP1ξ1P1 and the priors for the parameter β1j is constructed as in Equation (6).

Prior distributions for the scale parameter

Dividing the variances in Equation (4) by wi can be seen as a special case of the power prior method where a specific power is chosen for the likelihood of each study (Chen & Ibrahim, 2000; Efthimiou et al., 2017; Ibrahim et al., 2015). Values of wi close to 0 reflect a serious downweight of the evidence coming from the external subgroup P2, whereas values close to 1 imply a mild downweight. Typical choices of prior distributions for the scale parameter can be the beta or the uniform distribution. The parameters of these distributions should be chosen to reflect the prior beliefs regarding the specific characteristic according to which the downweight takes place. For example, a Beta(3, 3) can be used for cases where a moderate downweight needs to apply for the external information. This distribution places its mass around 0.5 with 95% range equal to [0.15, 0.84] thus reflecting more uncertainty about the magnitude of our downweighting. Left-skewed beta priors (e.g., Beta(1, 6)) can be used for weight values close to 0, while on the other hand, right-skewed beta priors (e.g., Beta(6, 1)) can be used for weight values close to 1. The previous downweighting schemes can be achieved with other types of prior distributions. For example, a Unif (0.4, 0.6), a Unif (0.1, 0.3) or a Unif (0.8, 1) can be used for moderate, serious, or mild downweight, respectively. Finally, it is generally not recommended to use priors for the scale parameter that can allow for values larger than 1. This could further increase the impact of external data and enable their dominance in the final NMA estimates.

4. Application

4.1. Implementation

We consider throughout CA being the target population subgroup P1 and GP the external subgroup P2. We applied in total nine models six of which refer to the combinations of the different prior distributions for the location (β1j) and scale parameters (wi).

  1. Naïve synthesis model for CA and GP with non-informative priors.

  2. NMA for CA with informative priors from GP.
    1. data-based prior distributions for the location parameters with
      • (i)
        no downweight for any GP study (wi = 1) using only treatments in Tc (No DW)
      • (ii)
        moderate downweight (i.e., wi ∼ Beta(3, 3)) to high risk of bias (RoB) GP studies using only treatments in Tc (RoB DW). The risk of bias was assessed according to the Cochrane’s RoB tool (Sterne et al., 2019) and in total 16 studies were rated at high RoB for GP.
      • (iii)
        no downweight for any GP study evaluating treatments in Tc and moderate downweight (i.e., wi ∼ Beta(3, 3)) for those evaluating at least one treatment in TaTc (NCT DW). This means that here we use the full network of GP with the 34 treatments and we downweight all studies that contain at least one treatment that has never been evaluated in CA (107 studies).
    2. prior distributions based on expert opinion for the location parameter combined the three above (i–iii) possibilities for the scale parameters.
  3. NMA for CA with non-informative priors (i.e., “standard” NMA).

  4. Pairwise meta-analysis for CA with non-informative priors.

All the analyses were conducted using the rjags (22) package through the R statistical software (R version 4.0.3, 2020-10-10) (23). For all models, we ran two chains in parallel, performed 50,000 iterations and discarded the first 10,000 samples of each chain. We checked the chains convergence using the Gelman–Rubin criterion (Brooks & Gelman, 1998) with a value below 1.1 to indicate failure of convergence. We also visually checked the Markov chains using trace plots to inspect the sampling behavior and assess mixing across chains and convergence. The R code and the data used in this article are available in the Appendix and can also be accessed at: https://github.com/TEvrenoglou/codes_sharing_information

4.2. Elicitation of expert’s opinion

We prepared a questionnaire and circulated it to psychiatrists with experience in treating schizophrenia for CA and GP (available in Appendix 2). We asked each expert to provide an estimate of the “expected” mean score reduction in the PANSS scale from baseline to endpoint for each drug for CA given the pooled mean score reduction obtained from the data for GP. We further asked them to provide a measure of uncertainty around the given expected mean scores (a) in the form of standard deviation and (b) in a 10-point scale of their confidence. We obtained in total 22 responses which we averaged for each drug using the meta-analysis model given in Equations (7)(8). We used a non-informative N(0, 10000) and a weakly informative HN(1) prior for the means ξjP1, jTc and the parameter σ, respectively.

4.3. Results

The estimates for the basic comparisons of the CA network are depicted in Figure 2 and for all relative comparisons in Online Appendix 1, Tables 10–17. As expected, for most comparisons, the use of informative priors leads to a substantial improvement in the precision of the relative effects. The naïve synthesis model provides the most precise results but relies on a strong assumption that is likely to be implausible. Interestingly, CA results from the standard NMA model (with non-informative priors) tend to be less precise than the respective direct estimates. This occurs often in sparse networks because indirect comparisons as well as heterogeneity are estimated with large uncertainty. Overall, standard NMA most often does not give any insight for comparisons without direct evidence as it yields very large credible intervals. For some of these comparisons (e.g., Haloperidol vs. Placebo), the NMA models with informative priors result in more conclusive relative effect estimates. In terms of point estimates, results appear generally robust across the different models with only few exceptions, such as the extreme cases of Fluphenazine and Trifluoperazine. Those two are very old drugs for which the evidence is outdated and sparse in both GP and CA networks. The latter implies that the informative prior can dominate in the analysis and yield to estimate closer to the GP. Overall, point estimates appeared to be more robust for the basic comparisons for which direct evidence is available in the network of CA. Finally, no convergence issues were identified across the chains. The corresponding trace plots are available in Appendix 1, Figures 8–13.

Figure 2. Results from the analysis of the network defined for the population of CA.

Figure 2

The outcome of interest is the reduction of the overall schizophrenia symptoms and the effect measure is the standardized mean difference. DW, downweight; RoB, risk of bias; NCT, non-common treatment.

Only small differences can be observed among the six models with informative priors, suggesting that the approach of informing the location and the scale parameters does not affect materially the final predictions; this is possibly due to the huge amount of data in the GP network. For a few comparisons, the data-based approach for β provides to some degree different relative effects than those obtained from the expert opinion approach, implying that the experience from clinical practice does not fully agree with the available data. Those comparisons might need to be prioritized for the design of new trials. Of course, when the full GP network was used and only studies comparing treatments in TaTc were downweighted some extra precision was gained. Such an approach might be more useful when the external network is not as dense as in the present dataset.

Assessment of inconsistency was performed using the node-splitting method (Dias et al., 2010). Inference regarding the presence of inconsistency was based on Bayesian p-values, and on the visual inspection of the posterior densities of direct and indirect evidence, Bayesian p-values were calculated as 2 × min{P, 1 − P} where P is the probability that the difference between direct and indirect comparisons is positive. Additionally, the different antipsychotics and placebo were ranked across all models using the surface under the cumulative ranking curve (SUCRA) (Salanti et al., 2011) method. Overall, the respective posterior densities of the direct and indirect estimates appeared to be quite similar across the different com-parisons and models and no p-value indicated lack of consistency (Online Appendix 1, Figures 2–6, Tables 2–9). In terms of ranking, the results appear to be generally robust across models including the naïve synthesis and the standard NMA model. Clozapine was always ranked as the most effective antipsychotic. Finally, Olanzapine, Risperidone, and Molindone were ranked in second, third, and fourth positions, respectively, in most of the fitted models (Appendix 1, Figure 7).

5. Discussion

In this paper, we present a Bayesian framework for NMA of sparse networks aiming to improve the performance of the relative effect estimates in terms of precision and reliability. To this end, we borrow information using external data from a dense network that targets a different subgroup of the population. We model the differences between the two populations with a two-stage approach. At the first stage, we analyze only the dense network, and we extrapolate its results to the sparse network though the incorporation of a location parameter in the distribution of the summary relative effects. Then, we use the extrapolated results to construct informative prior distributions for the target population subgroup. Similar approaches have been used previously at the level of primary studies (Dunne et al., 2011; Dunoyer, 2011; Hampson et al., 2014; Rhodes et al., 2015). We further introduce a scale parameter that inflates the variance of the studies in the dense network to reflect the potential uncertainty regarding the assumed relationship between the two population subgroups (Hlavin et al., 2016; Wadsworth et al., 2018).

We proposed two different approaches for informing the location parameters; one based on the data and one based on expert opinion. The former is easier to implement but the data are used in both stages: once to form the priors and once to perform the NMA. This is a drawback of the data-based approach as it might introduce some correlation that is ignored. In addition, this approach requires that a certain number of direct comparisons are available in both networks; otherwise, the use of non-informative priors would be necessary for several βs at the first stage. The prior elicitation approach is general as it does not rely on the amount of direct evidence. However, results from such an approach are always subjective to some degree and handling disagreements between different experts might be challenging. Here, we requested the necessary information from the psychiatrists through a questionnaire and some of them raised concerns that it was difficult for them to make a suggestion. Other ways to elicit information from experts exist. For example, Turner et al. used a three-stage elicitation approach where the experts were interviewed through 1:1 meetings or by telephone (Turner et al., 2022). Alternatively, information on these parameters may be obtained by using, when available, external data, such as from observational studies, electronic health records, or other networks pertaining to populations with similar characteristics.

In our motivating example, the two approaches for constructing priors for the location parameters were generally in agreement except the comparison between Loxapine and Placebo where, in contrast to the data-based approach, the analyses based on the expert opinion may not support the presence of an important effect. In all other comparisons, despite some differences in the magnitude of the estimated summary effects, the two approaches yielded similar conclusions. Regarding the scale parameters, we applied no downweight for trials considered more relevant to our target population and moderate downweight for studies that were considered less relevant or reliable. The results remained robust across the different downweighing schemes that we used. However, the use of Beta(3, 3) prior for the wis can be considered a conservative choice as it is centered at 0.5 and provides similar mass to both the left and right tails allowing for both strong and mild downweight, for example, wi ≈ 0 or wi ≈ 1, respectively. Sensitivity analysis could further consider left-skewed Beta distributions (e.g., Beta(1, 8)) that would assign weights mostly close to 0 and right-skewed Betas (e.g., Beta(8, 1)) that would assign weights mostly close to 1. Compared to the standard NMA with non-informative priors, our two-stage approach provided substantially more precise estimates. The lack of robustness of the standard NMA model for this dataset is also deduced from the fact that almost always it yielded less precise results than pairwise meta-analysis. As expected, the most precise results were obtained from the naïve synthesis model. Nevertheless, this model does not account for any differences across the two population subgroups.

In the presence of few studies, estimation of the heterogeneity is challenging (Turner et al., 2019). In our analysis, we used a weakly-informative half normal prior distribution for the heterogeneity parameter (Röver et al., 2021). However, a sensitivity analysis on the prior distribution for heterogeneity could be considered as in the absence of sufficient data, the estimation of heterogeneity might be affected by the choice of the prior. For example, using informative priors for heterogeneity could possibly further increase the precision of our approach. Usually, these are based on empirical distributions that depend on the type of outcome and treatment comparisons (Rhodes et al., 2015; Turner et al., 2012, 2015). Alternatively, more sophisticated options that allow the incorporation of external data in the estimation of the heterogeneity parameter may be applied (Turner et al., 2019).

A limitation of our work is that we did not explore under which conditions the introduced location parameters reduce or increase the bias of the summary estimates. This would require to evaluate the proposed Bayesian framework through a simulation study. Nevertheless, data generation in this setting might be challenging as it would require construction of the differences between the two populations based on clinically meaningful baseline characteristics. This merits a lot of further consideration and can form a follow-up article. In addition, such a study could only evaluate the data-based approach for the priors as it would be impossible to obtain expert opinion for many simulated datasets.

Our proposed method was built upon a Bayesian framework as this offers a straightforward and intuitive way to incorporate external information in the analysis through informative prior distributions. Alternatively, a two-stage frequentist framework could be used where the prior information for the location parameters could be incorporated through pseudodata as in Rhodes et al. (Rhodes et al., 2016). However, this approach uses meta-regression to fit an NMA model, and therefore, it is directly applicable only to NMA of two-arm studies (Rhodes et al., 2016). A further possibility may be to model the differences in the two populations through inconsistency parameters using the random inconsistency design-by-treatment interaction model proposed by Jackson et al. (2014). Such an approach would assume different inconsistency parameters for populations P1 and P2 but the same for studies within each population. A potential future simulation study could investigate similarities and differences in the estimation between the three approaches.

Overall, the proposed framework offers a reliable approach for assisting the analysis of sparse networks. Of course, its application requires close collaboration between statisticians and experienced clinicians to ensure that sharing of information between different population subgroups is clinically meaningful. Finally, performing NMA in cases of sparse networks will always be a challenging procedure and the best approach will likely be context-dependent.

Supplementary Material

Supplementary Material

Acknowledgments

Theodoros Evrenoglou received funding by the French National Research Agency under the project ANR-22-CE36-0013-01. Rebecca Turner was supported by the UK Medical Research Council (Programme MC_UU_00004/06). We are thankful to the following psychiatrists for providing expert opinion: Arango Celso, Carpenter Will, Crespo Facorro Benedicto, Davidson Michael, Dazzan Paola, Gerhard Gruender, Hasan Alkomiet, Heres Stephan, Hui Wu, Jauhar Sameer, Juckel Georg, McGorry Pat, Murray Robin, Musil Richard, Peter Natalie, Samara Myrto, Stip Emmanuel, Takeuchi Hiroyoshi, Weiser Mark, and Zhu Yikang.

Funding information

French National Research Agency, Grant/Award Number: ANR-22-CE36-0013-01; UK Medical Research Council, Grant/Award Number: MC_UU_00004/06

Footnotes

Conflict of Interest Statement

The authors declare no conflicts of interest.

Data Availability Statement

All data are available in the Supplementary Material and on the following public GitHub repository: https://github.com/TEvrenoglou/codes_sharing_information

References

  1. Baker RD, Jackson D. Meta-analysis inside and outside particle physics: Two traditions that should converge? Research Synthesis Methods. 2013;4(2):109–124. doi: 10.1002/jrsm.1065. [DOI] [PubMed] [Google Scholar]
  2. Brooks SP, Gelman A. General methods for monitoring convergence of iterative simulations. Journal of Computational and Graphical Statistics. 1998;7(4):434–455. [Google Scholar]
  3. Chaimani A, Caldwell DM, Li T, Higgins JP, Salanti G. In: Cochrane handbook for systematic reviews of interventions. Higgins JPT, Thomas J, Chandler J, Cumpston M, Li T, Page MJ, Welch VA, editors. John Wiley & Sons, Ltd; 2019. Undertaking network meta-analyses; pp. 285–320. [DOI] [Google Scholar]
  4. Chen MH, Ibrahim JG. Power prior distributions for regression models. Statistical Science. 2000;15(1):46–60. doi: 10.1214/ss/1009212673. [DOI] [Google Scholar]
  5. Dias S, Welton NJ, Caldwell DM, Ades AE. Checking consistency in mixed treatment comparison meta-analysis. Statistics in Medicine. 2010;29(7–8):932–944. doi: 10.1002/sim.3767. [DOI] [PubMed] [Google Scholar]
  6. Dunne J, Rodriguez WJ, Murphy MD, Beasley BN, Burckart GJ, Filie JD, Lewis LL, Sachs HC, Sheridan PH, Starke P, Yao LP. Extrapolation of adult data and other data in pediatric drug-development programs. Pediatrics. 2011;128(5):e1242–e1249. doi: 10.1542/peds.2010-3487. [DOI] [PubMed] [Google Scholar]
  7. Dunoyer M. Accelerating access to treatments for rare diseases. Orphanet Journal of Rare Diseases. 2011;10(7):475–476. doi: 10.1038/nrd3493. [DOI] [PubMed] [Google Scholar]
  8. Efthimiou O, Mavridis D, Debray TPA, Samara M, Belger M, Siontis GCM, Leucht S, Salanti G. Combining randomized and non-randomized evidence in network meta-analysis. Statistics in Medicine. 2017;36(8):1210–1226. doi: 10.1002/sim.7223. [DOI] [PubMed] [Google Scholar]
  9. Hampson LV, Whitehead J, Eleftheriou D, Brogan P. Bayesian methods for the design and interpretation of clinical trials in very rare diseases. Statistics in Medicine. 2014;33(24):4186–4201. doi: 10.1002/sim.6225. [DOI] [PMC free article] [PubMed] [Google Scholar]
  10. Higgins JPT, Thompson SG, Spiegelhalter DJ. A re-evaluation of random-effects meta-analysis. Journal of the Royal Statistical Society: Series A (Statistics in Society) 2009;172:137–159. doi: 10.1111/j.1467-985X.2008.00552.x. [DOI] [PMC free article] [PubMed] [Google Scholar]
  11. Hlavin G, Koenig F, Male C, Posch M, Bauer P. Evidence, eminence and extrapolation. Statistics in Medicine. 2016;35(13):2117–2132. doi: 10.1002/sim.6865. [DOI] [PMC free article] [PubMed] [Google Scholar]
  12. Ibrahim JG, Chen MH, Gwon Y, Chen F. The power prior: Theory and applications. Statistics in Medicine. 2015;34(28):3724–3749. doi: 10.1002/sim.6728. [DOI] [PMC free article] [PubMed] [Google Scholar]
  13. Jackson D, Barrett JK, Rice S, White IR, Higgins JPT. A design-by-treatment interaction model for network meta-analysis with random inconsistency effects. Statistics in Medicine. 2014;33(21):3639–3654. doi: 10.1002/sim.6188. [DOI] [PMC free article] [PubMed] [Google Scholar]
  14. Law M, Alam N, Veroniki AA, Yu Y, Jackson D. Two new approaches for the visualisation of models for network meta-analysis. BMC Medical Research Methodology [Electronic Resource] 2019;19(1):61. doi: 10.1186/s12874-019-0689-9. [DOI] [PMC free article] [PubMed] [Google Scholar]
  15. Leucht S, Chaimani A, Krause M, Schneider-Thoma J, Wang D, Dong S, Samara M, Peter N, Huhn M, Priller J, Davis JM. The response of schizophrenia subgroups to different antipsychotic drugs: A systematic review and meta-analysis. Lancet Psychiatry. 2022;9:884–893. doi: 10.1016/S2215-0366(22)00304-2. [DOI] [PubMed] [Google Scholar]
  16. Nikolakopoulou A, Mavridis D, Furukawa TA, Cipriani A, Tricco AC, Straus SE, Siontis GCM, Egger M, Salanti G. Living network meta-analysis compared with pairwise meta-analysis in comparative effectiveness research: Empirical study. British Medical Journal. 2018;360:k585. doi: 10.1136/bmj.k585. [DOI] [PMC free article] [PubMed] [Google Scholar]
  17. Parmar MK, Sydes MR, Cafferty FH, Choodari-Oskooei B, Langley RE, Brown L, Phillips PP, Spears MR, Rowley S, Kaplan R, James ND, et al. Testing many treatments within a single protocol over 10 years at MRC Clinical Trials Unit at UCL: Multi-arm, multi-stage platform, umbrella and basket protocols. Clinical Trials. 2017;14(5):451–461. doi: 10.1177/1740774517725697. [DOI] [PMC free article] [PubMed] [Google Scholar]
  18. Phillippo DM, Dias S, Ades AE, Welton NJ. Assessing the performance of population adjustment methods for anchored indirect comparisons: A simulation study. Statistics in Medicine. 2020;39(30):4885–4911. doi: 10.1002/sim.8759. [DOI] [PMC free article] [PubMed] [Google Scholar]
  19. Plummer M. rjags: Bayesian Graphical Models using MCMC. 2015. https://CRAN.R-project.org/package=rjags Published online 2015.
  20. R Core Team. R: A Language and Environment for Statistical Computing. 2016. Published online 2016. https://www.R-project.org/
  21. Renfro LA, Sargent DJ. Statistical controversies in clinical research: Basket trials, umbrella trials, and other master protocols: a review and examples. Annals of Oncology. 2017;28(1):34–43. doi: 10.1093/annonc/mdw413. [DOI] [PMC free article] [PubMed] [Google Scholar]
  22. Rhodes KM, Turner RM, Higgins JPT. Predictive distributions were developed for the extent of heterogeneity in meta-analyses of continuous outcome data. Journal of Clinical Epidemiology. 2015;68(1):52–60. doi: 10.1016/j.jclinepi.2014.08.012. [DOI] [PMC free article] [PubMed] [Google Scholar]
  23. Rhodes KM, Turner RM, White IR, Jackson D, Spiegelhalter DJ, Higgins JPT. Implementing informative priors for heterogeneity in meta-analysis using meta-regression and pseudo data. Statistics in Medicine. 2016;35(29):5495–5511. doi: 10.1002/sim.7090. [DOI] [PMC free article] [PubMed] [Google Scholar]
  24. Röver C, Bender R, Dias S, Schmid CH, Schmidli H, Sturtz S, Weber S, Friede T. On weakly informative prior distributions for the heterogeneity parameter in Bayesian random-effects meta-analysis. Research Synthesis Methods. 2021;12(4):448–474. doi: 10.1002/jrsm.1475. [DOI] [PubMed] [Google Scholar]
  25. Röver C, Wandel S, Friede T. Model averaging for robust extrapolation in evidence synthesis. Statistics in Medicine. 2019;38(4):674–694. doi: 10.1002/sim.7991. [DOI] [PubMed] [Google Scholar]
  26. Salanti G, Ades AE, Ioannidis JPA. Graphical methods and numerical summaries for presenting results from multiple-treatment meta-analysis: An overview and tutorial. Journal of Clinical Epidemiology. 2011;64(2):163–171. doi: 10.1016/j.jclinepi.2010.03.016. [DOI] [PubMed] [Google Scholar]
  27. Salanti G, Higgins JPT, Ades AE, Ioannidis JPA. Evaluation of networks of randomized trials. Statistical Methods in Medical Research. 2008;17(3):279–301. doi: 10.1177/0962280207080643. [DOI] [PubMed] [Google Scholar]
  28. Sterne JAC, Savović J, Page MJ, Blencowe NS, Boutron I, Cates CJ, Cheng HY, Corbett MS, Eldridge SM, Emberson JR, Hernán MA, et al. RoB 2: A revised tool for assessing risk of bias in randomised trials. Bmj. 2019;366:4898. doi: 10.1136/bmj.l4898. [DOI] [PubMed] [Google Scholar]
  29. Turner RM, Davey J, Clarke MJ, Thompson SG, Higgins JP. Predicting the extent of heterogeneity in meta-analysis, using empirical data from the Cochrane Database of Systematic Reviews. International Journal of Epidemiology. 2012;41(3):818–827. doi: 10.1093/ije/dys041. [DOI] [PMC free article] [PubMed] [Google Scholar]
  30. Turner RM, Domínguez-Islas CP, Jackson D, Rhodes KM, White IR. Incorporating external evidence on between-trial heterogeneity in network meta-analysis. Statistics in Medicine. 2019;38(8):1321–1335. doi: 10.1002/sim.8044. [DOI] [PMC free article] [PubMed] [Google Scholar]
  31. Turner RM, Jackson D, Wei Y, Thompson SG, Higgins JPT. Predictive distributions for between-study heterogeneity and simple methods for their application in Bayesian meta-analysis. Statistics in Medicine. 2015;34(6):984–998. doi: 10.1002/sim.6381. [DOI] [PMC free article] [PubMed] [Google Scholar]
  32. Turner RM, Turkova A, Moore CL, Bamford A, Archary M, Barlow-Mosha LN, Cotton MF, Cressey TR, Kaudha E, Lugemwa A, Lyall H, et al. Borrowing information across patient subgroups in clinical trials, with application to a paediatric trial. BMC Medical Research Methodology. 2022;22(1):49. doi: 10.1186/s12874-022-01539-3. [DOI] [PMC free article] [PubMed] [Google Scholar]
  33. U.S. Department of Health and Human Services Food and Drug Administration. Leveraging existing clinical data for extrapolation to pediatric uses of medical devices—Guidance for industry and Food and Drug Administration staff draft guidance. 2016 [Google Scholar]
  34. Veroniki AA, Tsokani S, Zevgiti S, Pagkalidou I, Kontouli KM, Ambarcioglu P, Pandis N, Lunny C, Nikolakopoulou A, Theodoros P, Chaimani A, et al. Do reporting guidelines have an impact? Empirical assessment of changes in reporting before and after the PRISMA extension statement for network meta-analysis. Systematic Reviews. 2021;10(1):246. doi: 10.1186/s13643-021-01780-9. [DOI] [PMC free article] [PubMed] [Google Scholar]
  35. Wadsworth I, Hampson LV, Jaki T. Extrapolation of efficacy and other data to support the development of new medicines for children: A systematic review of methods. Statistical Methods in Medical Research. 2018;27(2):398–413. doi: 10.1177/0962280216631359. [DOI] [PubMed] [Google Scholar]
  36. Yoon JH, Dias S, Hahn S. A method for assessing robustness of the results of a star-shaped network meta-analysis under the unidentifiable consistency assumption. BMC Medical Research Methodology [Electronic Resource] 2021;21(1):113. doi: 10.1186/s12874-021-01290-1. [DOI] [PMC free article] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Supplementary Material

Data Availability Statement

All data are available in the Supplementary Material and on the following public GitHub repository: https://github.com/TEvrenoglou/codes_sharing_information

RESOURCES