Abstract
We recently suggested new statistical metrics for routine reporting in random-effects meta-analyses to convey evidence strength for scientifically meaningful effects under effect heterogeneity. First, given a chosen threshold of meaningful effect size, we suggested reporting the estimated proportion of true effect sizes above this threshold. Second, we suggested reporting the proportion of effect sizes below a second, possibly symmetric, threshold in the opposite direction from the estimated mean. Our previous methods applied when the true effects are approximately normal, when the number of studies is relatively large, and when the proportion is between approximately 0.15 and 0.85. Here, we additionally describe robust methods for point estimation and inference that perform well under considerably more general conditions, as we validate in an extensive simulation study. The methods are implemented in the R package MetaUtility (function prop_stronger). We describe application of the robust methods to conducting sensitivity analyses for unmeasured confounding in meta-analyses.
Keywords: meta-analysis, effect sizes, heterogeneity, nonparametric, bootstrapping, confounding
Introduction
We recently suggested new statistical metrics for routine reporting in random-effects meta-analyses to convey evidence strength for scientifically meaningful effects under effect heterogeneity1. First, given a chosen threshold of meaningful effect size (q), we suggested reporting the estimated proportion of true effect sizes above this threshold . Second, we suggested reporting the proportion of effect sizes below a second, possibly symmetric, threshold in the opposite direction from the estimated mean. These metrics can help identify if: (1) there are few effects of scientifically meaningful size despite a “statistically significant” pooled point estimate; (2) there are some large effects despite an apparently null point estimate; or (3) strong effects in the direction opposite of the pooled estimate also regularly occur. Additionally, these metrics can sometimes adjudicate apparent “conflicts” between meta-analyses2 and can convey evidence strength in multisite replication projects3.
We had proposed parametric estimation methods with asymptotic inference based on the delta method1; those methods applied when the true effects are approximately normal, when the number of studies is relatively large, and when is between approximately 0.15 and 0.85. Here, we additionally describe robust methods for point estimation and inference that perform well under more general conditions.
Methods
Let θi, , and respectively denote the true effect size, the point estimate, and the estimated standard error of the ith study. If the parameter θi for each study were known, then a simple nonparametric estimate would simply be the sample proportion of θi greater than q. Given that the θi are in fact unknown, it would seem intuitive to instead compute the sample proportion of point estimates greater than q, but this approach is incorrect because dispersion in the reflects not only true effect heterogeneity, but also statistical error due to finite sample sizes in the meta-analyzed studies. Thus, the are overdispersed compared to the θi and would not themselves yield an unbiased estimate .
We therefore suggest computing the sample proportion using recently proposed “calibrated” estimates that have been appropriately shrunk to correct the overdispersion4. Let and represent classical Dersimonian-Laird5 meta-analytic estimates of the mean and variance of the true effects; these moment-based estimates do not require parametric assumptions. Wang et al. (2019)4 defined the calibrated point estimate for the ith study as:
and showed that , as desired. Intuitively, the calibrated estimate shrinks the point estimate toward the estimated meta-analytic mean with a degree of shrinkage that is inversely proportional to the study’s precision: relatively imprecise estimates (i.e., those with large ) receive strong shrinkage toward , while relatively precise estimates receive less shrinkage and remain closer to their original values. Wang et al. (2019)4 demonstrated that the calibrated estimates can be used to construct approximately unbiased prediction intervals for small meta-analyses and for non-normal true effect distributions. For our purposes, we propose estimating the proportion of scientifically meaningful effect sizes as the sample proportion of calibrated estimates above q, i.e., . For inference, one can bootstrap pairs of by drawing with replacement from the original sample and estimating in turn and , for each study, and finally . A bias-corrected and accelerated (BCa) confidence interval6;7 can then be constructed from the bootstrapped values of . (Naturally, analogous methods can be used to estimate the proportion of effects below another threshold.)
We also considered a simulation-based nonparametric method (here termed the “sign test method”) that was originally designed to estimate a given percentile of interest (e.g., the median) of a distribution of effect sizes and to construct a confidence interval8. The method involves first conducting nonparametric hypothesis tests that are similar to sign tests for each of many possible values for the percentile of interest, then inverting the rejection region to form a confidence interval. This method can be straightforwardly repurposed to provide an estimate and confidence interval for the proportion of effects above a threshold, P>q, as we show in the eAppendix.
We assessed all methods’ performance in an extensive simulation study of 480 scenarios with a range of true effect distributions (including normal, highly skewed, bimodal, and heavy-tailed distributions), meta-analyses with 5 to 50 studies of varying size, varying heterogeneity, and true proportions from 0.05 to 0.50. We included both realistic and extreme distributions of effect sizes in order to establish the boundary conditions under which the statistical methods would perform well. Some of the more extreme distributions may be unlikely to occur in practice, and conducting an aggregate meta-analysis when effects are clearly multimodal may be scientifically ill-advised in the first place. Details on the simulation study design appear in the eAppendix.
Results and conclusions
Based on the simulation results (eAppendix), we recommend reporting and inference only for meta-analyses with at least 10 studies. In such meta-analyses, we recommend by default estimating using the calibrated estimates; this method was the least biased for all distributions, though its root mean square error (RMSE) was sometimes higher than that of other methods. For inference, even when the effects appear normal, we recommend by default constructing the confidence interval by applying the bias-corrected and accelerated bootstrap to the calibrated estimates (“BCa-calibrated”); this method achieved nominal coverage in almost all scenarios and always achieved at least 90% coverage. The sign test method sometimes performed poorly when heterogeneity was low to moderate, and it offered few advantages over the BCa-calibrated method. We therefore do not recommend its use in practice to estimate or conduct inference for P>q.
The BCa-calibrated method did sometimes lose considerable precision compared to the parametric method in certain scenarios in which the latter achieved approximately nominal coverage (e.g., see eFigure 7), so for large meta-analyses with apparently normal effects and estimating a proportion close to 0.50, one might reasonably choose to substitute the parametric interval for the default BCa-calibrated interval. For example, in relatively small meta-analyses, estimating the amount of heterogeneity can be inherently imprecise9. This uncertainty propagates to the confidence interval for P>q and, in small meta-analyses, may result in confidence intervals that span most or all of the possible range [0, 1]. Reporting confidence intervals in these settings may nevertheless be informative: a very wide confidence interval may instill appropriate circumspection about what can be learned regarding the distribution of true effects in a small meta-analysis, even if itself may have a narrow confidence interval (for example, see the applied example in the eAppendix). A wide confidence interval may further suggest the value of performing a larger meta-analysis when more literature becomes available. As an additional limitation, the BCa-calibrated interval may sometimes fail to converge for small meta-analyses. (When the BCa-calibrated interval fails to converge, it may seem attractive to construct a simpler bootstrapped confidence interval using percentiles of the bootstrapped calibrated estimates. However, we recommend against this method; additional simulation results suggested that it performed quite poorly.)
The recommended methods are implemented in the function prop_stronger in the R package MetaUtility as of version 2.0.0. We illustrate this software and approach with an applied example in the eAppendix, where we also discuss extensions to sensitivity analysis for unmeasured confounding.
Supplementary Material
Acknowledgments
Source of funding: MBM and TVW were supported by NIH grant R01 CA222147. The funders had no role in the design, conduct, or reporting of this research.
Footnotes
Conflicts of interest: The authors declare that they have no conflicts of interest.
References
- [1].Mathur Maya B and VanderWeele Tyler J. New metrics for meta-analyses of heterogeneous effects. Statistics in Medicine, 2018. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [2].Mathur Maya B and VanderWeele Tyler J. Finding common ground in meta-analysis “wars” on violent video games. Perspectives on Psychological Science, page 1745691619850104, 2019. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [3].Mathur Maya B and VanderWeele Tyler J. New statistical metrics for multisite replication projects. Under review. Preprint retrieved from https://osf.io/w89s5/. [Google Scholar]
- [4].Wang Chia-Chun and Lee Wen-Chung. A simple method to estimate prediction intervals and predictive distributions: Summarizing meta-analyses beyond means and confidence intervals. Research Synthesis Methods, 10(2):255–266, 2019. [DOI] [PubMed] [Google Scholar]
- [5].DerSimonian Rebecca and Laird Nan. Meta-analysis in clinical trials. Controlled Clinical Trials, 7(3):177–188, 1986. [DOI] [PubMed] [Google Scholar]
- [6].Efron Bradley. Better bootstrap confidence intervals. Journal of the American Statistical Association, 82(397):171–185, 1987. [Google Scholar]
- [7].Carpenter James and Bithell John. Bootstrap confidence intervals: When, which, what? A practical guide for medical statisticians. Statistics in Medicine, 19(9):1141–1164, 2000. [DOI] [PubMed] [Google Scholar]
- [8].Wang Rui, Tian Lu, Cai Tianxi, and Wei LJ. Nonparametric inference procedure for percentiles of the random effects distribution in meta-analysis. The Annals of Applied Statistics, 4(1):520, 2010. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [9].Veroniki Areti Angeliki, Jackson Dan, Viechtbauer Wolfgang, Bender Ralf, Bowden Jack, Knapp Guido, Kuss Oliver, Higgins Julian, Langan Dean, and Salanti Georgia. Methods to estimate the between-study variance and its uncertainty in meta-analysis. Research Synthesis Methods, 2015. [DOI] [PMC free article] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.