Skip to main content
F1000Research logoLink to F1000Research
. 2021 Jun 8;10:455. [Version 1] doi: 10.12688/f1000research.53034.1

Development of a checklist to detect errors in meta-analyses in systematic reviews of interventions: study protocol

Raju Kanukula 1, Matthew Page 1, Kerry Dwan 2, Simon Turner 1, Elizabeth Loder 3,4, Evan Mayo-Wilson 5, Tianjing Li 6,7, Adya Misra 8, Steve McDonald 1, Andrew Forbes 1, Joanne McKenzie 1,a
PMCID: PMC8258702  PMID: 34249342

Abstract

Background : Systematic reviews underpin clinical practice and policies that guide healthcare decisions. A core component of many systematic reviews is meta-analysis, which is a statistical synthesis of results across studies. Errors in the conduct and interpretation of meta-analysis can lead to incorrect conclusions regarding the benefits and harms of interventions; and studies have shown that these errors are common. Enabling peer reviewers to better detect errors in meta-analysis through the use of a checklist provides an opportunity for these errors to be rectified before publication. To our knowledge, no such checklist exists.

Objective : To develop and evaluate a checklist to detect errors in pairwise meta-analyses in systematic reviews of interventions.

Methods : We will undertake a four-step process to develop the checklist. First, we will undertake a systematic review of studies that have evaluated errors in the conduct and interpretation of meta-analysis to generate a bank of items to consider for the checklist. Second, we will undertake a survey of systematic review methodologists and statisticians to seek their views on which items, of the bank of items generated in step 1, are most important to include in the checklist. Third, we will hold a virtual meeting to agree upon which items to include in the checklist. Fourth, before finalising the checklist, we will pilot with editors and peer reviewers of journals.

Conclusion : The developed checklist is intended to help journal editors and peer reviewers identify errors in the application and interpretation of meta-analyses in systematic reviews. Fewer errors in the conduct and improved interpretation will lead to more accurate review findings and conclusions to inform clinical practice.

Keywords: systematic review, meta-analysis, errors, checklist, reporting guideline, statistical issues, synthesis, pair-wise meta-analysis

1. Introduction

Systematic reviews (SRs) frequently underpin clinical practice guidelines and policies that guide healthcare decisions. A core component of many SRs is meta-analysis, a statistical technique used to synthesise study effect estimates from studies addressing similar questions, yielding a quantitative summary. 1 Extensions to meta-analysis (e.g. meta-regression, subgroup analysis) allow for investigation of factors that may explain variation of results across studies. These methods have the potential to provide valuable insights for healthcare decision-making; however, they are reliant on the methods being appropriately applied and interpreted.

Many errors can arise when conducting meta-analysis. For example, when meta-analysing continuous outcomes, calculations may be incorrect if standard errors are confused with standard deviations. When data are included from multi-arm trials, there is the risk that participants might be counted more than once when multiple comparisons from these trials are eligible for inclusion in the same meta-analysis. For example, from a three-arm trial of paroxetine, fluoxetine, and placebo, two comparisons would be eligible for a meta-analysis of ‘antidepressants versus placebo’ (i.e., ‘paroxetine versus placebo’ and ‘fluoxetine versus placebo’). When dealing with non-standard randomized trials – such as crossover trials, cluster-randomized trials, or split-body trials – there is a risk that variances of the effect estimates in the meta-analysis do not appropriately account for the correlation in observations induced by these designs. 2 5 Such errors can lead to studies receiving the incorrect weight in the meta-analysis with potential consequent impact on the combined estimate of intervention effect and its confidence interval, and other statistics, such as the estimated heterogeneity variance and measures of inconsistency. In some circumstances, these errors will lead to a different interpretation of the findings and review conclusions. 6

Statistical errors have been observed frequently in published SRs. For example, a study including 42 reviews from the Cochrane Cystic Fibrosis and Genetic Disorders Group found that nearly half of the SRs had at least one error (e.g., used standard error instead of standard deviation; calculated standard deviations incorrectly from the standard error given in the report; entered median instead of mean). 7 Another study in which the authors re-extracted the data from two randomly selected trials included in each of 27 meta-analyses, found errors in how the meta-analyst entered data for at least one of the two trials in 17 (63%) of the meta-analyses. 6 Furthermore, some published meta-analyses papers have been retracted because of errors in analyses and error in results and/or conclusions. 8 10

Research has shown that errors in the interpretation of statistical analyses in reviews are also common. For example, of 110 SRs indexed in MEDLINE® in 2014, 62 used the random-effects model, but 57 (92%) incorrectly interpreted the meta-analytic effect as the best estimate of a common intervention effect across studies, rather than as the average of the intervention effects across studies. In 42 of the 110 meta-analyses, a subgroup analysis was undertaken, but the findings were not interpreted with respect to a test for interaction in 29/42 (69%), and in 11/42 (26%), a claim of a subgroup difference was made based on a statistically significant effect in one group and not the other. 11 Furthermore, the potential for confounding by other factors as a possible explanation for observed subgroup effects, was not raised in any of the SRs. 11 , 12

Many strategies have been proposed to improve the conduct of meta-analysis (thus lessening the chance of errors occurring) and the interpretation of findings. These include, for example, textbooks, 13 15 training on meta-analysis methods, connection with support systems (e.g., Cochrane’s TaskExchange), and the inclusion of statisticians on review teams. However, even with these strategies, errors will still occur. A possible additional strategy is to enable peer reviewers to better detect possible errors in meta-analyses.

The peer-review process is regarded as a valuable approach for helping peer reviewers and journal editors to judge the quality, critically appraise and finally accept or reject the submitted manuscripts for publication. 16 Researchers have explored the impact of checklists to guide peer reviewers in assessing the completeness of reporting of submitted manuscripts, and have found some evidence that these are effective. 17 19 For example, training early career researchers to use the COBPeer tool (which is an online CONSORT-based peer-review tool assessing nine domains: the eight most important CONSORT domains and a switch in primary outcome(s)) helped them detect inadequate reporting in randomized trials compared to the usual review process. 17 , 20 To our knowledge, no such checklist has been developed to detect statistical errors in meta-analyses.

Aim: To develop and evaluate a checklist to detect conduct and interpretation errors in pairwise meta-analyses in systematic reviews of interventions.

2. Defining the concept of statistical errors

The notion of statistical conduct and interpretation errors is not simple. Brown and colleagues 21 defined errors to be “actions or conclusions that are demonstrably and unequivocally incorrect from a logical or epistemological point of view (e.g. … mathematical mistakes, statements not supported by the data, incorrect statistical procedures …)”. In this research, we will consider statistical errors to include those arising from underlying assumptions not being met, incorrect values used in the calculations, application of incorrect statistical methods, and misinterpretation of the results and statistical tests. We plan to initially group errors into categories (Section 3.2.4) and refine and revise these based on the types of errors identified through the systematic review (Section 3.2). Our focus will be on errors where it can be reasonably expected that a trained meta-analyst should have or could have known better, recognising that there is subjectivity in making this determination. 21

3. Methods

3.1 Contributors

A core team (RK, MJP, KD, SLT, EL, EMW, TL, AM, ABF, JEM) will lead the development of this checklist. The core team will conduct the systematic review, develop survey content and analyse survey responses, draft the checklist, coordinate piloting of the checklist, and decide the final content of the checklist. The core team consists of individuals with experience in meta-analysis methods and SR methodology, contributors of the Cochrane Handbook for Systematic Reviews of Interventions, and editors of medical journals who frequently publish SRs (BMJ, PLOS Medicine, Cochrane Database of Systematic Reviews, American Journal of Public Health, and Systematic Reviews).

3.2 Systematic review

We will conduct an SR of studies evaluating errors in the conduct and interpretation of pairwise meta-analysis, for the purpose of identifying types of errors, their prevalence, and to generate a bank of items to potentially be included in the checklist.

3.2.1 Eligibility criteria

Studies will need to meet the following eligibility criteria:

Inclusion criteria:

  • Studies evaluating types of errors (and potentially their prevalence) in the conduct and interpretation of meta-analyses (and its extensions, for example, subgroup analysis, sensitivity analysis) in SRs of interventions (irrespective of included study design);

  • Articles presenting a checklist or tool to evaluate the conduct of meta-analyses in SRs.

Exclusion criteria:

  • Studies evaluating the methodological or reporting quality or risk of bias in SRs using a tool that does not specifically examine statistical errors (e.g. PRISMA, AMSTAR-2, ROBIS) ;

  • Studies or checklists evaluating errors in statistical analyses in primary study designs (e.g. randomized trials and observational studies).

3.2.2 Search methods

We will search MEDLINE, Embase and Scopus from inception to January 2021, without any publication type or language restrictions. The search strategies for MEDLINE and Embase combine subject heading terms and text words related to statistical errors in meta-analyses are presented in Appendix (see Extended data). 22 The search was iteratively developed and tested by an experienced information specialist (SM) using a set of 10 methods articles relevant to the topic. The Scopus search closely replicates the MEDLINE search with respect to included terms and word adjacency but is limited to the following subject areas in Scopus: medicine, nursing, dentistry and health professions.

We will also search abstracts of papers and posters presented at Cochrane Colloquia since 2011 (available at https://abstracts.cochrane.org/), including the Global Evidence Summit 2017. The search strategy will be determined by assessing the relative recall of terms from eligible studies identified from searches of MEDLINE, Embase and Scopus. We will screen the reference lists and conduct a cited reference search in Web of Science of included articles and review our personal collections of reports or studies related to statistical issues in meta-analyses. In addition, we will contact organisations that produce SRs (e.g., Cochrane, Campbell Collaboration, National Institute for Health and Care Excellence), and journals that frequently publish SRs to seek any in-house checklists they are willing to share.

3.2.3 Selection of studies

Two authors will screen independently all titles and abstracts according to the aforementioned eligibility criteria and retrieve the full text of any potentially relevant articles. The same authors will screen the full texts of retrieved articles. In case of any discrepancies, a senior author will adjudicate and finalise the inclusion or exclusion of any article(s).

3.2.4 Data collection

Once we finalise the studies to be included, two authors will collect data independently from each article using a standardised data collection form. For studies evaluating types of errors (and potentially their prevalence), we will collect the following information: corresponding author name, email address, year of publication, journal name, objective(s), focus of error investigation (e.g., multi-arm trials, cross-over trials, cluster randomized trials), type and prevalence of errors, and recommendations provided for conducting meta-analyses. For articles presenting a checklist or tool to evaluate conduct or reporting of meta-analyses in SRs, we will collect the following information: checklist/tool name, method of checklist/tool development, number of items included in the checklist/tool and scope of the checklist/tool. In addition, we will collect all the items and response options pertinent to meta-analysis and its interpretation identified in the tools/checklists, and these will be added to our item bank.

Once we have extracted data from all articles, we will review the items in the item bank and remove any duplicate or redundant items. We will then group items into broader categories. For example, those relating to data type (e.g., continuous, binary), rare outcomes (i.e., handling of zero events in one or both arms), design of included studies (e.g. cross-over, cluster, multi-arm, non-randomised [e.g. interrupted time series, cohort]), type of analysis (meta-analysis, subgroup analysis, meta-regression, sensitivity analysis, publication bias analysis) and issues of interpretation. We will use Microsoft Excel 2016 for data management.

3.3 Survey

We will send an invitation to all SR methodologists and statisticians (identified from the Cochrane Methods community, Campbell Methods Coordinating Group, Society for Research Synthesis Methodology and other SR methodologists and statisticians known to the core team members) and SR editors and statistical editors (identified from Cochrane Review Groups, and those supporting other journals that frequently publish SRs) to complete a survey to inform the development of the checklist. The survey will ask respondents to:

  • 1)

    Provide their views on the most important items from the bank of items (generated from the systematic review) to include in the checklist. We will ask respondents to prioritise items that capture/identify the most common and consequential errors expected to occur in the conduct and interpretation of meta-analyses;

  • 2)

    Provide their views on specific signals (or ‘red flags’) that might lead them to conduct a more thorough investigation of statistical errors in reviews (e.g., size of the effect for some studies, meta-analysis methods used, I-squared value etc.).

We will ask researchers to provide rationale for their responses and to suggest additional items not listed in the survey. We will calculate frequencies of each response option for each item and specific signals. For an item or specific signal to meet consensus for discussion, one of the response options for the item or signal will need to be selected by more than 70% of survey respondents; this threshold was selected according to Sumsion 1998 et al. 23

3.4 Virtual meeting

Following the survey, the core team will hold a virtual meeting to agree upon which items to include in the statistical errors’ checklist for editors/peer reviewers, which items might trigger further investigation (by a statistical reviewer or the authors) and discuss how best to word each item. Attendees will discuss all items exceeding the 70% threshold in the survey. We will also send the items rated as important by fewer than 70% to meeting attendees prior to the meeting, to provide them with the chance of “rescuing” items for discussion at the meeting. Following the meeting, the core team will draft the checklist and an accompanying guidance document (with examples for each item).

3.5 Piloting

In the first stage of piloting, two reviewers will read the draft checklist and guidance document before independently applying it to a random sample of 20 reviews; 10 from each of two previous methodological studies that collated systematic reviews. Specifically, the first methodological study includes 42 SRs of nutrition research (the ROBUST study) 24 that were published between January 2018 and June 2019, and the second includes 31 SRs of interventions for arthritis, depression or anxiety (the SIM study) that were published between January 2010 and January 2012. 25 The reviewers will record issues on whether the wording of items is ambiguous or difficult to interpret, and those items will be discussed by the core team and improved.

In the second stage of piloting, we will invite associate editors and peer reviewers of journals that frequently handle SR submissions to pilot the checklist and provide feedback on its usability. After collating the feedback received from the peer reviewers and editors, we will finalise the checklist and the accompanying guidance document.

3.6 Dissemination and knowledge translation strategy

We plan to publish the developed checklist and guidance document in an open-access journal. We will disseminate the checklist via presentations and workshops at relevant conferences and workshops focused on SR methodology, health technology assessment and evidence-based medicine (e.g., Cochrane Colloquia, Evidence Live, HTAi Annual Meeting, The Global Evidence Summit, G-I-N Conference), and via social media, and in a series of international webinars. We also plan to create a user-friendly, online version of the checklist and guidance document for use by journals that publish SRs and meta-analyses.

The developed checklist and related work will be published in open access journals. Associated datasets, data collection forms and analyses not included in any publication will be made publicly available via an online repository.

4. Conclusion

The developed checklist could help journal editors and peer reviewers identify errors in the conduct and interpretation of meta-analyses in SRs. Fewer errors and improved interpretation would lead to more accurate SR findings and conclusions to inform clinical practice.

Study status: We have completed searching and screening for sections 3.2.1, 3.2.2 and 3.2.3 for systematic review.

Data availability

Underlying data

No underlying data are associated with this article.

Extended data

Bridges: Search strategy for systematic review of studies evaluating errors in the conduct and interpretation of pairwise meta-analysis. Monash University. DOI: https://doi.org/10.26180/14446293.v1. 22

This project contains the following extended data:

  • -

    Search strategy for systematic review of studies evaluating errors in the conduct and interpretation of pairwise meta-analysis

Data are available under the terms of the Creative Commons Attribution 4.0 International license (CC-BY 4.0).

Funding Statement

The author(s) declared that no grants were involved in supporting this work.

[version 1; peer review: 2 approved]

References

  • 1. McKenzie JE, Brennan SE: Chapter 12: Synthesizing and presenting findings using other methods. Cochrane Handbook for Systematic Reviews of Interventions. 2019:321–47. [Google Scholar]
  • 2. Richardson M, Garner P, Donegan S: Cluster Randomised Trials in Cochrane Reviews: Evaluation of Methodological and Reporting Practice. PLoS One. 2016;11(3):e0151818. 10.1371/journal.pone.0151818 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3. Donner A, Klar N: Issues in the meta-analysis of cluster randomized trials. Stat Med. 2002;21(19):2971–80. 10.1002/sim.1301 [DOI] [PubMed] [Google Scholar]
  • 4. Stedman MR, Curtin F, Elbourne DR, et al. : Meta-analyses involving cross-over trials: methodological issues. Int J Epidemiol. 2011;40(6):1732–4. 10.1093/ije/dyp345 [DOI] [PubMed] [Google Scholar]
  • 5. Nolan SJ, Hambleton I, Dwan K: The use and reporting of the cross-over study design in clinical trials and systematic reviews: a systematic assessment. PLoS One. 2016;11(7):e0159014. 10.1371/journal.pone.0159014 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6. Hrobjartsson A, Gotzsche PC: Unsubstantiated claims of large effects of placebo on pain: serious errors in meta-analysis of placebo analgesia mechanism studies. J Clin Epidemiol. 2006;59(4):336–8. discussion 9–41. 10.1016/j.jclinepi.2005.05.011 [DOI] [PubMed] [Google Scholar]
  • 7. Jones AP, Remmington T, Williamson PR, et al. : High prevalence but low impact of data extraction and reporting errors were found in Cochrane systematic reviews. J Clin Epidemiol. 2005;58(7):741–2. 10.1016/j.jclinepi.2004.11.024 [DOI] [PubMed] [Google Scholar]
  • 8. Roba AA, Tefera M, Worku T, et al. : Retraction Note: Application of 4% chlorhexidine to the umbilical cord stump of newborn infants in lower income countries: a systematic review and meta-analysis. Matern Health Neonatol Perinatol. 2020;6:4. 10.1186/s40748-020-00118-y [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9. Guo LQ, Chen Y, Mi BB, et al. : Retraction Note to: Ambient air pollution and adverse birth outcomes: a systematic review and meta-analysis. J Zhejiang Univ Sci B. 2020;21(9):756. 10.1631/jzus.B18r0122 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10. Ezenwa BN, Bello S, Oyenusi EE, et al. : RETRACTED: Galactogogues use Among Mothers With Preterm Births: A Systematic Review and Meta-Analysis. J Hum Lact. 2020;36(3):NP3–NP14. 10.1177/0890334420914629 [DOI] [PubMed] [Google Scholar]
  • 11. Page MJ, Altman DG, McKenzie JE, et al. : Flaws in the application and interpretation of statistical analyses in systematic reviews of therapeutic interventions were common: a cross-sectional analysis. J Clin Epidemiol. 2018;95:7–18. 10.1016/j.jclinepi.2017.11.022 [DOI] [PubMed] [Google Scholar]
  • 12. Fisher DJ, Carpenter JR, Morris TP, et al. : Meta-analytical methods to identify who benefits most from treatments: daft, deluded, or deft approach? BMJ. 2017;356. 10.1136/bmj.j573 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13. Cooper H, Hedges LV, Valentine JC: The handbook of research synthesis and meta-analysis: Russell Sage Foundation; 2019. [Google Scholar]
  • 14. Schmid CH, White IR, Stijnen T: Introduction to Systematic Review and Meta-Analysis. Handbook of Meta-Analysis. 2020. [Google Scholar]
  • 15. Higgins JP, Thomas J, Chandler J, et al. : Cochrane handbook for systematic reviews of interventions. John Wiley & Sons;2019. [Google Scholar]
  • 16. Rennie D: Editorial peer review in biomedical publication: the first international congress. JAMA. 1990;263(10):1317. 10.1001/jama.1990.03440100011001 [DOI] [PubMed] [Google Scholar]
  • 17. Chauvin A, Ravaud P, Moher D, et al. : Accuracy in detecting inadequate research reporting by early career peer reviewers using an online CONSORT-based peer-review tool (COBPeer) versus the usual peer-review process: a cross-sectional diagnostic study. BMC Medicine. 2019;17(1):205. 10.1186/s12916-019-1436-0 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18. Speich B, Schroter S, Briel M, et al. : Impact of a short version of the CONSORT checklist for peer reviewers to improve the reporting of randomised controlled trials published in biomedical journals: study protocol for a randomised controlled trial. BMJ Open. 2020;10(3):e035114. 10.1136/bmjopen-2019-035114 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19. Blanco D, Schroter S, Aldcroft A, et al. : Effect of an editorial intervention to improve the completeness of reporting of randomised trials: a randomised controlled trial. BMJ Open. 2020;10(5):e036799. 10.1136/bmjopen-2020-036799 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 20. Chauvin A, Moher D, Altman D, et al. : A protocol of a cross-sectional study evaluating an online tool for early career peer reviewers assessing reports of randomised controlled trials. BMJ Open. 2017;7(9):e017462. 10.1136/bmjopen-2017-017462 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21. Brown AW, Kaiser KA, Allison DB: Issues with data and analyses: Errors, underlying themes, and potential solutions. Proc Natl Acad Sci U S A. 2018;115(11):2563–70. 10.1073/pnas.1708279115 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22. Page M, Kanukula R, McKenzie J: Search strategy for systematic review of studies evaluating errors in the conduct and interpretation of pairwise meta-analysis. Monash University. Online resource. 2021. [database on the Internet]. 10.26180/14446293 [DOI]
  • 23. Sumsion T: The Delphi technique: an adaptive research tool. Br J Occup Ther. 1998;61(4):153–6. 10.1177/030802269806100403 [DOI] [Google Scholar]
  • 24. Page MJ, Bero L, Kroeger CM, et al. : Investigation of Risk Of Bias due to Unreported and SelecTively included results in meta-analyses of nutrition research: the ROBUST study protocol. F1000Res. 2019;8(1760):1760. 10.12688/f1000research.20726.2 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25. Page MJ, Forbes A, Chau M, et al. : Investigation of bias in meta-analyses due to selective inclusion of trial effect estimates: empirical study. BMJ Open. 2016;6(4):e011863. 10.1136/bmjopen-2016-011863 [DOI] [PMC free article] [PubMed] [Google Scholar]
F1000Res. 2021 Jul 5. doi: 10.5256/f1000research.56375.r87112

Reviewer response for version 1

Therese D Pigott 1

This work will provide an important contribution to the field, highlighting typical errors that can occur in pairwise meta-analyses of intervention effects. One question I have is the audience of the checklist. The study protocol is situated in meta-analysis for health and medical interventions. The search strategy for relevant systematic reviews for stage one of the process focuses on the health and medical literature, with only the Campbell Collaboration as a source for meta-analyses in the social sciences. If the intent of the checklist is to apply more broadly to social interventions rather than just health and medicine, then the search strategy needs to include a broader set of databases. There are relevant efforts in psychology for example (see Maasen et al. (2020 1 )) that could also inform the checklist. I understand if the authors wish to focus this effort only on health and medicine. 

In the social sciences, similar issues to those that are outlined in the protocol occur. In addition, many meta-analyses include a large number of studies, and thus key issues center on dependent effect sizes within studies and interpretation of effect size models for heterogeneity. If the authors intend the checklist to have a broader application, then these are a few issues that should be considered in the development of the checklist.

Is the study design appropriate for the research question?

Yes

Is the rationale for, and objectives of, the study clearly described?

Yes

Are sufficient details of the methods provided to allow replication by others?

Yes

Are the datasets clearly presented in a useable and accessible format?

Yes

Reviewer Expertise:

meta-analysis methods

I confirm that I have read this submission and believe that I have an appropriate level of expertise to confirm that it is of an acceptable scientific standard.

References

  • 1. : Reproducibility of individual effect sizes in meta-analyses in psychology. PLoS One .2020;15(5) : 10.1371/journal.pone.0233107 e0233107 10.1371/journal.pone.0233107 [DOI] [PMC free article] [PubMed] [Google Scholar]
F1000Res. 2021 Jun 21. doi: 10.5256/f1000research.56375.r87114

Reviewer response for version 1

Ram Bajpai 1

Authors have well prepared their systematic review protocol with clarity. I have following minor suggestions to improve the quality of this protocol.

From inclusion criteria it is clear that authors will include SRs of interventions only. However, if possible, authors should elaborate on not including SR of proportions or SR of observational studies or SR of biomarkers etc. Each set of SRs will require a different type of statistical skills to perform meta-analysis. It also important to mention about whether only pharmacological intervention SRs or non- pharmacological interventions such as educational interventions etc. will also be included in this methodological systematic review.

Did authors think that a statistician included in the author list will have any impact on error as it is possible that SRs with statistician as co-author may have less statistical errors? I think the authors have already thought about this point, however, appreciate authors reflection on this point and if they add few words in the protocol.

If I am not wrong, the planned survey and consensus meeting will follow Delphi survey methodology. I am wondering that authors did not mention this word explicitly or they are following some other equally useful methodology.

I believe that authors have well refined methodology to capture statistical errors. However, in many SRs, authors wrongly choose a meta-analysis model (i.e., fixed-effect vs. random-effect) to pool study estimates and possible other statistics right. How would authors capture such errors? Please elaborate if I have missed this point.

Is the study design appropriate for the research question?

Yes

Is the rationale for, and objectives of, the study clearly described?

Yes

Are sufficient details of the methods provided to allow replication by others?

Yes

Are the datasets clearly presented in a useable and accessible format?

Yes

Reviewer Expertise:

Meta-analysis, Statistical analysis

I confirm that I have read this submission and believe that I have an appropriate level of expertise to confirm that it is of an acceptable scientific standard.

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Data Availability Statement

    Underlying data

    No underlying data are associated with this article.

    Extended data

    Bridges: Search strategy for systematic review of studies evaluating errors in the conduct and interpretation of pairwise meta-analysis. Monash University. DOI: https://doi.org/10.26180/14446293.v1. 22

    This project contains the following extended data:

    • -

      Search strategy for systematic review of studies evaluating errors in the conduct and interpretation of pairwise meta-analysis

    Data are available under the terms of the Creative Commons Attribution 4.0 International license (CC-BY 4.0).


    Articles from F1000Research are provided here courtesy of F1000 Research Ltd

    RESOURCES