Skip to main content
Contemporary Clinical Trials Communications logoLink to Contemporary Clinical Trials Communications
. 2019 Mar 11;14:100343. doi: 10.1016/j.conctc.2019.100343

Study reporting guidelines: How valid are they?

Catherine Arundel 1,, Sophie James 1, Matthew Northgraves 1, Alison Booth 1
PMCID: PMC6421355  PMID: 30923775

Abstract

Reporting guidelines help improve the reporting of specific study designs, and clear guidance on the best approaches for developing guidelines is available. The methodological strength, or validation of guidelines is however unclear. This article explores what validation of reporting guidelines might involve, and whether this has been conducted for key reporting guidelines.

Keywords: Reporting guidelines, Validity, Validation

1. Introduction

Comprehensive reporting can reduce reporting bias, enable informed decision making in clinical practice, limit duplication of effort and inform subsequent research [1]. The quality of reporting of research activity continues to be inadequate, presenting readers with difficulties in judging the reliability of research findings, or how best to interpret results for individual settings [2,3].

Reporting guidelines have been developed to help improve the reporting of specific study designs. If followed by authors this should enable users to understand the design, conduct and analysis of the research, to critically appraise and review the findings and interpret the conclusions appropriately [4].

A guideline is a checklist, diagram or explicit text which guides authors in reporting research, and should be developed using explicit methodology [2]. Many already exist, mostly as checklists, and clear guidance, including a checklist of recommended steps, for developing such tools is available [2].

2. Use of guidelines

A search of the websites of five leading medical and health research journals (BMJ, Journal of the American Medical Association (JAMA), Lancet, New England Journal of Medicine (NEJM) and BMC Trials) identified the reporting guidelines included in the journals’ instructions to authors. These journals were purposively sampled as they: are prominent in the publication of a wide range of research topics and study designs; each publish significant volumes of research over a 3-month period (RCT publication rate, range 1–10 per month; 4–31 per quarter); and represent a range of impact factors (Range: 2.067 to 79.258).

All five journals require the use of the CONSORT reporting guidelines for randomised controlled trial (RCT) manuscripts. For RCTs, the BMJ also specifically recommend use of the TIDieR checklist to ensure accurate and complete reporting of a trial intervention [5]. BMC Trials, BMJ, JAMA and Lancet promote the use of reporting guidelines for other study designs and refer authors to the EQUATOR (Enhancing the QUAlity and Transparency Of health Research) Network database of reporting guidelines [6].

The EQUATOR Network is an international multidisciplinary group that promotes transparent and accurate health research reporting through use of reporting guidelines. The network provides access to a comprehensive range of such guidelines in a searchable database (http://www.equator-network.org). However, there is evidence that simply having reporting guidelines, even with journal endorsement of their use, is insufficient [7].

The EQUATOR Network takes an inclusive approach and are clear that there is no indication of methodological strength, or validation of the guidelines listed. The reporting guidelines for the main study types, such as CONSORT for RCTs, STROBE for observational studies, and STARD for diagnostic/prognostic studies, are highlighted on the front page of the EQUATOR Network website. We decided to explore what validation of reporting guidelines might involve.

3. Validation

Validation is ‘the action of checking or proving the validity or accuracy of something’ [8] principles already well established in the development of health care and research documentation.

In the UK the National Institute for Health and Care Excellence (NICE) requires validation of guidelines designed to inform decisions in health, public health and social care (https://www.nice.org.uk/about/what-we-do/our-programmes/nice-guidance). As a minimum, validation comprises stakeholder review, with fieldwork to trial implementation and discussions with service users, with external review also included if warranted, for example for guidelines in complex or sensitive areas (https://www.nice.org.uk/process/pmg20/chapter/the-validation-process-for-draft-guidelines-and-dealing-with-stakeholder-comments).

A number of initiatives provide standards for the development and validation of patient reported outcome measures (PROMs) for research [[9], [10], [11]]. This includes identifying the scope and focus of the measure, reviewing literature, engaging relevant stakeholders, and consensus assessment. Guidance for the development of health research reporting guidelines suggests similar features to those used for PROMs: 1) Literature review; 2) Delphi process; 3) Identification of key items; 4) Meeting with collaborators; 5) Iterative revision and review; 6) Pilot testing, all of which are derived from the authors’ comprehensive experience in the development of reporting guidelines [2]. While there are many similarities in the proposed activities for the development of PROMs and reporting guidelines, unlike PROMS, methods of validation of reporting guidelines are not explicitly mentioned.

In 2016 we conducted a systematic literature search using MEDLINE to identify validation methods commonly used for PROMs. Search strategies are provided in Supplementary Document 1. Our pre-defined inclusion criteria were for studies: focused on PROMs; detailing a validation method; references other publications regarding validation. Two authors independently screened the search results against the inclusion criteria, identified 73 relevant papers. Details of the included papers are provided in Supplementary Document 2. Data on PROM type, validation method, and if this was noted as a strength or limitation were extracted and a summary of the validation methods identified is detailed in Table 1, Table 2.

Table 1.

Types of validation method used for PROMS development.

Validation Method Number of studies using methoda Number of studies using single method only
Comparison to similar measures n = 23 n = 5
Delphi consensus n = 1 n = 1
Expert opinion n = 1 n = 0
Focus groups n = 3 n = 0
Interviews n = 7 n = 3
Statistical testing n = 56 n = 31
a

Note some studies used multiple methods, therefore total n exceeds number of papers included.

Table 2.

Combinations of validation methods used for PROMS development.

Validation Method Number of studies using method
Statistical testing plus comparison to similar measures n = 17
Statistical testing plus expert opinion n = 1
Statistical testing plus focus groups n = 1
Statistical testing plus interviews n = 2
Statistical testing plus comparison to similar measures plus focus groups n = 1
Statistical testing plus comparison to similar measures plus interviews n = 1
Statistical testing plus focus groups plus interviews n = 1

By far the most common method of validation was use of statistical testing either as a single validation method or in combination with other methods. The most common combination was statistical testing in conjunction with comparison with similar measures. This corresponds to guidance published in 2011 which indicates that comparison and correlation with similar, existing measures is critical in the development of PROMS [10].

4. Are reporting guidelines validated?

Having established the methods of validation, we went on to see which had been used in the reporting guidelines highlighted on the EQUATOR Network homepage.

We conducted a literature search in 2018 to identify papers reporting the development of guidelines for the main study types as highlighted on the EQUATOR network website. Two researchers independently extracted information about the development and validation methods reported; disagreements were resolved through discussion. We excluded papers where content analysis was the sole measure used, as this was not explicitly identified as a validation activity. The results are summarised in Table 3.

Table 3.

Validation methods used in reporting guidelines for main study types.

Research Type Key Guideline Validation Methods
RCT CONSORT [12]
  • -

    Stakeholder meeting

  • -

    Evidence synthesis

  • -

    Stakeholder review of draft

Observational Studies STROBE [13]
  • -

    Literature search

  • -

    Stakeholder meeting

  • -

    Stakeholder review of drafts (following iterative review)

  • -

    Peer review

Case Reports
Systematic Review PRISMA [14]
  • -

    Literature search

  • -

    Survey of peers using systematic review methodology

  • -

    Stakeholder meeting and review of drafts

Qualitative COREQ [4]
  • -

    Literature search

  • -

    Consensus discussions between authors

ENTREQ [15]
  • -

    Literature review

  • -

    Inductive generation of items

  • -

    Pilot testing

  • -

    Iterative review

  • -

    Pilot testing

Diagnostic/Prognostic STARD (2015) [16]
  • -

    Literature review

  • -

    Survey

  • -

    Stakeholder meeting

  • -

    Pilot testing

TRIPOD [17]
  • -

    Stakeholder meeting

  • -

    Literature search

  • -

    Survey

  • -

    Stakeholder meeting and review of drafts

Economic Evaluation CHEERS [18]
  • -

    Systematic review

  • -

    Survey of collaborators

  • -

    Modified Delphi panel (2 rounds)

Protocols SPIRIT [19]
  • -

    Delphi consensus survey

  • -

    Evidence synthesis

  • -

    Stakeholder meetings

  • -

    Pilot testing

PRISMA-P [20]
  • -

    Mapping of existing guidelines

  • -

    Stakeholder meeting and review of drafts

  • -

    Delphi consensus (from PROSPERO)

The methods described within the papers matched the principles outlined in the guidance for the development of health research reporting guidelines [2]. While some guideline developers utilised multiple components and others were more selective, we believe the overarching principles remained. In the absence of clear statements, a pragmatic interpretation would say, for example, that evidence synthesis requires a literature review, and having done a literature review, or convened a stakeholder meeting, it can be supposed that validation methods were used. Of note here, is that although following the key principles, this activity is not noted as being ‘validation’. This could account for why this term is not, used in the context of promoting the use of reporting guidelines.

5. Discussion

Reporting guidelines, are available for a wide range of health care research methodologies [6]. Many journals request their use to increase transparent research reporting, however mandated use is rare, despite evidence that reporting guidelines can have a positive impact on completeness of reporting [7,21].

Validation is important to ensure the validity and accuracy of tools used within the conduct and reporting of research. Whilst the validation of PROMS is frequently reported, we have identified that while validation activities for reporting guidelines do occur, the activities are not always explicitly reported as such. This may occur because some validation activities are also part of the development process, for example a consensus exercise. Reporting of the development of future reporting guidelines for research, may benefit from clearly identifying the work undertaken to ensure the accuracy of the guidelines proposed. This could be within a ‘Validation’ section of a guideline publication or by simple use of the words ‘validated/validation’ in the context of the activity being reported. For completeness of reporting, it may also be appropriate to request use of a development and validation checklist, for example that provided by Moher et al. {2}, where guideline development is reported.

The EQUATOR Network database currently contains around 406 reporting guidelines which cover a variety of research methodologies, many either specialised or narrow in scope. It is unclear how many of these included validation activities in their development, and we have not been able to identify any post development or publication validation work from our literature search. Ensuring validation activities are not only undertaken but also clearly reported could add weight to the value of reporting guidelines for both those promoting their use and those authoring papers. By understanding what validation looks like, we would suggest that journals and peer reviewers could be encouraged to mandate the use of validated checklists.

Despite being included as one of the elements for the development of health research reporting guidelines, it is surprising that a limited number of the guideline development papers used pilot testing prior to publication. Given that reporting guidelines are intended to ensure transparent reporting across similar research methodologies, pilot testing may be applicable to the development of reporting guidelines.

Although we used systematic review methods to identify and select papers, we acknowledge that some may have been missed, however any impact from missed papers is likely to be limited.

6. Conclusion

The reporting of guidelines while including details of their development, frequently fail to explicitly identify validation activities even when they have clearly been undertaken. While this may appear to be a semantic or even pedantic issue, emphasising that reporting guidelines have been validated could help encourage authors to use the guidelines, publishers and journals to mandate checklist submission with manuscripts, and peer reviewers to monitor accuracy of completion. An improvement in any, and ideally all, of these approaches would be beneficial in promoting high quality research and reducing research waste.

Authorship

Conception and design: CA, AB.

Analysis and interpretation of data: CA, MN, SJ.

Drafting of the article: CA.

Critical revision for important intellectual content: AB, MN, SJ.

Conflicts of interest

CA, MN, SJ declare they have no conflicts of interest. AB is a member of the PRISMA-P group.

Footnotes

Appendix A

Supplementary data to this article can be found online at https://doi.org/10.1016/j.conctc.2019.100343.

Contributor Information

Catherine Arundel, Email: catherine.arundel@york.ac.uk.

Sophie James, Email: Sophie.j.james@york.ac.uk.

Matthew Northgraves, Email: matt.northgraves@york.ac.uk.

Alison Booth, Email: Alison.booth@york.ac.uk.

Funding

This research did not receive any specific grant from funding agencies in the public, commercial, or not-for-profit sectors.

Appendix A. Supplementary data

The following are the Supplementary data to this article:

Multimedia component 1
mmc1.docx (11.1KB, docx)
Multimedia component 2
mmc2.docx (17.6KB, docx)
Multimedia component 3
mmc3.xml (258B, xml)

References

  • 1.Chan A., Song F., Vickers A., Jefferson T., Dickersin K., Gotzsche P.C. Increasing value and reducing waste: addressing inaccessible research. The Lancet. 2014;383:257–266. doi: 10.1016/S0140-6736(13)62296-5. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 2.Moher D., Schulz K.F., Simera I., Altman D.G. Guidance for developers of health research reporting guidelines. PLoS Med. 2010;7(2) doi: 10.1371/journal.pmed.1000217. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3.Macleod M., Michie S., Roberts I., Dirnagl U., Chalmers I., Ioannidis J.P.A. Biomedical research: increasing value, reducing waste. The Lancet. 2014;383(9912):101–104. doi: 10.1016/S0140-6736(13)62329-6. [DOI] [PubMed] [Google Scholar]
  • 4.Tong A., Sainsbury P., Craig J. Consolidated criteria for reporting qualitative research (COREQ): a 32-item checklist for interviews and focus groups. Int. J. Qual. Health Care. 2007;19(6):349–357. doi: 10.1093/intqhc/mzm042. [DOI] [PubMed] [Google Scholar]
  • 5.Hoffmann T., Glasziou P.P., Boutron I., Milne R., Perera R., Moher D. Better reporting of interventions: template for intervention description and replication (TIDieR) checklist and guide. BMJ. 2014;348(g1687) doi: 10.1136/bmj.g1687. [DOI] [PubMed] [Google Scholar]
  • 6.Altman D., Simera I., Hoey J., Moher D., Schulz K.F. EQUATOR: reporting guidelines for health research. The Lancet. 2008;31:1149–1150. doi: 10.1016/S0140-6736(08)60505-X. [DOI] [PubMed] [Google Scholar]
  • 7.Stevens A., Shamseer L., Weinstein E., Yazdi F., Turner L., Thielman J. Relation of completeness of reporting of health research to journals' endorsement of reporting guidelines: systematic review. BMJ. 2014;348(g3804) doi: 10.1136/bmj.g3804. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.The Oxford English Dictionary. 2018. [PMC free article] [PubMed] [Google Scholar]
  • 9.COMET Initiative 2018 [09.10.18] http://www.comet-initiative.org/ Available from:
  • 10.Rothrock N., Kaiser K.A., Cella D. Developing a valid patient reported outcome measure. Clin. Pharmacol. Ther. 2011;90(5):737–742. doi: 10.1038/clpt.2011.195. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11.Alonso J., Bartlett S., Rose M., Aaronson N., Chaplin J., Efficace F., Forrest C. The case for an international patient-reported outcomes measurement information system (PROMIS(R)) initiative. Health Qual. Life Outcomes. 2013;11(1):2010. doi: 10.1186/1477-7525-11-210. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12.Schulz K., Altman D.G., Moher D., CONSORT Group CONSORT 2010 Statement: updated guidelines for reporting parallel group randomised trials. BMJ. 2010;340(C):332. doi: 10.3736/jcim20100702. [DOI] [PubMed] [Google Scholar]
  • 13.von Elm E., Altman D.G., Egger M., Pocock S.J., Gotzsche P.C., Vandenbroucke J.P. The strengthening the reporting of observational studies in epidemiology (STROBE) statement: guidelines for reporting observational studies. BMJ. 2007;335(7624):806–808. doi: 10.1136/bmj.39335.541782.AD. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14.Moher D., Liberati A., Tetzlaff J., Altman D.G. The PRISMA group. Preferred reporting items for systematic reviews and meta-analyses: the PRISMA statement. BMJ. 2009;339 b2535. [PMC free article] [PubMed] [Google Scholar]
  • 15.Tong A., Fleming K., McInnes E., Oliver S., Craig J. Enhancing transparency in reporting the synthesis of qualitative research: ENTREQ. BMC Med. Res. Methodol. 2012;12(1):181. doi: 10.1186/1471-2288-12-181. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16.Bossuyt P., Reitsma J.B., Bruns D.E., Gatsonis C.A., Glasziou P.P., Irwig L. STARD 2015: an updated list of essential items for reporting diagnostic accuracy studies. BMJ. 2015;351 doi: 10.1136/bmj.h5527. h5527. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17.Collins G., Reitsma J.B., Altman D.G., Moons K.G. Transparent reporting of a multivariable prediction model for individual prognosis or diagnosis (TRIPOD): the TRIPOD statement. BMJ. 2015;350 doi: 10.1136/bmj.g7594. G7594. [DOI] [PubMed] [Google Scholar]
  • 18.Husereau D., Drummond M., Petrou S., Carswell C., Moher D., Greenberg D. Consolidated health economic evaluation reporting standards (CHEERS) statement. Eur. J. Health Econ. 2013;14(3):367–372. doi: 10.1007/s10198-013-0471-6. [DOI] [PubMed] [Google Scholar]
  • 19.Chan A., Tetzlaff J.M., Altman D.G., Laupacis A., Gotzsche P.C., Krleza-Jeric K. SPIRIT 2013 Statement: defining standard protocol items for clinical trials. Ann. Intern. Med. 2013;158(3):200–207. doi: 10.7326/0003-4819-158-3-201302050-00583. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 20.Moher D., Shamseer L., Clarke M., Ghersi D., Liberati A., Petticrew M. Preferred reporting items for systematic review and meta-analysis protocols (PRISMA-P) 2015 statement. Syst. Rev. 2015;4(1):1. doi: 10.1186/2046-4053-4-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21.Turner L., Shamseer L., Altman D.G., Schulz K.F., Moher D. Does use of the CONSORT statement impact the completeness of reporting of randomised controlled trials published in medical journals? Cochrane Syst. Rev. 2012;1(60) doi: 10.1186/2046-4053-1-60. [DOI] [PMC free article] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Multimedia component 1
mmc1.docx (11.1KB, docx)
Multimedia component 2
mmc2.docx (17.6KB, docx)
Multimedia component 3
mmc3.xml (258B, xml)

Articles from Contemporary Clinical Trials Communications are provided here courtesy of Elsevier

RESOURCES