Skip to main content
PLOS One logoLink to PLOS One
. 2020 Jul 8;15(7):e0235485. doi: 10.1371/journal.pone.0235485

Methods used to assess outcome consistency in clinical studies: A literature-based evaluation

Ewelina Rogozińska 1,2,*, Elizabeth Gargon 3, Rocío Olmedo-Requena 4,5,6, Amani Asour 2, Natalie A M Cooper 2, Claire L Vale 1, Janneke van’t Hooft 7
Editor: Tim Mathes8
PMCID: PMC7343158  PMID: 32639999

Abstract

Evaluation studies of outcomes used in clinical research and their consistency are appearing more frequently in the literature, as a key part of the core outcome set (COS) development. Current guidance suggests such evaluation studies should use systematic review methodology as their default. We aimed to examine the methods used. We searched the Core Outcome Measures in Effectiveness Trials (COMET) database (up to May 2019) supplementing it with additional resources. We included evaluation studies of outcome consistency in clinical studies across health subjects and used a subset of A MeaSurement Tool to Assess systematic Reviews (AMSTAR) 2 (items 1–9) to assess their methods. Of 93 included evaluation studies of outcome consistency (90 full reports, three summaries), 91% (85/93) reported performing literature searches in at least one bibliographic database, and 79% (73/93) was labelled as a “systematic review”. The evaluations varied in terms of satisfying AMSTAR 2 criteria, such that 81/93 (87%) had implemented PICO in the research question, whereas only 5/93 (6%) had included the exclusions list. None of the evaluation studies explained how inconsistency of outcomes was detected, however, 80/90 (88%) concluded inconsistency in individual outcomes (66%, 55/90) or outcome domains (20%, 18/90). Methods used in evaluation studies of outcome consistency in clinical studies differed considerably. Despite frequent being labelled as a “systematic review”, adoption of systematic review methodology is selective. While the impact on COS development is unknown, authors of these studies should refrain from labelling them as “systematic review” and focus on ensuring that the methods used to generate the different outcomes and outcome domains are reported transparently.

Introduction

Inconsistency (or heterogeneity) of outcomes measured in clinical studies is a widely recognised problem hindering evidence synthesis. [15] Core outcome sets (COS), defined as a minimum set of outcomes to be reported from all intervention trials sharing a common research objective, have been advocated as a solution to this problem. [6, 7] A growing number of studies aiming to develop a COS for conditions across a range of health areas [8, 9], is accompanied by reviews aiming to assess the consistency of outcomes in a formal way. [1012]

Prior to the release of the Core Outcome Measures in Effectiveness Trials (COMET) handbook [13], guidance on a COS-related methodology was mostly focused on aspects of the consensus process. [14] Review of past research appears in the handbook in the context of assessing a need for a COS—described as an optional step—and informing a list of outcomes for a consensus process. The guidance on the conduct of this type of assessments is succinct and suggests the adoption of a systematic review approach. [1315]

The method of systematically reviewing literature was introduced as a comprehensive way of summarising the evidence for the purpose of medical decision-making and identification of unanswered research questions. [1619] The methodological rigour required for systematic review, which intends to minimise biases and provide a robust estimation of an underlying treatment effect, [16] requires considerable time and resources. [19, 20] The benefits of applying the same approach to the assessment of outcome consistency or generating a long list of outcomes for Delphi survey is unclear. [21]

Therefore, we set out to assess the methods adopted in evaluation studies of outcome consistency published to-date and examine their adoption of systematic review methods. Furthermore, we checked if identified evaluation studies were part of a COS project, examined methods specific to determining outcome consistency, such as type of collected outcomes, methods of their identification, and how authors assessed and presented outcome consistency or need for a COS.

Materials and methods

Our work was guided by a prospectively developed protocol registered with PROSPERO (CRD42018100481).

Identification of relevant studies

We included full texts of evaluation studies of outcome consistency in clinical studies on any health condition. We recognise that, even though, the COMET handbook refers to these type of evaluations as “systematic reviews” there is no consensus on the type of study design for this type of evaluation studies. Aiming to gain a thorough overview of the practices in this area and acknowledging lack of consensus regarding the study design, we decided to include any study design regardless of the design labelling. We searched the COMET database from its inception to June 2018; the search was updated in May 2019. [8] The COMET database is an annually updated repository of the international COS literature based on the systematic searches run in MEDLINE, SCOPUS, and Cochrane Methodology Register. [9] The search was supplemented with the resource of the Core Outcomes in Women’s and Newborn Health (CROWN) initiative [10, 22] and check of references of the included studies. Systematic reviews of treatment effects with a secondary conclusion regarding outcome reporting, reviews of outcomes from non-clinical settings (registers, audits, population databases), COS protocols and reports mentioning outcome assessment but not describing them and studies focusing on a very narrow group of outcomes (e.g. only pain-related outcomes) were excluded.

Data collection

All data were collected using a prospectively developed and piloted data collection form (S1 Text) and subsequently amalgamated into a master file in MS Excel. We collected information on publication year, medical speciality, evaluation’s aims, scope (number of included studies and their type), whether the design was labelled as a “systematic review” and was it a part of a COS project (based on study acronym or the information provided in the publication). In order to examine the adoption of systematic review methodology, all evaluation studies (regardless of declared study design) were assessed against a tailored subset of A MeaSurement Tool to Assess systematic Reviews (AMSTAR) 2 items [23] we felt were relevant (S1 Table). AMSTAR was designed as a practical critical appraisal instrument enabling rapid and reproducible quality assessments of systematic reviews. In our work, we used items covering the development of a study protocol (items 1–3), identification and selection of eligible studies (items 4–5), data collection (item 6), reporting (item 7–8) and study quality assessment (item 9). The remaining items (10 to 16) were not assessed as we felt they were not applicable to evaluation studies of outcome consistency (e.g. study funding, aspects of meta-analysis, publication bias, etc.).

Information collected on the outcome-specific methods covered following elements: type of outcome (primary/secondary); a way of determining outcome type (clearly specified as primary or secondary, used in power calculation, etc.); approach to outcome extraction (any outcome reported in the publication or just those from the methods or results sections); use of any tool to assess the quality of outcome description; outcome unit as reported by the publication (individual outcomes or outcome groups); presentation of the findings (text, table, graphic format); and the conclusions. The conclusion section was examined for presence of any statement regarding inconsistency of assessed outcomes (heterogeneity, variation, etc.) or a need for a COS. Where detection of outcome inconsistency was concluded, we examined the methods and protocol of the evaluation, if available, for any description of how outcome inconsistency was defined. Where the need for a COS development was concluded, we looked for arguments given in the publications to support this conclusion. Double data extraction (AA, ROR or ER) and assessment of the subset of AMSTAR 2 items were performed on half of the included evaluations. Remaining evaluations were extracted and assessed by a single senior reviewer (ER). All discrepancies and statements supporting final judgements were recorded in the final Excel dataset.

Strategy for data synthesis

Extracted data were cross-tabulated and presented either as frequencies or as median with corresponding interquartile range (IQR). When examining detection of outcome consistency, we assessed only studies with a full description of outcome assessment. Fisher’s exact test was used to explore the potential difference between the evaluation studies labelled as “systematic review” and those that were not labelled as such. All descriptive summaries and graphs were computed with Excel software (MS Office 2016). The comparison of two groups—labelled and not labelled as “systematic review” was performed using Stata version 15.1.

Results

Study selection process and description of included studies

Out of 237 records assessed, 93 evaluation studies met the inclusion criteria—90 full reports and three summaries of outcome assessment in the final report of the COS project (S1 Fig). The main reason for records exclusion was that the objective was outside of this work’s scope, e.g. assessment of measurements, instruments, scales or definitions (S2 Table). Included evaluation studies were mainly published between 2014 and 2019 (Fig 1), included a median of 68 studies (Table 1) and covered topics across 24 medical areas (S3 Table). Around two-thirds (63/93) of the evaluation studies set out to assess outcome consistency or need for a COS in their objectives and around half (48/93) could have been linked to a COS project. Almost 80% (73/93) described their design as a “systematic review” (Table 1).

Fig 1. Number of evaluation studies of outcome consistency in clinical studies over time.

Fig 1

Table 1. Characteristics of included evaluation studies of outcome consistency.

Characteristic N Descriptive
Number of clinical studies per evaluation (median, Q1-3) 93 68 (34–133)
Distinction between primary and subsequent publications, n (%) 93 30 (32)
Objective(s) of evaluation study 93
To assess a need for COS or consistency in outcome selection, n (%) 63 (67)
To identify outcomes for Delphi survey in COS development, n (%) 18 (47)
To explore outcome-reporting, n (%) 13 (14)
Evaluation study linked with a core outcome set project 93 48 (51)
Evaluation study labelled as a “systematic review” 93
Yes, n (%) 73 (79)
No, n (%) 20 (21)
When the evaluation study was not labelled as “systematic review”, what other term was used to describe the design: 20
“Literature review” 11 (55)
“Review of outcomes” 2 (10)
“Analysis of outcome reporting” 1 (5)
“Electronic database search” 1 (5)
“Survey” 1 (5)
“Outcome mapping” 1 (5)
“Systematic exploration” 1 (5)
“Analysis of studies 1 (5)
“Scoping review” 1 (5)
Types of research included in the evaluation studies of outcome consistency
Primary research
Full scale clinical study (e.g. RCTs, cohort studies), n (%) 93 59 (63)
Feasibility or pilot study, n (%) 93 8 (9)
Secondary research (literature review) 93 27 (29)
When the literature review was included, the purpose was to:
Identify outcomes, n (%) 27 10 (37)
Identify clinical studies for outcome assessment, n (%) 27 15 (56)

Systematic review methodology in evaluations of outcome consistency

In most evaluation studies, we could identify a structured research question (87%, 81/93). Half of them declared having a protocol (50%, 46/93), of which 46% had been prospectively registered with PROSPERO (21/46) (S4 Table). In many, study identification (66%, 61/93) or data extraction (50%, 49/93) were carried out by two researchers (in duplicate). Authors rarely provided an exclusions list (5%, 5/93), a rationale behind the inclusion of specific study design (20%, 19/93), or performed quality assessment of included clinical studies (28%, 26/93) (Table 2). Even though in 91% of the evaluation studies (85/93) the literature search was performed in at least one bibliographical database (median of three databases per review) (S3 Table), we were able to classify only 33% (28/85) of them as comprehensive, as per AMSTAR 2 criteria, based on provided search details. The difference between evaluation studies labelled and not labelled as “systematic review” was statistically significant for the proportion of evaluation studies where study selection or data extraction were performed by at least two researchers, or the authors provided adequate description of included clinical studies (Table 2).

Table 2. Systematic review methods as specified in the subset of AMSTAR 2 items in evaluation studies of outcome consistency.

Item Assessed aspect Group Yes (n,%)* Unclear (n,%)* No (n,%)* N/A Fisher exact test (p-value)
1 Did the research questions and inclusion criteria include the components of PICO? Overall 81 (87) 9 (10) 3 (3) 0
Labelled as “Syst rev” 64 (88) 8 (11) 1 (1) 0.129
Not labelled as “Syst rev” 17 (85) 1 (5) 2 (10)
2 Did the report contain an explicit statement that the methods were established prior to the conduct of the review and did the report justify any significant deviations from the protocol? Overall 46 (50) 5 (5) 42 (45) 0
Labelled as “Syst rev” 42 (58) 4 (5) 27 (37) 0.06
Not labelled as “Syst rev” 4 (20) 1 (5) 15 (75)
3 Did the authors explain their selection of the study designs for inclusion? Overall 19 (20) 0 74 (80) 0
Labelled as “Syst rev” 17 (23) 0 56 (77) 0.346
Not labelled as “Syst rev” 2 (10) 0 18 (90)
4 Did the authors use a comprehensive literature search strategy? Overall 28 (33) 4 (5) 53 (62) 8
Labelled as “Syst rev” 23 (32) 3 (4) 43 (59) 0.200
Not labelled as “Syst rev” 5 (25) 1 (5) 10 (50)
5 Did the authors perform study selection in duplicate? Overall 61 (66) 21 (23) 10 (11) 1
Labelled as “Syst rev” 56 (77) 11 (15) 5 (7) <0.001
Not labelled as “Syst rev” 5 (25) 10 (50) 5 (25)
6 Did the authors perform data extraction in duplicate? Overall 49 (53) 32 (34) 12 (13) 0
Labelled as “Syst rev” 45 (62) 21 (29) 7 (10) 0.002
Not labelled as “Syst rev” 4 (20) 11 (55) 5 (25)
7 Did the authors provide a list of excluded studies and justify the exclusions? Overall 5 (6) 0 84 (94) 4
Labelled as “Syst rev” 4 (5) 0 66 (90) 1.00
Not labelled as “Syst rev” 1 (5) 0 18 (90)
8 Did the authors describe the included studies in adequate detail? Overall 37 (42) 2 (2) 50 (56) 4
Labelled as “Syst rev” 35 (48) 2 (3) 33 (45) 0.006
Not labelled as “Syst rev” 2 (10) 0 17 (85)
9 Did the authors assess the quality of included studies? Overall 26 (29) 0 64 (71) 3
Labelled as “Syst rev” 22 (30) 0 49 (67) 0.474
Not labelled as “Syst rev” 4 (20) 0 15 (75)

N/A, not applicable; Syst rev, systematic review;

*Percentages calculated using as denominator number without evaluation studies in N/A category;

Methods used to identify outcomes and their consistency

The number of identified outcomes was reported in 93% (86/93) of the evaluation studies with 40% (37/93) specifying from which section of the trial publication, the outcomes were extracted. In two-thirds (63/93) of the evaluation studies, the authors made the distinction between primary and secondary outcomes, of which many also described how they ascertained whether the outcome was primary or not (41/63) (Table 3). Over half of the evaluation studies (55/90) looked at individual outcomes, 20% (18/90) assessed outcome domains (i.e. groups of individual outcomes referring to the same phenomena) and the remainder (19%, 17/90) applied both approaches (Table 3). Outcomes were frequently (90%, 81/90) presented in a tabulated format, and around one fifth (21%, 19/90) used a matrix to present outcome distribution (Fig 2).

Table 3. Methods used to identify and assess outcomes in evaluation studies of outcome consistency.

Characteristic n/N* (%)
Evaluation studies reporting the number of identified outcomes (individual or domains) 86/90 (96)
 Number of extracted outcomes (individual or domains) per review (median, Q1 to 3) 80 (43 to 158)
Distinction between primary and secondary outcomes 63/93 (68)
When the distinction between the type of outcome(s) was made:
 The primary outcome(s) was clearly specified as a primary 27/63 (43)
 The primary outcome(s) was clearly specified as a primary or used in the power calculation 11/63 (17)
 The primary outcome(s) was identified using other measures (e.g. outcome mentioned in the trial title, first reported) 3/63 (5)
 There were no details of how the primary outcome(s) was identified 22/63 (35)
Approach to outcome extraction
 Outcomes mentioned anywhere in the report 22/93 (24)
 Outcomes mentioned only in methods 11/93 (12)
 Outcomes mentioned only in results 4/93 (4)
 Not specified or unclear 56/93 (60)
Quality assessment of outcome reporting 17/93 (18)
Outcomes presented
 Only individually 55/90 (61)
 Only grouped into domains** 18/90 (20)
 In both ways (individually and grouped in domains) 17/90 (19)
Presentation of extracted outcomes
 Table of outcomes (individual or domains) with the number of studies 81/90 (90)
 Matrix (outcomes per study) 20/90 (22)
 Other graphic formats (e.g. bar chart, Venn diagram, Spiral graph) 41/90 (46)
Conclusion regarding outcome consistency
 Detection of inconsistency in outcome choice 80/90 (88)
 Detection of a need for COS 74/90 (82)
  Due to identified inconsistency in outcome choice 68/90 (76)
  Due to other reasons than inconsistency in outcome choice 6/90 (7)
Conclusions of evaluations
 Wide range of identified outcomes (individual or domains) 55/90 (61)
 Difference in how, rather than which, the outcome was measured 52/90 (58)
 Infrequent reporting of outcomes relevant to patient care 37/90 (41)
 Impact of inconsistency in outcome choice on evidence synthesis 13/90 (14)
 Difference in when the outcomes were measured 12/90 (13)

* Three reports (3/93) of full core outcome sets were excluded from some assessments (N = 90) as they focused only on providing a list of identified outcomes without reporting more details (see S3 Table)

** Outcome domain, constructs used to broadly classify individual outcomes referring to the same phenomena into a group

Fig 2. Matrix of outcomes—example.

Fig 2

We found statements on the inconsistency of outcomes in conclusions of 88% (80/90) of the assessed evaluation studies. None provided a description of how this inconsistency was defined or detected. The need for a COS was declared in 82% (74/90) and frequently justified by the encountered inconsistency of assessed outcomes (76%, 68/90) (Table 3).

Discussion

The methods adopted in the evaluation studies of outcome consistency were variable, and none implemented all nine of the methodological expectations of systematic reviews as described in AMSTAR 2. The proportion of evaluation studies labelled as a “systematic review” and those that were not differed in three AMSTAR 2 items (duplicate identification of eligible studies, duplicate data extraction, and description of included clinical studies in sufficient detail). Most of the evaluations concluded that there was evidence of inconsistency in the outcomes used across clinical studies. Although none described how this inconsistency was determined, it was frequently cited to justify the need for COS development.

Our work provides the first extensive evaluation of methods used in evaluation studies of outcome consistency in clinical studies to date. It was guided by a prospectively developed protocol using broad inclusion criteria to facilitate identification of a representative sample of the studies of outcomes. Despite the benefits of publication of methodological study protocols are yet to be determined [23], their development can enhance the study’s reproducibility. [24] Our methods heavily borrow from a systematic review process; however, our work is not a systematic review and should not be judged like one. For example, we did not set out to run an extensive search across numerous databases instead of relying on the COMET database and supplemented its results with the resources of the CROWN initiative. [25] Use of such a specific source of studies might be perceived as a limiting factor leading to the omission of some relevant publications. On the other hand, the COMET database is updated periodically, using extensive search strategy, for COS-related literature. [8] In this exploratory study of published literature, we aimed to obtain a pragmatic sample of studies using relevant sources. We did not search the international prospective register of systematic reviews (PROSPERO) [26], due to lack of specific indexing of this type of evaluation studies resulting in the inability to retrieve relevant records. As only less than a quarter of the included evaluation studies registered their protocols with PROSPERO, we believe that omission of this source did not have a substantial impact on our work.

AMSTAR 2 was primarily developed to examine the robustness of systematic reviews of healthcare interventions. [27] Thus, the use of a subset of its items in this work might be contested. We decided to use AMSTAR 2 for two reasons, the COMET handbook reference to the evaluation studies of outcome consistency as “systematic reviews”, and previous evaluation of COS-related literature in the area of women’s health. [10] We selected the most relevant and applicable across a broader spectrum of study designs AMSTAR 2 items (S1 Table). Their application to all included evaluation studies allowed us to have a uniform comparison of the methods regardless of how they were labelled. Even though, a quarter of the evaluations in our sample did not self-identified as a “systematic review”, the proportion of those satisfying AMSTAR 2 criteria labelled and not labelled as “systematic review” did not differ in six out of nine evaluated items.

Publication of the COMET Handbook [13] and related resources [14, 15, 28] increased access to guidance on robust methods for COS development. However, advice on methodology to evaluate outcomes to inform COS development is currently limited to a suggestion of adopting systematic review methodology. [13, 14] Evaluation studies of outcome consistency are a relatively new phenomenon with only a handful of this type of studies published before 2010 (Fig 1). Thus, the observed differences in the adopted methods probably arise from this limited guidance. In recent years, there has been a growing body of literature focusing on the methodological aspects of medical research [29]. The methodology of methods-orientated studies is inherently heterogeneous [29] and may involve features of a systematic review [21]. Nevertheless, the authors should refrain from labelling their work as “systematic review” when this is not the case [21], ensuring that their methodological explorations are guided by a robust methodology.

A well-executed systematic review is a resource-intensive and meticulous synthesis of available evidence. [16] Application of systematic review methods to assess outcome consistency raises a question over a value of applying such workload intense methodology—mainly in the form of researchers’ time. Evaluation studies included in this work frequently aimed to assess the need for a COS and, simultaneously, collect outcomes to inform a subsequent Delphi survey. Delphi process tends to comprise multiple sources of information for generating the initial list of outcomes with a literature assessment of outcome being only one of them. Delphi enables involved stakeholders to propose relevant outcomes not yet captured on the list. [30] Thus, the rationale for conducting a systematic review to inform this process is unclear. Given that the impact of the methods of discussed evaluation studies of outcome consistency on the COS development process is unknown; authors should consider whether a systematic review approach is warranted.

None of the assessed evaluation studies specified how inconsistency of outcomes was detected; yet, the majority claimed in their conclusion to have identified such inconsistency. This finding highlights the need for more transparency in the design and reporting of these type of evaluations and is consistent with the conclusion of a recent COS literature. In their work, Young et al. examined number of reported outcomes, their definitions, timing, and approach to their grouping in 132 studies (development papers, protocols, and reviews) and found meaningful differences. [31] Based on their findings, they proposed a definition for a unique outcome and inconsistency of outcomes referred to in their work as outcome reporting heterogeneity (ORH). [31] The proposed definition brings attention to a difference between the number of identified outcomes (however defined) and their overlap between the clinical studies.

The methods of evaluation studies of outcome consistency published to date differ in their approach to the identification of eligible studies and their outcomes. Although frequently labelled as “systematic review”, their methods seldom satisfied the standards set for systematic reviews in the aspects of the process described by the first nine items of AMSTAR 2. In the absence of evidence of a benefit of adopting the systematic review approach in this context, we encourage researchers embarking on these type of project to follow general principles of systematic review practice. This should include the development of a research protocol with a clear description of intended methods and how researchers plan to determine if the main objective has been achieved and transparent reporting of the study. [32, 33] However, we discourage labelling this type of evaluation, as a “systematic review”. Instead, we propose the use of terms such as “evaluation of outcomes” or “outcome mapping.” Alternatively, authors may choose to use an established approach, such as “scoping review” and adopt the appropriate methodological and reporting standards [34, 35].

Supporting information

S1 Fig. Study selection flow diagram.

(DOCX)

S1 Text. Data collection form.

(DOCX)

S1 Table. Description of how A MeaSurement Tool to Assess systematic Reviews (AMSTAR 2) items were applied.

(DOCX)

S2 Table. List of assessed records with exclusion reasons.

(XLSX)

S3 Table. Detailed characteristic of included evaluation studies of outcome consistency.

(DOCX)

S4 Table. Details of identification of clinical study reports in the evaluation studies of outcome consistency.

(DOCX)

Acknowledgments

We would like to thank Professor Paula R. Williamson for her helpful comments at the initial stages of this work.

Data Availability

All relevant data are within the paper and its Supporting Information files.

Funding Statement

ER and CLV were supported by the UK Medical Research Council (MC_UU_12023/24).

References

  • 1.Kirkham JJ, Dwan KM, Altman DG, Gamble C, Dodd S, Smyth R, et al. The impact of outcome reporting bias in randomised controlled trials on a cohort of systematic reviews. BMJ. 2010;340:c365 10.1136/bmj.c365 [DOI] [PubMed] [Google Scholar]
  • 2.Williamson PR, Gamble C. Identification and impact of outcome selection bias in meta-analysis. Statist Med. 2005;24:1547–61. [DOI] [PubMed] [Google Scholar]
  • 3.Mayo-Wilson E, Fusco N, Li T, Hong H, Canner JK, Dickersin K, et al. Multiple outcomes and analyses in clinical trials create challenges for interpretation and research synthesis. Journal of clinical epidemiology. 2017;86:39–50. 10.1016/j.jclinepi.2017.05.007 [DOI] [PubMed] [Google Scholar]
  • 4.JL H, PR W. Bias in meta-analysis due to outcome variable selection within studies. Appl Stat. 2000;49:359–70. [Google Scholar]
  • 5.Dwan K, Altman DG, Arnaiz JA, Bloom J, Chan AW, Cronin E, et al. Systematic review of the empirical evidence of study publication bias and outcome reporting bias. PLoS One. 2008;3(8):e3081 10.1371/journal.pone.0003081 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6.Williamson P, Altman D, Blazeby J, Clarke M, Gargon E. Driving up the quality and relevance of research through the use of agreed core outcomes. J Health Serv Res Policy. 2012;17(1):1–2. 10.1258/jhsrp.2011.011131 [DOI] [PubMed] [Google Scholar]
  • 7.Clarke M. Standardising outcomes for clinical trials and systematic reviews. Trials. 2007;8:39 10.1186/1745-6215-8-39 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.Gargon E, Gorst SL, Harman NL, Smith V, Matvienko-Sikar K, Williamson PR. Choosing important health outcomes for comparative effectiveness research: 4th annual update to a systematic review of core outcome sets for research. PLoS One. 2018;13(12):e0209869 10.1371/journal.pone.0209869 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9.Gargon E, Gurung B, Medley N, Altman DG, Blazeby JM, Clarke M, et al. Choosing important health outcomes for comparative effectiveness research: a systematic review. PLoS One. 2014;9(6):e99111 10.1371/journal.pone.0099111 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10.Duffy JMN, Rolph R, Gale C, Hirsch M, Khan KS, Ziebland S, et al. Core Outcome Sets in Women’s and Newborn Health: A Systematic Review. BJOG. 2017. [DOI] [PubMed] [Google Scholar]
  • 11.Sinha I, Jones L, Smyth RL, Williamson PR. A systematic review of studies that aim to determine which outcomes to measure in clinical trials in children. PLoS Med. 2008;5(4):e96 10.1371/journal.pmed.0050096 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12.Gargon E. Developing the agenda for core outcome set development: University of Liverpool; 2016. [Google Scholar]
  • 13.Williamson PR, Altman DG, Bagley H, Barnes KL, Blazeby JM, Brookes ST, et al. The COMET Handbook: version 1.0. Trials. 2017;18(Suppl 3):280 10.1186/s13063-017-1978-4 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14.Williamson PR, Altman DG, Blazeby JM, Clarke M, Devane D, Gargon E, et al. Developing core outcome sets for clinical trials: issues to consider. Trials. 2012;13:132 10.1186/1745-6215-13-132 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15.Kirkham JJ, Davis K, Altman DG, Blazeby JM, Clarke M, Tunis S, et al. Core Outcome Set-STAndards for Development: The COS-STAD recommendations. PLoS Med. 2017;14(11):e1002447 10.1371/journal.pmed.1002447 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16.Sutton AJ, Cooper NJ, Jones DR. Evidence synthesis as the key to more coherent and efficient research. BMC medical research methodology. 2009;9:29 10.1186/1471-2288-9-29 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17.Evidence-Based Medicine Working Group. Evidence-based medicine. A new approach to teaching the practice of medicine. JAMA. 1992;268(17):2420–5. 10.1001/jama.1992.03490170092032 [DOI] [PubMed] [Google Scholar]
  • 18.Clarke M. History of evidence synthesis to assess treatment effects: Personal reflections on something that is very much alive. J R Soc Med. 2016. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19.Clayton GL, Smith IL, Higgins JPT, Mihaylova B, Thorpe B, Cicero R, et al. The INVEST project: investigating the use of evidence synthesis in the design and analysis of clinical trials. Trials. 2017;18(1):219 10.1186/s13063-017-1955-y [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 20.Bown MJ, Sutton AJ. Quality control in systematic reviews and meta-analyses. European journal of vascular and endovascular surgery: the official journal of the European Society for Vascular Surgery. 2010;40(5):669–77. [DOI] [PubMed] [Google Scholar]
  • 21.Puljak L. Methodological studies evaluating evidence are not systematic reviews. Journal of clinical epidemiology. 2019;110:98–9. 10.1016/j.jclinepi.2019.02.002 [DOI] [PubMed] [Google Scholar]
  • 22.CROWNinitiative. Core Outcomes in Women’s and Newborn Health London: The Royal College of Obstetricians and Gynaecologists; 2014 http://www.crown-initiative.org/.
  • 23.Puljak L, Pieper D. Registration of methodological studies, that is, "research-on-research" studies-should it be mandatory? Journal of clinical epidemiology. 2019;115:35–6. 10.1016/j.jclinepi.2019.06.017 [DOI] [PubMed] [Google Scholar]
  • 24.Murad MH, Wang Z. Guidelines for reporting meta-epidemiological methodology research. Evid Based Med. 2017;22(4):139–42. 10.1136/ebmed-2017-110713 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25.Khan K. The CROWN Initiative: journal editors invite researchers to develop core outcomes in women’s health. BJOG. 2014;121(10):1181–2. 10.1111/1471-0528.12929 [DOI] [PubMed] [Google Scholar]
  • 26.Chien PF, Khan KS, Siassakos D. Registration of systematic reviews: PROSPERO. BJOG. 2012;119(8):903–5. 10.1111/j.1471-0528.2011.03242.x [DOI] [PubMed] [Google Scholar]
  • 27.Shea BJ, Reeves BC, Wells G, Thuku M, Hamel C, Moran J, et al. AMSTAR 2: a critical appraisal tool for systematic reviews that include randomised or non-randomised studies of healthcare interventions, or both. BMJ. 2017;358:j4008 10.1136/bmj.j4008 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 28.Kirkham JJ, Gorst S, Altman DG, Blazeby JM, Clarke M, Tunis S, et al. Core Outcome Set-STAndardised Protocol Items: the COS-STAP Statement. Trials. 2019;20(1):116 10.1186/s13063-019-3230-x [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 29.Ioannidis JP, Fanelli D, Dunne DD, Goodman SN. Meta-research: Evaluation and Improvement of Research Methods and Practices. PLoS Biol. 2015;13(10):e1002264 10.1371/journal.pbio.1002264 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 30.Sinha IP, Smyth RL, Williamson PR. Using the Delphi technique to determine which outcomes to measure in clinical trials: recommendations for the future based on a systematic review of existing studies. PLoS Med. 2011;8(1):e1000393 10.1371/journal.pmed.1000393 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 31.Young AE, Brookes ST, Avery KNL, Davies A, Metcalfe C, Blazeby JM. A systematic review of core outcome set development studies demonstrates difficulties in defining unique outcomes. Journal of clinical epidemiology. 2019;115:14–24. 10.1016/j.jclinepi.2019.06.016 [DOI] [PubMed] [Google Scholar]
  • 32.Ioannidis JP. How to make more published research true. PLoS Med. 2014;11(10):e1001747 10.1371/journal.pmed.1001747 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 33.Ioannidis JPA, Greenland S, Hlatky MA, Khoury MJ, Macleod MR, Moher D, et al. Increasing value and reducing waste in research design, conduct, and analysis. The Lancet. 2014;383(9912):166–75. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 34.Peters MD, Godfrey CM, Khalil H, McInerney P, Parker D, Soares CB. Guidance for conducting systematic scoping reviews. Int J Evid Based Healthc. 2015;13(3):141–6. 10.1097/XEB.0000000000000050 [DOI] [PubMed] [Google Scholar]
  • 35.Tricco AC, Lillie E, Zarin W, O’Brien KK, Colquhoun H, Levac D, et al. PRISMA Extension for Scoping Reviews (PRISMA-ScR): Checklist and Explanation. Ann Intern Med. 2018;169(7):467–73. 10.7326/M18-0850 [DOI] [PubMed] [Google Scholar]

Decision Letter 0

Tim Mathes

26 May 2020

PONE-D-20-08427

Evaluation of methods in studies of outcomes consistency in clinical studies

PLOS ONE

Dear Dr. Rogozińska,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

You can ignore the comment on highlighting  statistical significance (reviewer #2). 

Please submit your revised manuscript by Jul 10 2020 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: http://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols

We look forward to receiving your revised manuscript.

Kind regards,

Tim Mathes

Academic Editor

PLOS ONE

Journal requirements:

When submitting your revision, we need you to address these additional requirements:

1.    Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at http://www.plosone.org/attachments/PLOSOne_formatting_sample_main_body.pdf and http://www.plosone.org/attachments/PLOSOne_formatting_sample_title_authors_affiliations.pdf

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

Reviewer #2: Yes

**********

2. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

Reviewer #2: Yes

**********

3. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: Yes

**********

4. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

**********

5. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: The authors propose to evaluate the quality of the methods used in studies on outcomes consistencies. The topic is important and has not been investigated. However a few point could be improved

Major comments:

1) The author declare that their study was not a systematic review and reported having searched in the “COMET Database”. It is not clear what this COMET database is and the reference 8 does not mention it clearly either (the weblink of reference 8 is broken : http://www.comet-initiative.org/studies/search) It is time consuming for the reader to find out how this COMET database is provided (medline and scopus mainly? ).

ð In the method section, the authors should explain with a few sentences how the COMET database is fuelled.

2) The title of the article does not give any idea about the method they used. We propose to add more precision about the design of their study as a scoping review see the reference of the Joanna briggs institute (https://wiki.joannabriggs.org/display/MANUAL/Chapter+11%3A+Scoping+reviews) or of the PRISMA group (http://www.prisma-statement.org/Extensions/ScopingReviews)

ð “Evaluation of methods in studies of outcomes consistency in clinical studies: a scoping review”

3) At the end of the discussion, the author recommend not to use the term “systematic review” when studies are not : “However, we discourage labelling for this type of evaluation, as “systematic review” instead, proposing alternative terms such as an “evaluation of outcomes” or “outcome mapping”.”

ð We also suggest to use the term of “scoping review” in order to standardized the terminology according to PRISMA and the Joanna Briggs institute.

Minor comments:

Table 1 : we do not understand what “total number of participants per review” does mean. Who are these participants ? Is it the number of reviewers involved in each study or the number of patients included in the studies evaluated by the review? In case of the later, how is it relevant for the current study?

Reviewer #2: The abstract is not fully understandable without reading the main text. You should state concisely which type of studies are included in your review (e.g. synthesis studies evaluating outcome consistency in clinical studies). Please do not use undefined abbreviations (AMSTAR).

Please provide a brief description of AMSTAR in the main text.

Table 1: please move the notes into the table since they include relevant information

Table 2: please explain why you used Fisher's exact test (instead of chi-squared) and to which groups of studies you applied the test; please also highlight significant results (p<0.05).

Table 3: it is not understandable why denominators differ across the rows (90 or 93).

Tables (in general): when you use the word "studies", please verify whether it is easy to distinguish between studies included in your review and clinical studies included in the synthesis studies you examined

Please check all tables and figure captions and see if they can be more informative of the underlying content.

Please discuss any practical implications of your work (such as improvements in clinical research and practice).

Please identify any health area that requires more attention in future research.

**********

6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

PLoS One. 2020 Jul 8;15(7):e0235485. doi: 10.1371/journal.pone.0235485.r002

Author response to Decision Letter 0


2 Jun 2020

Reviewer #1

1. The authors propose to evaluate the quality of the methods used in studies on outcomes consistencies. The topic is important and has not been investigated. However a few point could be improved

We thank the reviewer for the recognition of the importance of this topic.

2. The author declare that their study was not a systematic review and reported having searched in the “COMET Database”. It is not clear what this COMET database is and the reference 8 does not mention it clearly either (the weblink of reference 8 is broken: http://www.comet-initiative.org/studies/search) It is time consuming for the reader to find out how this COMET database is provided (medline and scopus mainly?). In the method section, the authors should explain with a few sentences how the COMET database is fuelled.

We agree with the reviewer that we had not adequately described the COMET database and have now addressed this shortcoming by adding the following description in the methods:

“We searched the COMET database from its inception to June 2018; the search was updated in May 2019. (8) The COMET database is an annually updated repository of the international COS literature based on the systematic searches run in MEDLINE, SCOPUS, and Cochrane Methodology Register. (9)

3. The title of the article does not give any idea about the method they used. We propose to add more precision about the design of their study as a scoping review see the reference of the Joanna Briggs Institute (https://wiki.joannabriggs.org/display/MANUAL/Chapter+11%3A+Scoping+reviews) or of the PRISMA group (http://www.prisma-statement.org/Extensions/ScopingReviews)

“Evaluation of methods in studies of outcomes consistency in clinical studies: a scoping review”

We thank the reviewer for the suggestion. However, our work was not set out to be a scoping review; thus we feel that labelling its design this way would be inappropriate. According to our protocol (see the PROSPERO registration) the design was described as “exploratory study” and was developed in the spirit of meta-research; discipline aiming to evaluate and improve research practices (Ioannidis et al. 2015 PLoS Biol. 13(10): e1002264). We have now added the information about the study design in the title as follows: “Methods to assess outcome consistency in clinical research: literature-baseded evaluation”

4. At the end of the discussion, the author recommend not to use the term “systematic review” when studies are not: “However, we discourage labelling for this type of evaluation, as “systematic review” instead, proposing alternative terms such as an “evaluation of outcomes” or “outcome mapping”.” We also suggest to use the term of “scoping review” in order to standardized the terminology according to PRISMA and the Joanna Briggs institute.

We agree that a scoping review may be another alternative and have amended the last sentence in our final paragraph as follows:

“However, we discourage labelling for this type of evaluation, as “systematic review.” Instead, we propose terms such as an “evaluation of outcomes”, “outcome mapping.” Alternatively, authors ma choose to use an established approach, such as “scoping review”, and adopt the appropriate methodological and reporting standards (34, 35).”

5. Table 1: we do not understand what “total number of participants per review” does mean. Who are these participants? Is it the number of reviewers involved in each study or the number of patients included in the studies evaluated by the review? In case of the later, how is it relevant for the current study?

On reflection we agree with the reviewer and have now removed this information from Table 1.

Reviewer #2

1. The abstract is not fully understandable without reading the main text. You should state concisely which type of studies are included in your review (e.g. synthesis studies evaluating outcome consistency in clinical studies). Please do not use undefined abbreviations (AMSTAR).

We have now amended the abstract and checked the manuscript for consistency in the used terminology referring to included studies as “evaluations of outcome consistency in clinical studies”

AMSTAR abbreviation has been explained in the abstract as follows: “The methods were assessed using a subset of A MeaSurement Tool to Assess systematic Reviews (AMSTAR) 2 items (1-9) and additional outcome-specific questions.”

2. Please provide a brief description of AMSTAR in the main text.

We have now added a following description in the methods section: “In order to examine adoption of systematic review methodology, all studies (regardless of declared study design) were assessed against a tailored subset of A MeaSurement Tool to Assess systematic Reviews (AMSTAR) 2 items (23) we felt were relevant to evaluated studies (S1 Table). AMSTAR was designed as a practical critical appraisal instrument enabling a rapid and reproducible quality assessments of systematic reviews.”

3. Table 1: please move the notes into the table since they include relevant information

The notes have been now incorporated into the main body of Table 1.

4. Table 3: it is not understandable why denominators differ across the rows (90 or 93).

The explanation of the difference is provided at the bottom of the table and has been rewritten for clarity: “*Three reports (3/93) of full core outcome sets were excluded from some assessments due to none applicability (for details see S3 Table)”. We also moved asterisks to the column heading (n/N*).

5. Tables (in general): when you use the word "studies", please verify whether it is easy to distinguish between studies included in your review and clinical studies included in the synthesis studies you examined

We have now amended all tables to improve their clarity according to the reviewer’s suggestion.

6. Please check all tables and figure captions and see if they can be more informative of the underlying content.

We have now checked all tables and figures to improve their clarity according to this suggestion.

7. Please discuss any practical implications of your work (such as improvements in clinical research and practice).

The practical implications of our work are discussed in the final paragraph of our discussion, notably that authors of these studies need to be transparent in reporting their methods and that they should not label such evaluations as systematic reviews if they do not fully adhere to stringent systematic review methodologies.

Although we do not feel that our work has any direct impact on clinical practice; we hope that it will improve research practice relating to core outcome set development and reporting of such research.

8. Please identify any health area that requires more attention in future research.

As our work focuses on the methods rather than a specific clinical issue and is applicable across all health areas. We have recorded and reported the health areas in which evaluations of outcome consistency were carried out, however, we have not compared methodological practices between different health areas.

Attachment

Submitted filename: 20Jun01 Response.docx

Decision Letter 1

Tim Mathes

5 Jun 2020

PONE-D-20-08427R1

Methods to assess outcome consistency in clinical research: a literature-based evaluation

PLOS ONE

Dear Dr. Rogozińska,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

Please submit your revised manuscript by Jul 20 2020 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: http://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols

We look forward to receiving your revised manuscript.

Kind regards,

Tim Mathes

Academic Editor

PLOS ONE

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation.

Reviewer #1: All comments have been addressed

Reviewer #2: (No Response)

**********

2. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

Reviewer #2: Yes

**********

3. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

Reviewer #2: N/A

**********

4. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: Yes

**********

5. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

**********

6. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: (No Response)

Reviewer #2: Thanks for addressing my previous comments. I have few additional minor suggestions:

1) replace "evaluations of outcome consistency" with "evaluation studies of outcome consistency"

2) if possibile, include such denomination among the keywords

3) check Table 1 caption (it should be outcome "consistency")

4) Table 3 caption: also in this case I would specify that methods are those of the included studies

5) Table 3: the third category ("when the distinction was made..") is a sub-category of the second one, and the way of reporting should reflect this

6) Table 3: the note is still unclear (you should fully understand the table without looking at supplementary files)

7) S3 Table caption: it should be "detailed characteristics" instead of "detail characteristics"

8) S4 Table caption is unclear: do you mean the clinical studies, or the evaluation studies that you assessed?

**********

7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

PLoS One. 2020 Jul 8;15(7):e0235485. doi: 10.1371/journal.pone.0235485.r004

Author response to Decision Letter 1


14 Jun 2020

1 Thanks for addressing my previous comments. I have few additional minor suggestions:

We thank the reviewer for his further suggestions. We have reviewed all the points and amended the manuscript accordingly.

2 replace "evaluations of outcome consistency" with "evaluation studies of outcome consistency" if possibile, include such denomination among the keywords

The term "evaluations of outcome consistency" has been changed to "evaluation studies of outcome consistency" and listed the term among the keywords.

3 check Table 1 caption (it should be outcome "consistency")

The caption of Table 1 has been corrected to: “Characteristics of included evaluations of outcome consistency”

4 Table 3 caption: also in this case I would specify that methods are those of the included studies

The caption of Table 3 has been changed to: “Methods used to identify and assess outcomes in evaluation studies of outcome consistency”

5 Table 3: the third category ("when the distinction was made..") is a sub-category of the second one, and the way of reporting should reflect this

The category and subcategories have been amended as follows:

When the distinction between the type of outcome(s) was made:

• The primary outcome(s) was clearly specified as a primary

• The primary outcome(s) was clearly specified as a primary or used in the power calculation

• The primary outcome(s) was identified using other measures (e.g. outcome mentioned in the trial title, first reported)

• There were no details of how the primary outcome(s) was identified

6 Table 3: the note is still unclear (you should fully understand the table without looking at supplementary files)

The note has been amended as follows:

“Three reports (3/93) of full core outcome sets were excluded from some assessments (N = 90) as they focused only on providing a list of identified outcomes without reporting more details (see S3 Table)”

7 S3 Table caption: it should be "detailed characteristics" instead of "detail characteristics"

The caption of S3 Table has been changed to: “Detailed characteristics of included evaluation studies of outcome consistency”

8) S4 Table caption is unclear: do you mean the clinical studies, or the evaluation studies that you assessed?

The caption of S4 Table has been changed to: “Details of identification of clinical study reports in the evaluation studies of outcome consistency”

Attachment

Submitted filename: 20Jun14 Response to reviewers.docx

Decision Letter 2

Tim Mathes

17 Jun 2020

Methods used to assess outcome consistency in clinical studies: a literature-based evaluation

PONE-D-20-08427R2

Dear Dr. Rogozińska,

We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements.

Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication.

An invoice for payment will follow shortly after the formal acceptance. To ensure an efficient process, please log into Editorial Manager at http://www.editorialmanager.com/pone/, click the 'Update My Information' link at the top of the page, and double check that your user information is up-to-date. If you have any billing related questions, please contact our Author Billing department directly at authorbilling@plos.org.

If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

Kind regards,

Tim Mathes

Academic Editor

PLOS ONE

Additional Editor Comments (optional):

Reviewers' comments:

Acceptance letter

Tim Mathes

22 Jun 2020

PONE-D-20-08427R2

Methods used to assess outcome consistency in clinical studies: a literature-based evaluation

Dear Dr. Rogozińska:

I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now with our production department.

If your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information please contact onepress@plos.org.

If we can help with anything else, please email us at plosone@plos.org.

Thank you for submitting your work to PLOS ONE and supporting open access.

Kind regards,

PLOS ONE Editorial Office Staff

on behalf of

Dr. Tim Mathes

Academic Editor

PLOS ONE

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    S1 Fig. Study selection flow diagram.

    (DOCX)

    S1 Text. Data collection form.

    (DOCX)

    S1 Table. Description of how A MeaSurement Tool to Assess systematic Reviews (AMSTAR 2) items were applied.

    (DOCX)

    S2 Table. List of assessed records with exclusion reasons.

    (XLSX)

    S3 Table. Detailed characteristic of included evaluation studies of outcome consistency.

    (DOCX)

    S4 Table. Details of identification of clinical study reports in the evaluation studies of outcome consistency.

    (DOCX)

    Attachment

    Submitted filename: 20Jun01 Response.docx

    Attachment

    Submitted filename: 20Jun14 Response to reviewers.docx

    Data Availability Statement

    All relevant data are within the paper and its Supporting Information files.


    Articles from PLoS ONE are provided here courtesy of PLOS

    RESOURCES