Table 2.
Dimension and/or tool | Authors | Construct(s), as described by the authors | Target population | Domains, nr. of items | Response options | Development and validation |
---|---|---|---|---|---|---|
“Applicability”-dimension of LEGEND | Clark et al. [77] | Applicability of results to treating patients | P1: RCTs and CCTs P2: reviewers and clinicians | 3 items | 3-point-scale | Deductive and inductive item-generation. Tool was pilot tested among an interprofessional group of clinicians. |
“Applicability”-dimension of Carr´s evidence-grading scheme | Carr et al. [63] | Generalizability of study population |
P1: clinical trials P:2 authors of SRs |
1 item | 3-point-classification-scale | No specific information on tool development. |
Bornhöft´s checklist | Bornhöft et al. [78] | External validity (EV) and Model validity (MV) of clinical trials |
P1: clinical trials P2: authors of SRs |
4 domains with 26 items for EV and MV each | 4-point-scale | Development with a comprehensive, deductive item-generation from the literature. Pilot-tests were performed, but not for the whole scales. |
Cleggs´s external validity assessment | Clegg et al. [64] | Generalizability of clinical trials to England and Wales |
P1: clinical trials P2: authors of SRs and HTAs |
5 items | 3-point-scale | No specific information on tool development |
Clinical applicability | Haraldsson et al. [66] | Report quality and applicability of intervention, study population and outcomes |
P1: RCTs P2: reviewers |
6 items | 3-point-scale and 4-point-scale | No specific information on tool development |
Clinical Relevance Instrument | Cho & Bero [79] | Ethics and Generalizability of outcomes, subjects, treatment and side effects |
P1: clinical trials P2: reviewers |
7 items | 3-point-scale |
Tool was pilot tested on 10 drug studies. Content validity was confirmed by 7 reviewers with research experience. - interrater reliability: ICC = 0.56 (n = 127) [80] |
“Clinical Relevance” according to the CCBRG | Van Tulder et al. [81] | Applicability of patients, interventions and outcomes |
P1: RCTs P2: authors of SRs |
5 items | 3-point-scale (Staal et al., 2008) | Deductive item-generation for Clinical Relevance. Results were discussed in a workshop. After two rounds, a final draft was circulated for comments among editors of the CCBRG. |
Clinical Relevance Score | Karjalainen et al. [68] | Report quality and applicability of results |
P1: RCTs P2: reviewers |
3 items | 3-point-scale | No specific information on tool development. |
Estrada´s applicability assessment criteria | Estrada et al. [82] | Applicability of population, intervention, implementation and environmental context to Latin America |
P1: RCTs P2: reviewers |
5 domains with 8 items | 3-point-scale for each domain | Deductive item generation from the review by Munthe-Kaas et al. [17]. Factors and items were adapted, and pilot tested by the review team (n = 4) until consensus was reached. |
EVAT (External Validity Assessment Tool) | Khorsan & Crawford [83] | External validity of participants, intervention, and setting |
P1: RCTs and non-randomized studies P2: reviewers |
3 items | 3-point-scale | Deductive item-generation. Tool developed based on the GAP-checklist [76] and the Downs and Black-checklist [22]. Feasibility was tested and a rulebook was developed but not published. |
“External validity”-dimension of the Downs & Black-Checklist | Downs & Black [22] | Representativeness of study participants, treatments and settings to source population or setting |
P1: RCTs and non-randomised studies P2: reviewers |
3 items | 3-point-scale |
Deductive item-generation, pilot test and content validation of pilot version. Final version tested for: - internal consistency: KR-20 = 0.54 (n = 20), - reliability: test-retest: k = -0.05-0.48 and 10–15% disagreement (measurement error) (n = 20), [22] interrater reliability: k = -0.08-0.00 and 5–20% disagreement (measurement error) (n = 20) [22]; ICC = 0.76 (n = 20) [84] |
“External validity”-dimension of Foy´s quality checklist | Foy et al. [65] | External validity of patients, settings, intervention and outcomes |
P1: intervention studies P2: reviewers |
6 items | not clearly described | Deductive item-generation. No further information on tool development. |
“External validity”-dimension of Liberati´s quality assessment criterias | Liberati et al. [69] | Report quality and generalizability |
P1: RCTS P2: reviewers |
9 items | dichotomous and 3-point-scale | Tool is a modified version of a previously developed checklist [85] with additional inductive item-generation. No further information on tool development. |
“External validity”-dimension of Sorg´s checklist | Sorg et al. [71] | External validity of population, interventions, and endpoints |
P1: RCTs P2: reviewers |
4 domains with 11 items | not clearly described | Developed based on Bornhöft et al. [78] No further information on tool development. |
“external validity”-criteria of the USPSTF | USPSTF Procedure manual [73] | Generalizability of study population, setting and providers for US primary care |
P1: clinical studies P2: USPSTF reviewers |
3 items |
Sum-score- rating: 3-point-scale |
Tool developed for USPSTF reviews. No specific information on tool development. - interrater reliability: ICC = 0.84 (n = 20) [84] |
FAME (Feasibility, Appropriateness, Meaningfulness and Effectiveness) scale | Averis et al. [70] | Grading of recommendation for applicability and ethics of intervention |
P1: intervention studies P2: reviewers |
4 items | 5-point-scale | The FAME framework was created by a national group of nursing research experts. Deductive and inductive item-generation. No further information on tool development. |
GAP (Generalizability, Applicability and Predictability) checklist |
Fernandez-Hermida et al. [76] |
External validity of population, setting, intervention and endpoints |
P1: RCTs P2: Reviewers |
3 items | 3-point-scale | No specific information on tool development. |
Gartlehner´s tool | Gartlehner et al. [86] | To distinguish between effectiveness and efficacy trials |
P1: RCTs P2: reviewers |
7 items | Dichotomous |
Deductive and inductive item-generation. - criterion validity testing with studies selected by 12 experts as gold standard.: specificity = 0.83, sensitivity = 0.72 (n = 24) - measurement error: 78.3% agreement (n = 24) - interrater reliability: k = 0.42 (n = 24) [86]; k = 0.11–0.81 (n = 151) [87] |
Green & Glasgow´s external validity quality rating criteria | Green & Glasgow [88] | Report quality for generalizability | P1: trials (not explicitly described) P2: reviewers | 4 Domains with 16 items | Dichotomous |
Deductive item-generation. Mainly based on the Re-Aim framework.[89] - interrater reliability: ICC = 0.86 (n = 14) [90] - discriminative validity: TREND studies report on 77% and non-TREND studies report on 54% of scale items (n = 14) [90] - ratings across included studies (n = 31) [91], no hypothesis was defined |
“Indirecntess”-dimension of the GRADE handbook | Schünemann et al. [92] | Differences of population, interventions, and outcome measures to research question |
P1: intervention studies P2: authors of SRs, clinical guidelines and HTAs |
4 items |
Overall: 3-point-scale (downgrading options) |
Deductive and inductive item-generation, pilot-testing with 17 reviewers (n = 12) [48]. - interrater reliability: ICC = 0.00–0.13 (n > 100) [93] |
Loyka´s external validity framework | Loyka et al. [75] |
Report quality for generalizability of research in psychological science |
P1: intervention studies P2: researchers |
4 domains with 15 items |
Dichotomous |
Deductive item generation (including Green & Glasgow [88]) and adaptation for psychological science. No further information on tool development. - measurement error: 60-100% agreement (n = 143) |
Modified “Indirectness” of the Checklist for GRADE | Meader et al. [94] | Differences of population, interventions, and outcome measures to research question. |
P1: meta-analysis of RCTs P2: authors of SRs, clinical guidelines and HTAs |
5 items |
Item-level: 2-and 3-point-scale Overall: 3-point-scale (grading options) |
Developed based on GRADE method, two phase pilot-tests, - interrater reliability: kappa was poor to almost perfect on item-level [94] and k = 0.69 for overall rating of indirectness (n = 29) [95] |
external validity checklist of the NHMRC handbook | NHMRC handbook [74] | external validity of an economic study |
P1: clinical studies P2: clinical guideline developers, reviewers |
6 items | 3-point-scale | No specific information on tool development. |
revised GATE in NICE manual (2012) | NICE manual [72] | Generalizability of population, interventions and outcomes |
P1: intervention studies P2: reviewers |
2 domains with 4 items | 3-point-scale and 5-point-scale | Based on Jackson et al. [96] No specific information on tool development. |
RITES (Rating of Included Trials on the Efficacy-Effectiveness Spectrum) | Wieland et al. [47] | To characterize RCTs on an efficacy-effectiveness continuum. |
P1: RCTs P2: reviewers |
4 items | 5-point-likert-scale |
Deductive and inductive item-generation, modified Delphi procedure with 69–72 experts, pilot testing in 4 Cochrane reviews, content validation with Delphi procedure and core expert group (n = 14) [47], - interrater reliability: ICC = 0.54-1.0 (n = 22) [97] - convergent validity with PRECIS 2 tool: r = 0.55 correlation (n = 59) [97] |
Section A (Selection Bias) of EPHPP (Effective Public health Practice Project) tool | Thomas et al. [98] | Representativeness of population and participation rate. |
P1: clinical trials P2: reviewers |
2 items |
Item-level: 4-point-scale and 5-point-scale Overall: 3-point-scale |
Deductive item-generation, pilot-tests, content validation by 6 experts, - convergent validity with Guide to Community Services (GCPS) instrument: 52.5–87.5% agreement (n = 70) [98] - test-retest reliability: k = 0.61–0.74 (n = 70) [98] k = 0.60 (n = 20) [99] |
Section D of the CASP checklist for RCTs | CASP Programme [100] | Applicability to local population and outcomes |
P1: RCTs P2: participants of workshops, reviewers |
2 items | 3-point-scale | Deductive item-generation, development and pilot-tests with group of experts. |
Whole Systems research considerations´ checklist | Hawk et al. [67] | Applicability of results to usual practice | P1: RCTs P2: Reviewers (developed for review) | 7 domains with 13 items |
Item-level: dichotomous Overall: 3-point-scale |
Deductive item-generation. No specific information on tool development. |
Abbreviations: CASP Critical Appraisal Skills Programme, CCBRG Cochrane Collaboration Back Review Group, CCT controlled clinical trial, GATE Graphical Appraisal Tool for Epidemiological Studies, GRADE Grading of Recommendations Assessment, Development and Evaluation, HTA Health Technology Assessment, ICC intraclass correlation, LEGEND Let Evidence Guide Every New Decision, NICE National Institute for Health and Care Excellence, PRECIS PRagmatic Explanatory Continuum Indicator Summary, RCT randomized controlled trial, TREND Transparent Reporting of Evaluations with Nonrandomized Designs, USPSTF U.S. Preventive Services Task Force