Table 1.
Inter-rater reliability
Items | Kappa (95% CI) |
---|---|
1. Was an “a priori” design provided? | 0.65 (0.54, 0.76) |
2. Was there duplicate study selection and data extraction? | 0.94 (0.91, 0.96) |
3. Was a comprehensive literature search performed? | 0.67 (0.62, 0.72) |
4. Was the status of publication (i.e., grey literature) used as an inclusion criterion? | 0.85 (0.81, 0.88) |
5. Was a list of studies (included and excluded) provided? | 0.75 (0.71, 0.79) |
6. Were the characteristics of the included studies provided? | 0.91 (0.88, 0.94) |
7. Was the scientific quality of the included studies assessed and documented? | 0.61 (0.56, 0.66) |
8. Was the scientific quality of the included studies used appropriately in formulating conclusions? | 0.57 (0.51, 0.62) |
9. Were the methods used to combine the findings of studies appropriate? | 0.53 (0.48, 0.57) |
10. Was the likelihood of publication bias assessed? | 1 |
11. Were potential conflicts of interest included? | 0.58 (0.54, 0.63) |
CI Confidence interval.