Skip to main content
PMC Canada Author Manuscripts logoLink to PMC Canada Author Manuscripts
. Author manuscript; available in PMC: 2012 Jul 3.
Published in final edited form as: Top Spinal Cord Inj Rehabil. 2007 Summer;13(1):1–10. doi: 10.1310/sci1301-1

Spinal Cord Injury Rehabilitation Evidence: Methods of the SCIRE Systematic Review

Janice J Eng 1, Robert Teasell 2, William C Miller 3, Dalton L Wolfe 4, Andrea F Townson 5, Jo-Anne Aubut 6, Caroline Abramson 7, Jane T Hsieh 8, Sandra Connolly 9, Kristin Konnyu 10; the SCIRE Research Team
PMCID: PMC3389040  CAMSID: CAMS2211  PMID: 22767989

Abstract

The Spinal Cord Injury Rehabilitation Evidence (SCIRE) is a synthesis of the research evidence underlying rehabilitation interventions to improve the health of people living with SCI. SCIRE covers a comprehensive set of topics and in this issue we present six papers relevant to SCI rehabilitation clinicians (SCI inpatient rehabilitation practices, gait strategies, upper extremity reconstructive surgery, spasticity treatments, cardiovascular health and bone health). The SCIRE used a systematic and well-defined protocol to assess and synthesize the evidence. Each article was scored for its methodological quality using either the Physiotherapy Evidence Database (PEDro) Score for randomized controlled trials or the Downs and Black Tool for other types of studies. Following the individual study assessment, conclusions were drawn about the accumulated studies for each topic of interest based on the levels of evidence, quality of studies and concurring evidence. The SCIRE project was designed for health professionals to inform them of best practices.

Keywords: spinal cord injury, methods, systematic review, rehabilitation, evidence-based practice, interventions, knowledge translation

Introduction

The field of spinal cord injury (SCI) rehabilitation has experienced significance advances over the past decade. Neurophysiological principles founded on animal models have developed into human trials, e.g., utilization of central pattern generators in training locomotion or epidural spinal stimulation to enhance recovery. Advances in microprocessors and signal processing have resulted in more portable and efficient functional electrical stimulation systems to facilitate movement. Clinical trials have started to identify the required intensity and duration of exercise required for functional gains. Randomized controlled trials have provided quality evidence on the effect of pharmacological interventions on function. Unquestionably, the amount of literature in SCI rehabilitation is growing. Given the expansive growth of such increasingly complex information, it is often difficult for a front-line clinician not intimately familiar with the research methods and analyses to interpret the results of a study. In addition, the interpretation is further complicated by the presence of multiple studies on an intervention, often with what appears to be conflicting messages. The Spinal Cord Injury Rehabilitation Evidence (SCIRE) project is dedicated to providing an up-to-date, accurate research synthesis about the effect of rehabilitation interventions for people with SCI.

In this issue of Topics in Spinal Cord Rehabilitation, we present six papers relevant to SCI rehabilitation clinicians (SCI inpatient rehabilitation practices, gait strategies, upper extremity reconstructive surgery, spasticity treatments, cardiovascular health and bone health). The SCIRE used a systematic and well-defined protocol to assess and synthesize the evidence of the effects of rehabilitation interventions in SCI and is designed for health professionals to inform them of best practices. The full SCIRE compendium is available at www.icord.org/scire.

Evaluation of the evidence

Literature Search Strategy

A systematic review was undertaken using multiple databases (MEDLINE/PubMed, CINAHL®, EMBASE, PsycINFO) to identify all relevant literature published from 1980–2006. An initial broad search was performed with five types of SCI therapies: drug therapy, radiotherapy, diet therapy, rehabilitation therapy and therapy. Articles were included if they involved human subjects, a specific intervention, and they were published in English. Exclusion criteria included: less than half the reported population had a SCI or no measurable outcome associated with the intervention. In the absence of supporting literature, studies with less than 3 subjects were also included. Meta-analyses, systematic reviews, and review articles were also captured. RefWorks, an on-line research management system was used to create the database. References of all articles were reviewed for pertinent articles, which were sought through hand searching. It is known that hand searching may provide higher rates of return than electronic searching within a particular subject area.1

MeSH headings were used with the keywords. Key words were paired with spinal cord injury, tetraplegia, quadriplegia or paraplegia. Specific SCI rehabilitation topics (e.g., spasticity) were identified by a multi-disciplinary team of expert scientists, clinicians, consumers with SCI and policy-makers. These specific topics were searched with additional keywords generated from expert scientists and clinicians in SCI rehabilitation familiar with the topic and more titles and abstracts were reviewed. The specific keywords utilized for the topics reviewed in this issue are listed in the SCIRE compendium.2 The search involved the review of over 17,000 titles and 8400 abstracts, and a final extraction and synthesis of almost 700 articles.

Quality Assessment Tools and Data Extraction

The research design of the study was initially determined (e.g., randomized controlled trial, see Table 1 for full list of research designs). Next, each article was scored for its methodological quality using either the Physiotherapy Evidence Database (PEDro) Score3 or the Downs and Black Tool.4 Two independent raters reviewed each article and determined the quality score. Intraclass correlations (ICC (1,K)) for the two independent scores were 0.91 (95%CI 0.70, 0.98) for the PEDro Score (F test not significant) and 0.90 (95%CI 0.88, 0.92) for the Downs and Black Tool (F test not significant). Scoring discrepancies were resolved through discussion. Tables were generated from the extracted data which included sample subject characteristics, nature of the intervention, outcome measures and key results.

Table 1.

Five levels of evidence

Level Research Design Description
Level 1 Randomized controlled trial (RCT) Randomized controlled trial, PEDro score ≥ 6. Includes within subjects comparison with randomized conditions and cross-over designs
Level 2 RCT Randomized controlled trial, PEDro score < 6.
Prospective controlled trial Prospective controlled trial (not randomized)
Cohort Prospective longitudinal study using at least 2 similar groups with one exposed to a particular condition.
Level 3 Case control A retrospective study comparing conditions, including historical controls
Level 4 Pre-post A prospective trial with a baseline measure, intervention, and a post-test using a single group of subjects.
Post-test A prospective post-test with two or more groups – intervention, then post-test (no pre-test or baseline measurement) using a single group of subjects.
Case Series A retrospective study usually collecting variables from a chart review.
Level 5 Observational Study using cross-sectional analysis to interpret relations.
Clinical Consensus Expert opinion without explicit critical appraisal, or based on physiology, biomechanics or “first principles”
Case Report Pre-post or case series involving one subject

The PEDro score3 was used to assess the methodological quality of individual randomized controlled trials (RCTs). PEDro was developed for the purpose of accessing bibliographic details and abstracts of RCTs, quasi-randomized studies and systematic reviews in physiotherapy. PEDro has been used to assess both pharmacological and non-pharmacological studies with good agreement between raters at an individual item level and in total PEDro scores.5 Maher et al.6 found the reliability of PEDro item ratings varied from “fair” to “substantial,” while the reliability of the total PEDro score was “fair” to “good. The PEDro is an 11-item scale, in which the first item relates to external validity and the other ten items assess the internal validity of a clinical trial. One point is given for each satisfied criterion (except for the first item, which is given a YES or NO), yielding a maximum score of ten. The higher the score, the better the quality of the study as assessed by the following cut-points defined by Foley et al.7: 9–10: excellent; 6–8: good; 4–5: fair; <4: poor. A point for a particular criterion was awarded only if the article explicitly reported that the criterion was met. The scoring system is detailed in Appendix 1.

All other studies with an intervention were assessed with the Downs and Black Tool4 for methodological quality. Saunders et al.8 compared 18 instruments which assessed the quality of non-randomized trials. From their analyses, they suggested that the Downs and Black Tool was the best because it was developed using psychometric methods and had been tested for reliability and validity. This tool consists of 27 questions in the following sub-sections: Reporting, External Validity, Internal Validity (bias) and Internal Validity (confounding). The original tool ranged from 0 to 32. However, the wording for the last question was ambiguous and difficult to score. Thus, we modified the last question from a scale of 0 to 5 to a scale of 0 to 1 where 1 was scored if a power calculation or sample size calculation was present while 0 was scored if there was no power calculation, sample size calculation or explanation whether the number of subjects was appropriate. Thus, our modified version ranged from 0 to 28, with a higher score indicating higher methodological quality. This modified Downs and Black tool is attached in Appendix 2.

Determining Levels of Evidence and Formulating Conclusions

Following the individual study assessment, conclusions were drawn about the accumulated studies for each topic of interest (e.g., effect of treadmill training on gait outcomes) based on the levels of evidence developed by Sackett et al.9 The subcategories with each level (e.g., level 1a, 1b, 1c) proposed by Sackett et al.9 were collapsed into a single level. This was performed to reduce the 10 subcategories from Sackett et al.9 to a less complex system from level 1 to level 5 and which included the key research designs encountered in the SCI rehabilitation literature (Table 1). We provided additional descriptions specific to the types of research designs found in SCI rehabilitation to facilitate the decision-making process. Sackett et al.9 distinguishes high and low quality randomized controlled trials (RCTs) into level 1b and level 2b, respectively. However, the specific criteria for this distinction are not provided. To provide a more reliable decision-making process, we required that a level 1 RCT had a PEDro score of greater than or equal to 6 (good to excellent quality), while a level 2 RCT had a PEDro score of 5 or less. Note that we report sample sizes so the reader is aware of the numbers of subjects that the conclusions are based on. However, there was no sample size requirement for level 1 evidence, but these studies were rated highly according to the standardized PEDro Scale. The appropriateness of the control group was assessed per study. In some studies, an able-bodied group may not have been an adequate control for the particular intervention used, but simply provided “normative’ values for comparison. In those studies, the study was considered “not controlled” and the level of evidence reduced (e.g., level 4 pre-post).

RCTs received priority when formulating conclusions. If studies addressing the same treatment differed in the quality, more weighting was applied to the studies with higher quality scores when deriving the final conclusions. Although conclusions were based on the study designs, levels of evidence and concurring evidence, there were some difficult instances, e.g., the results of a single high quality study conflicting with those of several lower quality studies. When conflicting data was present, an explanation was provided as to how the conclusions were derived to make the process as transparent as possible.

Concluding Remarks

The SCIRE provides a synthesis of the research evidence underlying rehabilitation interventions that is transparent and reproducible. However, it is not without its limitations. Interventions published in the literature may not exactly match what is currently practiced in the clinical setting. As emphasized by Sackett et al.,10 the evidence from systematic research should be integrated with clinical expertise and patients’ choice to form best practice. For example, the extensive literature on functional electrical stimulation in spinal cord rehabilitation is not matched by its practice in the clinical or community setting and could be due to a variety of reasons, e.g., lack of uptake of practice by clinicians, clinician and patient preferences and priorities, lack of an inexpensive commercial product, physical difficulties to use the system or excessive cost (monetary expense, time for treatment, cosmetic acceptance) for the resulting benefits in function.1113

SCIRE allows for objective identification of areas in which there is substantial quality evidence for practice, as well as the gaps in which additional research is urgently required. Although the conclusions reached by SCIRE may change as additional studies are undertaken, new evidence can be incorporated with the same rigor with minimal bias.

Acknowledgments

We would like to acknowledge the financial support for the SCIRE project from the Rick Hansen Man-in-Motion Foundation and Ontario Neurotrauma Fund. In addition, salary support was provided from the Canadian Institutes of Health Research (to Drs. Eng (MSH-63617) and Miller) and the Michael Smith Foundation for Health Research (to Dr. Eng).

Appendix 1. The PEDro Score (adapted from Moseley et al.3 with permission from the Australian Journal of Physiotherapy)

  1. “Subjects were randomly allocated to groups.” (in a crossover study, subjects were randomly allocated an order in which treatments were received). A point for random allocation was awarded if random allocation of patients was stated in its methods. The precise method of randomization need not be specified. Procedures such as coin-tossing and dice-rolling were considered random. Quasi-randomization allocation procedures such as allocation by bed availability did not satisfy this criterion.

  2. “Allocation was concealed.” A point was awarded for concealed allocation if this was explicitly stated in the methods section or if there was reference that allocation was by sealed opaque envelopes or that allocation involved contacting the holder of the allocation schedule who was “off-site.”

  3. “The groups were similar at baseline regarding the most important prognostic indicators.” A trial was awarded a point for baseline comparability if at least one key outcome measure at baseline was reported for the study and control groups. This criterion was satisfied even if only baseline data of study completed-only subjects were presented.

  4. “There was blinding of all subjects.” The person in question (subject, therapist or assessor) was considered blinded if he/she did not know which group the subject had been allocated to. In addition, subjects and therapists were only considered to be “blind” if it could be expected that they would have been unable to distinguish between the treatments applied to different groups. In drug therapy trials, the administrator of the drug was considered the therapist and was considered blinded if he/she did not prepare the drug and was unaware of the drug being administered.

  5. “There was blinding of all therapists who administered the therapy.” (criteria 4.)

  6. “There was blinding of all assessors who measured at least one key outcome” (criteria 4).

  7. “Adequacy of follow-up.” For the purposes of this review, follow-up was considered adequate if all of the subjects that had been originally randomized could be accounted for at the end of the study. The interpretation of this criterion differs from that described by PEDro, where adequacy is defined as the measurement of the main outcome in more than 85% of subjects.

  8. “Intention to treat.” All subjects for whom outcome measures were available received the treatment or control condition as allocated or, where this was not the case, data for at least one key outcome was analyzed by “intention to treat”. For purpose of the present evidence-based review, a trial was awarded a point for intention-to-treat if the trial explicitly stated that an intention-to-treat analysis was performed.

  9. “The results of between-group statistical comparisons are reported for at least one key outcome.” Scoring of this criterion was design dependent. As such, between groups comparison may have involved comparison of two or more treatments, or comparison of treatment with a control condition. The analysis was considered a between-group analysis if either a simple comparison of outcomes measured after the treatment was administered was made, or a comparison of the change in one group with the change in another was made. The comparison may be in the form of hypothesis testing (e.g. p-value) or in the form of an estimate (e.g. the mean, median difference, difference in proportion, number needed to treat, relative risk or hazard ratio) and its confidence interval. A trial was awarded a point for this criterion if between group comparison on at least one outcome measure was made and its analysis of comparison was provided.

  10. “The study provides both point measures and measures of variability for at least one key outcome.” A point measure was referred as to the measure of the size of the treatment effect. The treatment effect was described as being either a difference in group outcomes, or as the outcome in (each of) all groups. Measures of variability included standard deviations, standard errors, confidence intervals, interquartile ranges (or other quartile ranges), and ranges. Point measures and/or measures of variability that were provided graphically (for example, SDs may be given as error bars in a Figure) were awarded a point as long as it was clear what was being graphed (e.g. whether error bars represent SDs or SEs). For those outcomes that were categorical, this criterion was considered to have been met if the number of subjects in each category was given for each group.

Appendix 2. Modified Downs and Black tool (Modified from Downs and Black4 with permission from J Epidemiol Community Health)

Reporting

  • 1)

    Is the hypothesis/aim/objective of the study clearly described?

  • 2)

    Are the main outcomes to be measured clearly described in the Introduction or Methods section?

  • 3)

    Are the characteristics of the patients included in the study clearly described?

  • 4)

    Are the interventions of interest clearly described?

  • 5)

    Are the distributions of principal confounders in each group of subjects to be compared clearly described?

  • 6)

    Are the main findings of the study clearly described?

  • 7)

    Does the study provide estimates of the random variability in the data for the main outcomes?

  • 8)

    Have all important adverse events that may be a consequence of the intervention been reported?

  • 9)

    Have the characteristics of patients lost to follow-up been described?

  • 10)

    Have actual probability values been reported (e.g. 0.035 rather than <0.05) for the main outcomes except where the probability value is less than 0.001?

External validity

  • 11)

    Were the subjects asked to participate in the study representative of the entire population from which they were recruited?

  • 12)

    Were those subjects who were prepared to participate representative of the entire population from which they were recruited?

  • 13)

    Were the staff, places, and facilities where the patients were treated, representative of the treatment the majority of patients receive?

Internal validity - bias

  • 14)

    Was an attempt made to blind study subjects to the intervention they have received?

  • 15)

    Was an attempt made to blind those measuring the main outcomes of the intervention?

  • 16)

    If any of the results of the study were based on “data dredging”, was this made clear?

  • 17)

    In trials and cohort studies, do the analyses adjust for different lengths of follow-up of patients, or in case-control studies, is the time period between the intervention and outcome the same for cases and controls?

  • 18)

    Were the statistical tests used to assess the main outcomes appropriate?

  • 19)

    Was compliance with the intervention/s reliable?

  • 20)

    Were the main outcome measures used accurate (valid and reliable)?

Internal validity – confounding (selection bias)

  • 21)

    Were the patients in different intervention groups (trials and cohort studies) or were the cases and controls (case-control studies) recruited from the same population?

  • 22)

    Were study subjects in different intervention groups (trials and cohort studies) or were the cases and controls (case-control studies) recruited over the same period of time?

  • 23)

    Were study subjects randomized to intervention groups?

  • 24)

    Was the randomised intervention assignment concealed from both patients and health care staff until recruitment was complete and irrevocable?

  • 25)

    Was there adequate adjustment for confounding in the analyses from which the main findings were drawn?

  • 26)

    Were losses of patients to follow-up taken into account?

  • 27)

    Did the study have sufficient power to detect a clinically important effect where the probability value for a difference being due to chance is less than 5%?

Contributor Information

Janice J Eng, Department of Physical Therapy, University of BC, GF Strong Rehab Centre and ICORD, Vancouver, Canada.

Robert Teasell, Department of Physical Medicine and Rehabilitation, Schulich School of Medicine, University of Western Ontario and St. Joseph’s Health Care London.

William C Miller, Department of Occupational Sciences and Occupational Therapy, University of BC, GF Strong Rehab Centre and ICORD, Vancouver, Canada.

Dalton L Wolfe, Lawson Health Research Institute, London, Canada.

Andrea F Townson, Division of Physical Medicine and Rehabilitation, University of British Columbia and ICORD, Vancouver, Canada.

Jo-Anne Aubut, Lawson Health Research Institute, London, Canada.

Caroline Abramson, ICORD, Vancouver, Canada.

Jane T Hsieh, Lawson Health Research Institute, London, Canada.

Sandra Connolly, St. Joseph’s Health Centrel, London, Canada.

Kristin Konnyu, ICORD, Vancouver, Canada.

References

  • 1.Hopewell S, Clarke M, Lefebvre C, Scherer R. Handsearching versus electronic searching to identify reports of randomized trials. The Cochrane Database of Methodology Reviews. 2002;(4):Art. No. MR000001. doi: 10.1002/14651858.MR000001.pub2. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 2.Eng JJ, Teasell RW, Miller WC, Wolfe DL, Townson AF, Aubut J, Abramson C, Hsieh JTC, Connolly S. Methods of the Systematic Reviews. In: Eng JJ, Teasell RW, Miller WC, Wolfe DL, Townson AF, Aubut J, Abramson C, Hsieh JTC, Connolly S, editors; Spinal Cord Injury Rehabilitation Evidence. Vancouver: 2006. pp. 2.1–2.11. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3.Moseley AM, Herbert RD, Sherrington C, Maher CG. Evidence for physiotherapy practice: A survey of the Physiotherapy Evidence Database (PEDro) Aust J Physiother. 2002;48:43–49. doi: 10.1016/s0004-9514(14)60281-6. [DOI] [PubMed] [Google Scholar]
  • 4.Downs SH, Black N. The feasibility of creating a checklist for the assessment of the methodological quality both of randomised and non-randomised studies of health care interventions. J Epidemiol Community Health. 1998;52:377–384. doi: 10.1136/jech.52.6.377. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5.Foley NC, Bhogal SK, Teasell RW, Bureau Y, Speechley MR. Estimates of quality and reliability with the physiotherapy evidence-based database scale to assess the methodology of randomized controlled trials of pharmacological and nonpharmacological interventions. Phys Ther. 2006;86:817–24. [PubMed] [Google Scholar]
  • 6.Maher CG, Sherrington C, Herbert RD, Moseley AM, Elkins M. Reliability of the PEDro scale for rating quality of randomized controlled trials. Phys Ther. 2003;83:713–21. [PubMed] [Google Scholar]
  • 7.Foley NC, Teasell RW, Bhogal SK, Speechley MR. Stroke Rehabilitation Evidence-Based Review: methodology. Top Stroke Rehabil. 2003;10:1–7. [PubMed] [Google Scholar]
  • 8.Saunders LD, Soomro GM, Buckingham J, Jamtvedt G, Raina P. Assessing the methodological quality of nonrandomized intervention studies. West J Nurs Res. 2003;25:223–37. doi: 10.1177/0193945902250039. [DOI] [PubMed] [Google Scholar]
  • 9.Sackett DL, Straus SE, Richardson WS, Rosenberg W, Haynes RB. Evidence-based medicine: how to practice and teach EBM. Toronto, Ontario: Churchill Livingstone; 2000. [Google Scholar]
  • 10.Sackett DL, Rosenberg WM, et al. Evidence based medicine: what it is and what it isn’t. BMJ. 1996;312:71–72. doi: 10.1136/bmj.312.7023.71. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11.Brissot R, Gallien P, Le Bot MP, Beaubras A, Laisne D, Beillot J, Dassonville J. Clinical experience with functional electrical stimulation-assisted gait with Parastep in spinal cord-injured patients. Spine. 2000;25:501–8. doi: 10.1097/00007632-200002150-00018. [DOI] [PubMed] [Google Scholar]
  • 12.Ditunno PL, Patrick M, Stineman M, Morganti B, Townson AF, Ditunno JF. Cross-cultural differences in preference for recovery of mobility among spinal cord injury rehabilitation professionals. Spinal Cord. 2006;44:567–75. doi: 10.1038/sj.sc.3101876. [DOI] [PubMed] [Google Scholar]
  • 13.Estores IM. The consumer’s perspective and the professional literature: what do persons with spinal cord injury want? J Rehabil Res Dev. 2003;40(4 Suppl 1):93–8. doi: 10.1682/jrrd.2003.08.0093. [DOI] [PubMed] [Google Scholar]

RESOURCES