Skip to main content
Indian Journal of Orthopaedics logoLink to Indian Journal of Orthopaedics
. 2008 Apr-Jun;42(2):150–164. doi: 10.4103/0019-5413.40251

Checklists to improve the quality of the orthopaedic literature

Raman Mundi 1,, Harman Chaudhry 1, Ishu Singh 1, Mohit Bhandari 1
PMCID: PMC2759624  PMID: 19826520

Abstract

Several checklists have been developed in an effort to help journals and researchers improve the quality of reporting in research. The CONSORT statement and the CLEAR NPT evaluate randomized trials. The MOOSE and QUOROM checklists evaluate meta-analyses. The STROBE checklists assists readers in evaluating observational studies and the STARD checklist was developed for diagnostic test evaluation. The checklists presented here provide an invaluable source of guidance to authors, journal editors and readers who are seeking to prepare and evaluate reports. As evidence-based medicine continues to establish itself as the new paradigm by which medicine is practiced, the need for good reporting for all research designs must also become commonplace as opposed to the exception.

Keywords: Critical appraisal, checklists, meta-analysis, diagnostic tests, quality assessment, randomized trials

INTRODUCTION

The current quality of reporting in both the medical and surgical literature is poor and in need of immediate improvement.15 The orthopaedic literature is no exception. In fact, many studies have highlighted the substandard quality of reporting in even ‘high quality’ study designs such as randomized controlled trials and systematic reviews.69

The true benefits of well conducted studies with valid results can only be realized if they are presented to readers in a comprehensive and transparent manner. Owing to the need for improved reporting, several checklists have been developed to guide authors preparing manuscripts for various types of study designs. Although some journals have endorsed certain checklists with marginal improvements in reporting, the quality is still often poor as authors do not adhere to many of the checklist recommendations.5,10

There is a dire need to promote awareness and understanding of these available checklists so that authors can begin taking advantage of these invaluable guides. The objective of this report is to introduce several of the existing checklists for various study designs. In particular, the study designs focused on in this article include randomized controlled trials, systematic reviews, observational studies, diagnostic trials and qualitative studies.

CHECKLISTS FOR RANDOMIZED CONTROLLED TRIALS

Although randomized controlled trials (RCTs) are considered the ‘gold standard’ study designs for evaluating treatment effectiveness, they nevertheless remain subject to bias unless methodological and statistical safeguards are implemented into the trial. Briefly, such safeguards include allocation concealment, blinding, ensuring complete patient follow-up and analyses according to the intention-to-treat principle. This is just a sample of many safeguards and other components of randomized trials that should be reported in trial manuscripts. For instance, there are methodological issues unique to nonpharmacological trials (NPTs) that also merit reporting, such as standardization of the intervention and ensuring adequate care provider skill.

To aid investigators in preparing comprehensive and high quality manuscripts for randomized trials, two checklists have gained much attention: 1) the Consolidated Standards of Reporting Trials (CONSORT) statement and 2) the Checklist to Evaluate a Report of a Nonpharmacological Trial (CLEAR NPT).

The CONSORT statement was first published in The Journal of the American Medical Association in 1996 and revised in 2001. This statement consists of a 22-item checklist and flow diagram that serve as a detailed set of recommendations on how to prepare a report for a randomized trial or conversely, aid in critically appraising the reports of others. In particular, the checklist provides recommendations on how to report the design, analysis and interpretation of the study, whereas the flow diagram offers guidance on how to report the progress of participants through the trial.11 Realizing that randomized trials of nonpharmacological interventions have unique challenges—such as complex interventions and difficulty blinding patients—not fully addressed by the revised CONSORT statement, an extension to the CONSORT was recently made to specifically address the issues facing these trials.12 Both the standard CONSORT checklist and the modified version for NPTs can be seen in Figure 1. Figure 2 illustrates the standard CONSORT flow diagram and Figure 3 illustrates the modified flow diagram for NPTs. Adherence in the orthopaedic literature to the CONSORT recommendations has been demonstrated to be poor.8,9 For instance, Bhandari and colleagues8 evaluated the reports of 196 randomized trials investigating fracture care across 32 journals and found that the average report adhered to only 32% ± 29% of the CONSORT criteria.

Figure 1.

Figure 1

CONSORT checklist12

Figure 2.

Figure 2

Standard CONSORT flow diagram11

Figure 3.

Figure 3

CONSORT flow diagram for Nonpharmacological Trials12

Developed in 2005, the CLEAR NPT is a 15-item checklist (10 main items and 5 sub-items), that serves to critically appraise the reports of randomized trials of nonpharmacological interventions [Figure 4].13 As implied by its name, this checklist is useful for assessing nonpharmacological trials due to its focus on key methodological issues surrounding NPTs. Each item on the checklist can be answered with a quick Yes, No or Unclear, making it an efficient tool to evaluate the literature with. Further supporting the claims of studies utilizing the CONSORT statement, Chan and Bhandari14 have demonstrated that the quality of reporting for randomized trials in the orthopaedic literature as assessed by the CLEAR NPT is also suboptimal. Although this checklist's primary utility is to evaluate reports, it still serves as a useful guide to authors preparing manuscripts of randomized trials.

Figure 4.

Figure 4

CLEAR NPT checklist13

CHECKLISTS FOR SYSTEMATIC REVIEWS AND META-ANALYSES

With the plethora of studies being published constantly, summarizing the results of primary articles on a given topic is a useful and helpful practice for health care providers and policy makers.15 Systematic reviews are considered high quality evidence due to their systematic approach at collecting, critically appraising and synthesizing data from original articles on a specific topic. If a quantitative analysis is performed to arrive at a single best estimate of the treatment effect, these reviews are better known as meta-analyses. Due to their systematic nature and ability to put forth a single best estimate of the treatment effect, meta-analyses can have a significant impact on patient care. However, meta-analyses may vary in their methodological rigor and produce results of varying credibility. For instance, the quality of a systematic review or meta-analyses is directly dependent on the quality of the studies included. Thus, systematic reviews and meta-analyses that consider only RCTs would provide stronger evidence than those which consider non-randomized studies as well. However, between 1996 and 2001, it was found that the majority of orthopaedic systematic reviews published in peer-reviewed journals outside the Cochrane collaboration included non-randomized trials.16 Clinicians reading the reports of systematic reviews and meta-analyses must be able to appraise the methods and validity of the study in order to confidently interpret their results.

The Quality of Reporting of Meta-Analyses (QUOROM) statement, consisting of an 18-item checklist and flow diagram, was developed to aid authors preparing reports of meta-analyses of RCTs [Figure 5 and 6].15 The QUOROM checklist outlines a set of recommendations on how to prepare the abstract, introduction, methods, results and discussion sections of a meta-analysis. The ultimate goal of these reporting guidelines is to provide readers with transparency regarding the search, selection, validity assessments, data abstraction, study characteristics, quantitative data synthesis and trial flow of the study.15 For instance, under the “methods” section of the checklist, authors are encouraged to report the criteria used to assess the quality of the included RCTs and the outcome of such quality assessments. This is imperative, as RCTs with deficiencies in certain methodological safeguards have been shown to produce biased results.5 Incorporating these ‘biased’ studies without caution into a meta-analysis would also result in a biased estimate of the treatment effect in the meta-analysis. The purpose of the flow diagram is to help authors on reporting details of the inclusion and exclusion of RCTs. Although the QUOROM statement is designed specifically to guide reporting of meta-analyses, authors of systematic reviews of RCTs can also benefit from these recommendations—with the exception of the reporting recommendations geared towards the quantitative analysis, as this step is only carried out in a meta-analysis.15

Figure 5.

Figure 5

QUOROM checklist15

Figure 6.

Figure 6

QUOROM flow diagram15

Not all meta-analyses can rely solely on RCTs to answer a question of interest. First and foremost, RCTs are relatively scarce in the orthopaedic literature, making it impractical to always exclusively use data from RCTs.6 Secondly, for issues surrounding risk factors for disease and harm, it would be unethical to randomize patients to groups in which they would be subject to any potentially harmful risks.17 For instance, if the question of interest was, “What is the effect of smoking on fracture nonunion rates?”, it would not be feasible to carry out a randomized study in which patients were randomly allocated to either a smoking or non-smoking group.18 As a result, several meta-analyses rely upon observational studies - those in which patients are naturally exposed to risk factors or in which physician or patient preference determines allocation to a treatment or control intervention.19 The importance of comprehensive reporting for such meta-analyses can not be overstated, as observational studies are more prone to biased results than RCTs. In fact, arriving at a single estimate of the effect of a treatment or exposure when pooling data from observational studies requires extreme caution, as these results can often be misleading.16

In 2000, the Meta-Analyses of Observational Studies in Epidemiology (MOOSE) group produced a 35-item checklist that details how meta-analyses of observational studies should be reported. Specifically, the checklist provides recommendations on how to report background information, the search strategy and the methods, results, discussion and conclusion sections of the paper [Figure 7]. For each of these six categories, there are several corresponding items that are listed by the MOOSE groups as essential for reporting. For instance, under search strategy, reports should include the qualifications of the searchers, a detailed description of the search strategy and the method by which articles in foreign languages were utilized, among other details regarding the search.17

Figure 7.

Figure 7

MOOSE checklist17

CHECKLISTS FOR OBSERVATIONAL STUDIES

As mentioned, not all research questions can be answered through RCTs. Therefore, observational studies have an important role in answering questions of treatment effectiveness and disease etiology. Three primary observational study designs include the cohort, case-control and cross-sectional studies. Briefly, cohort studies usually follow two groups of patients; one group in which everyone has been exposed to a risk factor or treatment and the other in which no exposure has occurred. The groups are then compared for the rate of development of disease or outcome of interest. In case-control studies, a group that has already developed an outcome of interest is compared to a group without the outcome for factors that may be associated with the outcome. Cross-sectional surveys are carried out at a single time point, at which both the outcome and factors of interest are measured. Due to the lack of randomization, observational studies are inherently more prone to potential biases. Even if investigators attempt to match groups for known prognostic factors, there may be underlying imbalances in unknown prognostic factors that may produce misleading and biased results.20 Furthermore, case-control studies are always retrospective in nature (cohort studies may also be retrospective) which increases the potential for incomplete and biased data collection.18 Despite these limitations, observational studies have a crucial role to play in medical research and as such, satisfactory reporting to allow readers to evaluate these studies is of utmost importance.

The Strengthening the Reporting of Observational Research in Epidemiology (STROBE) statement, outlines how to prepare good manuscripts for these three observational study designs. It consists of a 22-item checklist which provides reporting recommendations for all sections of the paper, as well as on funding sources [Figure 8]. Of the 22-items, 18 are general to all three study designs and 4 are design-specific. In particular, information in the methods section regarding participants (item 6) and statistical methods (item12), as well as in the results section regarding descriptive data (item 14) and outcome data (item 15) are design-specific. Although the STROBE group emphasizes that reporting of all 22-items in this checklist is essential, they encourage authors to utilize their preferences and creativity when selecting the order and format of presenting such details.21

Figure 8.

Figure 8

STROBE checklist22

CHECKLISTS FOR STUDIES OF DIAGNOSTIC ACCURACY

Diagnostic tests are widely used by clinicians to diagnose health states and subsequently initiate, alter or terminate various treatment options.22 Diagnostic studies evaluate the accuracy of a diagnostic test (by its level of agreement to the current ‘gold standard’ for diagnosis) in predicting a disease, stage of a disease, health status or any health condition that could prompt clinical action.22 The ‘gold standard’ is typically impractical to use in regular clinical encounters and therefore the study is attempting to offer a more practical alternative.23 As such, studies of diagnostic test accuracy have the potential to directly impact treatment decisions and, therefore, patient care. Unfortunately, it has been demonstrated that methodologically compromised studies are more likely than methodologically rigorous studies to overestimate the accuracy of diagnostic tests.24 In the hands of the uncritical clinician, such poor studies may lead to the unwarranted use and interpretation of a diagnostic test, ultimately to the detriment of high quality patient care.

Recognizing the importance of studies evaluating diagnostic accuracy, Bossuyt and colleagues25 developed the Standards for the Reporting of Diagnostic Accuracy studies (STARD) statement. The STARD statement includes a 25-item checklist which outlines crucial information that authors should include in the abstract, introduction, methods, results and discussion sections of a report to enable an adequate assessment of both external validity (i.e. how generalizable study results are) and internal validity (i.e. the potential for bias) [Figure 9].26 In addition to some relatively common elements, such as inclusion/exclusion criteria, method of data collection and method of data analysis, the STARD checklist also includes some unique items. For instance, it asks for a description and rationale of the gold standard to which the diagnostic test (referred to as the index test) is being compared. This is because even positive study results will be limited by the effectiveness of the ‘gold standard’ as a diagnostic tool.23 The STARD statement also includes and encourages authors to use a flow diagram to report the number of patients included and excluded in the diagnostic and/or ‘gold standard’ tests [Figure 10].26

Figure 9.

Figure 9

STARD checklist27

Figure 10.

Figure 10

STARD flow diagram27

Rama and colleagues recently published an investigation of diagnostic accuracy studies using the STARD criteria in three orthopaedic journals.27 They found that the majority of studies had deficiencies in reporting of methodology and, overall, reported less than two-thirds of the STARD criteria. Currently, no major orthopaedic journals have adopted the STARD statement;28 this may be partly attributable to the scarcity of diagnostic accuracy studies, which constitute only 1% of the orthopaedic literature.27 However, owing to the enormous implications that a newly implemented diagnostic test can have on patient care, we believe that the STARD statement must be endorsed to enable readers to adequately interpret study results and prevent untenable treatment decisions.

CHECKLISTS FOR QUALITATIVE STUDIES

Qualitative studies are useful in the surgical literature for comprehensively describing phenomena (from social, emotional and experiential perspectives) as well as for generating hypotheses that can subsequently be quantitatively verified or disproven.29,30 Giacomini, Cook and Guyatt have stated that qualitative studies can provide a “rigorous alternative to armchair hypothesizing”.29 For instance, one study published in the Journal of Bone and Joint Surgery (American Volume) explored the reasons why (from the patient's perspective) many elderly arthritic patients are unwilling to undergo a total joint replacement procedure.31 The principles of evidence-based medicine demand that articles should be critically appraised before results are implemented into clinical practice. However, the methodological rigour of qualitative studies has come under criticism; commentators have stated that there is a need for rigorous methodological standards in order to minimize the effect of bias on study results.32,33

The most comprehensive available checklist for qualitative studies is the RATS guidelines developed by Clark and adopted as a 28-item checklist by BioMed Central in the instructions to authors section [Figure 11].34,35 RATS is an acronym which describes four components of a rigorously reported qualitative study: 1) Relevance of the study question; 2) Appropriateness of qualitative method; 3) Transparency of procedures; and 4) Soundness of interpretive approach.35 In addition to the 28-item checklist, the RATS guideline offers a section on possible “red flags” authors should avoid.34 Unfortunately, there does not appear to be an overwhelming consensus on the effectiveness of this particular checklist to ensure all pertinent methodological criteria have been met, as there is for the CONSORT criteria. However, authors of qualitative studies are advised to consider and incorporate the checklist criteria in reporting findings.

Figure 11.

Figure 11

RATS checklist35

CONCLUSION

For over a decade now, several checklists have been developed in an effort to help investigators prepare reports for a variety of different study designs. If the time, effort and resources put forth in carrying out medical research is to make its impact on patient care and policy decisions, then the importance of complete and comprehensive reporting can not be overstated. The checklists presented here provide an invaluable source of guidance to authors, journal editors and readers who are seeking to prepare and evaluate reports. To gain greater information on these and other available reporting checklists, we encourage readers to locate the original articles in which these checklists are published. Furthermore, certain checklists, such as the CONSORT statement, have corresponding explanation and elaboration papers which are informative and aid in promoting understanding of the checklists. As evidence-based medicine continues to establish itself as the new paradigm by which medicine is practiced, the need for good reporting for all research designs must also become commonplace as opposed to the exception.

Footnotes

Disclaimer: Dr. Bhandari was supported in part by a Canada Research Chair. Raman Mundi is the recipient of the ‘Studentships in Musculoskeletal Research’ from Canadian Institutes of Health Research. Raman Mundi and Harman Chaudhry are recipients of the Bachelor of Health Sciences Programme ‘2007 Research Scholarships’.

Source of Support: Nil

Conflict of Interest: None.

REFERENCES

  • 1.Tiruvoipati R, Balasubramanian SP, Atturu G, Peek GJ, Elbourne D. Improving the quality of reporting randomized controlled trials in cardiothoracic surgery: The way forward. J Thorac Cardiovasc Surg. 2006;132:233–40. doi: 10.1016/j.jtcvs.2005.10.056. [DOI] [PubMed] [Google Scholar]
  • 2.Balasubramanian SP, Wiener M, Alshameeri Z, Tiruvoipati R, Elbourne D, Reed MW. Standards of reporting of randomized controlled trials in general surgery: Can we do better? Ann Surg. 2006;244:663–7. doi: 10.1097/01.sla.0000217640.11224.05. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3.Agha R, Cooper D, Muir G. The reporting quality of randomised controlled trials in surgery: A systematic review. Int J Surg. 2007;5:413–22. doi: 10.1016/j.ijsu.2007.06.002. [DOI] [PubMed] [Google Scholar]
  • 4.Jacquier I, Boutron I, Moher D, Roy C, Ravaud P. The reporting of randomized clinical trials using a surgical intervention is in need of immediate improvement: A systematic review. Ann Surg. 2006;244:677–83. doi: 10.1097/01.sla.0000242707.44007.80. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5.Mills E, Wu P, Gagnier J, Heels-Ansdell D, Montori V. An analysis of general medical and specialist journals that endorse CONSORT found that reporting was not enforced consistently. J Clin Epidemiol. 2005;58:662–7. doi: 10.1016/j.jclinepi.2005.01.004. [DOI] [PubMed] [Google Scholar]
  • 6.Bhandari M, Richards RR, Sprague S, Schemitsch EH. The quality of reporting randomized trials in the Journal of Bone and Joint Surgery from 1988 through 2000. J Bone Joint Surg Am. 2002;84:388–96. doi: 10.2106/00004623-200203000-00009. [DOI] [PubMed] [Google Scholar]
  • 7.Poolman R, Struijs PA, Krips R, Sierevelt IN, Lutz KH, Bhandari M. Does a “Level I Evidence” rating imply high quality of reporting in orthopaedic randomised controlled trials? BMC Med Res Methodol. 2006;6:44–51. doi: 10.1186/1471-2288-6-44. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.Bhandari M, Guyatt GH, Lochner H, Sprague S, Tornetta P., 3rd Application of the consolidated standards of reporting trials (CONSORT) in the fracture care literature. J Bone Joint Surg Am. 2002;84:485–9. doi: 10.2106/00004623-200203000-00023. [DOI] [PubMed] [Google Scholar]
  • 9.Cowan J, Lozano-Calderon S, Ring D. Quality of prospective controlled randomized trials: Analysis of trials of treatment for lateral epicondylitis as an example. J Bone Joint Surg Am. 2007;89:1693–9. doi: 10.2106/JBJS.F.00858. [DOI] [PubMed] [Google Scholar]
  • 10.Devereaux PJ, Manns BJ, Ghali WA, Quan H, Guyatt GH. The reporting of methodological factors in randomized controlled trials and the association with a journal policy to promote adherence to the Consolidated Standards of Reporting Trials (CONSORT) checklist. Control Clin Trials. 2002;23:380–8. doi: 10.1016/s0197-2456(02)00214-3. [DOI] [PubMed] [Google Scholar]
  • 11.The CONSORT group. CONSORT 2001 checklist and CONSORT 2001 flow diagram? [cited on 2008 Mar 1] Available from: http://www.consort-statement.org/index.aspx?o=1031.
  • 12.Boutron I, Moher D, Altman DG, Schulz KF, Ravaud P. Extending the CONSORT statement to randomized trials of nonpharmacological treatments: Explanation and elaboration. Ann Intern Med. 2008;148:205–9. doi: 10.7326/0003-4819-148-4-200802190-00008. [DOI] [PubMed] [Google Scholar]
  • 13.Boutron I, Moher D, Tugwell P, Giraudeau B, Poiraudeau S, Nizard R, et al. A checklist to evaluate a report of a nonpharmacological trial (CLEAR NPT) was developed using consensus. J Clin Epidemiol. 2005;58:1233–40. doi: 10.1016/j.jclinepi.2005.05.004. [DOI] [PubMed] [Google Scholar]
  • 14.Chan S, Bhandari M. The quality of reporting of orthopaedic randomized trials with use of a checklist for nonpharmacological therapies. J Bone Joint Surg Am. 2007;89:1970–8. doi: 10.2106/JBJS.F.01591. [DOI] [PubMed] [Google Scholar]
  • 15.Moher D, Cook DJ, Eastwood S, Olkin I, Rennie D, Stroup DF. Improving the quality of reports of meta-analyses of randomized controlled trials: the QUOROM statement. Lancet. 1999;354:1896–900. doi: 10.1016/s0140-6736(99)04149-5. [DOI] [PubMed] [Google Scholar]
  • 16.Audige L, Bhandari M, Griffin D, Middleton P, Reeves BC. Systematic reviews of nonrandomized clinical studies in the orthopaedic literature. Clin Orthop Relat Res. 2004;427:249–57. doi: 10.1097/01.blo.0000137558.97346.fb. [DOI] [PubMed] [Google Scholar]
  • 17.Stroup DF, Berlin JA, Morton SC, Olkin I, Williamson GD, Rennie D, et al. Meta-analysis of observational studies in epidemiology: A proposal for reporting. JAMA. 2000;283:2008–12. doi: 10.1001/jama.283.15.2008. [DOI] [PubMed] [Google Scholar]
  • 18.Petrisor BA, Bhandari M. The hierarchy of evidence: levels and grades of recommendation. Indian J Orthop. 2007;41:11–5. doi: 10.4103/0019-5413.30519. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19.Guyatt GH. Therapy and harm: An introduction. In: Guyatt GH, Rennie D, editors. Users’ guides to the medical literature: A manual for evidence-based clinical practice/The Evidence-Based Medicine Working Group. Chicago: AMA Press; 2002. pp. 49–53. [Google Scholar]
  • 20.Levine M, Haslam D, Walter S, Cumming R, Lee H, Haines T, et al. Harm. In: Guyatt GH, Rennie D, editors. Users’ guides to the medical literature: A manual for evidence-based clinical practice/The Evidence-Based Medicine Working Group. Chicago: AMA Press; 2002. pp. 81–100. [Google Scholar]
  • 21.von Elm E, Altman DG, Egger M, Pocock SJ, Gøtzsche PC, Vandenbroucke JP. The Strengthening the Report of Observational Studies in Epidemiology (STROBE) Statement: Guidelines for Reporting Observational Studies. Ann Intern Med. 2007;147:573–7. doi: 10.7326/0003-4819-147-8-200710160-00010. [DOI] [PubMed] [Google Scholar]
  • 22.Bossuyt PM, Reitsma JB, Bruns DE, Gatsonis CA, Glasziou PP, Irwig LM, et al. The STARD statement for reporting studies of diagnostic accuracy: Explanation and elaboration. Ann Intern Med. 2003;138:W1–12. doi: 10.7326/0003-4819-138-1-200301070-00012-w1. [DOI] [PubMed] [Google Scholar]
  • 23.Furukawa TA, Guyatt GH. Sources of bias in diagnostic accuracy studies and the diagnostic process. CMAJ. 2006;174:481–2. doi: 10.1503/cmaj.060014. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 24.Lijmer JG, Mol BW, Heisterkamp S, Bonsel GJ, Prins MH, van der Meulen JH, et al. Empirical evidence of design-related bias in studies of diagnostic tests. JAMA. 1999;282:1061–6. doi: 10.1001/jama.282.11.1061. [DOI] [PubMed] [Google Scholar]
  • 25.Bossuyt PM, Reitsma JB, Bruns DE, Gatsonis CA, Glasziou PP, Irwig LM, et al. Towards complete and accurate reporting of studies of diagnostic accuracy: The STARD initiative. Clin Chem. 2003;49:1–6. doi: 10.1373/49.1.1. [DOI] [PubMed] [Google Scholar]
  • 26.The STARD group. STARD checklist and STARD flow diagram [cited on 2008 Mar 1] Available from: http://www.stard-statement.org.
  • 27.Rama S, Boddu KR, Sharmila P, Apsingi S. Quality of reporting of orthopaedic diagnostic accuracy studies is suboptimal. Clin Ortho Rel Res. 2006;446:247. doi: 10.1097/01.blo.0000205906.44103.a3. [DOI] [PubMed] [Google Scholar]
  • 28.The STARD group. Adopters of STARD. [cited on 2008 Mar 1] Available from: http://www.stard-statement.org.
  • 29.Giacomini M, Cook D, Guyatt GH. Therapy and applying the results: Qualitative research. In: Guyatt GH, Rennie D, editors. Users’ guides to the medical literature: a manual for evidence-based clinical practice/The Evidence-Based Medicine Working Group. Chicago: AMA Press; 2002. pp. 433–48. [Google Scholar]
  • 30.Rich JA, Grey CM. Qualitative research on trauma surgery: getting beyond the numbers. World J Surg. 2003;27:957–61. doi: 10.1007/s00268-003-7099-0. [DOI] [PubMed] [Google Scholar]
  • 31.Clark JP, Hudak PL, Hawker GA, Coyte PC, Mohamed NN, Kreder HJ, et al. The moving target: A qualitative study of elderly patients’ decision-making regarding total joint replacement surgery. J Bone Joint Surg Am. 2004;86:1366–74. [PubMed] [Google Scholar]
  • 32.Poses RM, Isen AM. Qualitative research in medicine and health care. J Gen Intern Med. 1998;13:32–8. doi: 10.1046/j.1525-1497.1998.00005.x. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 33.Poses RM, Isen AM. In defense of qualitative research: responses to Poses and Isen perspective article. J Gen Intern Med. 1998;13:64–72. doi: 10.1046/j.1525-1497.1998.00017.x. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 34.Clark JP. Qualitative research review guidelines: RATS. [cited on 2008 Mar 1]. Available from: http://www.biomedcentral.com/info/ifora/rats.
  • 35.Clark JP. How to peer review a qualitative manuscript. In: Godlee F, Jefferson T, editors. Peer review in health sciences Second edition. London: BMJ Books; 2003. pp. 219–35. [Google Scholar]

Articles from Indian Journal of Orthopaedics are provided here courtesy of Indian Orthopaedic Association

RESOURCES