Skip to main content
International Dental Journal logoLink to International Dental Journal
. 2020 Nov 8;61(1):37–42. doi: 10.1111/j.1875-595X.2011.00007.x

Quality of randomised controlled trials in dentistry

Iacopo Cioffi 1,*, Mauro Farella 2
PMCID: PMC9374841  PMID: 21382032

Abstract

Randomised controlled trials (RCTs) are regarded as the best study designs to test the efficacy of medical and dental intervention. Many reports, however, have shown that at the moment the quality of dental RCT reports is still poor, and further efforts to improve it are necessary. It has been suggested that trials that are not well designed provide biased estimates of the treatment effects and that a journal’s impact factor is not related to the quality of RCTs published. For these reasons, before trusting RCT reports, a careful assessment of the study selected is needed. Randomisation, blinding, allocation concealment, drop outs analysis are essential quality components of RCTs. Many systems for RCTs quality assessment are available. In this report the concept of quality of RCTs will be critically evaluated and the most commonly used instruments available for quality assessment of RCTs in dental research will be reviewed.

Key words: Randomized controlled trials (RCTs), quality


Evidence Based Medicine (EBM) is the procedure by which the outcomes of the scientific research are applied to daily clinical practice1. Currently, this concept is popular among dentists and for some years has been known as evidence based dentistry. Randomised controlled trials (RCTs) represent the gold standard for testing the efficacy of medical interventions2., 3. and are generally regarded as the most reliable source of ‘scientific evidence’ in healthcare4.

The reliability of the results coming from clinical trials strongly depends on a sound methodology and proper management of research. Indeed, biases resulting from poorly conducted trials could mislead clinicians and address questionable clinical decisions. In recent years the need for predictable and effective dental treatments has lead clinicians to look for therapeutic approaches validated by means of experimental trials but, unfortunately, the quality of many published medical and dental RCTs is variable5. Also, no correlation between the quality of the trials and the journal impact factor has been noted5. This may cause incorrect evaluations by readers who may assess quality of trials by the impact factor of the journal in which they are published.

Moreover, basing therapeutic approaches on trials that are not well designed might be a risk for patients and, from a legal perspective for dentists also. For this reason, it is important that clinicians are enabled to recognise good-quality evidence and to use only those studies in support of their daily practice. In this report the concept of quality of RCTs will be critically evaluated and the most commonly used instruments available for quality assessment of RCTs in dental research will be described.

RANDOMISED CONTROLLED TRIALS

Clinical trials are scientific experiments involving subjects or/and patients designed to find out the most effective therapy for treatment of a disease2. Generally the results coming from clinical trials help to set-up a treatment protocol for populations who are affected by the same condition of the research sample or to choose among the different therapies available2. The efficacy of the treatment tested in a trial is determined by comparing the findings obtained from a treated group, on which the treatment protocol object of the research is tested, with those obtained from a control group, which is treated with a different treatment protocol or is not treated at all.

Without using a control group, the efficacy of the treatment tested cannot be assessed adequately being generally overestimated. The allocation of the experimental sample to both groups should be randomised. Randomisation prevents selection biases during group allocation and allows that control and treated subjects will have similar characteristics at baseline.

Another important aspect of reliable clinical trials is blinding. This is the process by which people involved in research or patients (single blinding), or both (double blinding) are not aware of one of more aspects of the research process. Single or double blinding should be considered when planning clinical trials because they can drastically increase the internal validity of a trial and reduce the risk of examiner’s or patient’s biases.

Quality of randomised controlled trials: concepts and tools

The assessment of the methodological quality of a trial is essential as quality can considerably influence the scientific outcomes and the clinical interpretation of the research. The definition of the construct ‘quality’ is the first needed but difficult step for evaluating the quality of a trial. Quality has been defined as the likelihood of the trial design to generate unbiased results, sufficiently valid that they can be applied in clinical practice3. The quality of a trial quantifies the likelihood that the experimental outcomes are valid estimates of the truth.

Thus far, some issues have been considered when assessing the quality of RCTs6:

  • The clinical relevance of the research question

  • The internal validity of the trial (the degree to which the trial design, conduct and statistical analysis have minimised biases during comparison of the interventions)

  • The external validity (the validity of the results in real-world settings)

  • Correct data analysis and presentation

  • The ethical implications of the intervention.

When talking about quality, two different issues should be considered. The methodological quality of the trial itself that is related to its internal and external validity, and the reporting quality, that concerns the reporting of the research design, conduct, and data analysis. Since the only instrument available to readers for assessing the quality of a trial is the published manuscript, RCT reports must provide the most accurate information about the trial design, conduct and data analysis7.

Quality scales and checklists may be used for assessing RCT quality. Scales provide a quantitative assessment of the quality of a trial by giving a quality numerical score, while checklists do not provide a quantitative score. Quality scales and checklists consist of lists of methodological issues that should be included when performing RCTs. Generally, the tools available include methodological issues that have been selected by experts in the field, such as statisticians, or include items that have been considered as ‘accepted criteria’ in textbooks.

However, the most common issues considered of these tools are randomisation, allocation concealment, blinding, withdrawals and drop-out analysis. These issues have been reported to strongly influence the outcomes of a trial and the efficacy of the treatments tested. For instance, a clinically and statistically 30–50% exaggeration of treatment efficacy in lower quality trials has been shown8 and larger estimates of treatment effects have been reported in trials in which allocation concealment was unclear or inadequate9., 10.. Therefore quality assessment tools such as scales and checklists are needed before assessing published RCTs.

In 2001 the number of quality scales and checklists available in literature was estimated to be between fifty and sixty3. An extensive review of scales can be found in Moher et al.11 and Olivo et al.12. The scales available differ for the number of items incorporated and for the importance given to each of them. Although these tools provide an overall summary score of the quality of a trial, they should be first tested for their validity and reliability. In many cases the scales available are the result of modifications of primary quality scales used for other disciplines and cannot be considered valid and reliable unless they are tested.

Among the RCTs quality assessment tools, the Jadad Scale13 is the most commonly used, which provides an easy score for a quantitative evaluation of the quality of a trial (Figure 1). Randomisation, blinding, and the account for withdrawals and drop-outs are considered within this scale. The quality assessment of research reports is easy and quick. For each item a unit score is added or detracted from an initial given score (0). Score range is −5 to 5. Low quality trials result in a score lower than three.

Figure 1.

Figure 1.

Jadad Scale. Modified from Jadad et al7.

The inter-rater reliability of the Jadad scale has been reported to be high in many studies, ranging from 0.48 to 1.00 (Intraclass correlation coefficients) and 0.37 to 0.89 (Kappa values)12., 14., 15., 16., 17., 18., 19., 20.. The test retest reliability was shown to be 0.9821. Nevertheless, the high simplicity of the scale and the large influence of blinding on the overall summary score might be questionable since in some disciplines blinding is not always applicable because of the nature of the clinical interventions. For these reasons a reassessment of the validity of the Jadad Scale has been recently claimed12.

The Delphi list22 (Figure 2) is a checklist devised in 1998 by statisticians and experts in RCTs. With this tool, a consensus among experts about a core of items for quality assessment of RCTs was achieved. The Delphi method represents the first step toward a standard in RCT quality assessment. After three sessions, a set of items was chosen to be included in the Delphi list. Inter-observer reliability was reported to range from 0.5423 to 0.8524. Although several efforts were made in order to reach a consensus, this instrument has not been as popular as the Jadad scale.

Figure 2.

Figure 2.

Delphi list. Modified from Verhagen et al22.

The PEDro scale25 (Figure 3) has been developed for the Physiotherapy Evidence Database (PEDro). This instrument comprises a set of 11 items, and consists of a modification of the primary 9-item Delphi list, with two items more (items 8 and 10). The inter-rater reliability has been reported to range from −0.61 to 0.8826., 27.. The PEDro scale was developed in order to help consultants of the PEDro database to quickly identify RCTs with internal validity and with sufficient statistical information among the uploaded RCTs. The scoring of PEDro scale is significantly influenced by blinding since different types of blinding are scored by three items. Therefore the quality score of RCTs related to disciplines in which blinding cannot be completely achieved is underestimated with this tool.

Figure 3.

Figure 3.

PEDro Scale. Modified from PEDro Physiotherapy Evidence Database: http://www.pedro.org.au/25

The question of which scale to use is a difficult one since a gold standard is lacking and the choice is not reliable28., 29.. The availability of various scales implies that the quantitative evaluation of the quality of a trial is different according to the system used. Thus, the use of scales developed with a consensus should be largely promoted. However, it is important to note that only the research report (i.e. the manuscript) is the object of the quality assessment. Hence, a trial designed with many biases, but well reported, can get a high score. Conversely, well designed trials poorly reported may receive a low score. As a consequence of this, systems have been developed to increase the quality of reporting trials. The introduction of CONSORT Statement30 in publication process has improved the reporting quality of RCTs, allowing more accurate and reproducible assessments31.

The CONSORT (Consolidation of the Standards of Reporting trials) statement was developed in 1996 to improve the standard of quality of RCTs and currently offers a standard for preparing reports of trials. A 25-item checklist and a flowchart were attached to this document and uploaded on-line with the intent of enhancing reporting of RCTs. The checklist items focus on trial design while the flow chart guides the progress of all participants through the research phases.

Currently a growing number of journals support the CONSORT statement and invite authors to follow its guidelines before submitting papers. The use of CONSORT has been shown to significantly increase the quality of reporting of RCTs in different medical fields in the recent past. Currently, modified versions of the CONSORT checklist have been used for quality assessment of trials in different medical fields. However, better reporting of RCTs is still claimed in nursing journals32, orthopaedic surgery33 and plastic surgery34. Better reporting of clinical trials in the five leading Chinese medical journals has been also requested35.

QUALITY OF RANDOMISED CONTROLLED TRIALS IN DENTISTRY

So far, the quality of randomised controlled trials has been assessed in few dental disciplines and has been object of only few studies. In a report concerning RCTs in paediatric dentistry36, the quality of trials published between 1985 and 1997, and between 1998 and 2006 were compared to assess the influence of CONSORT introduction on quality of published RCTs. On average, the reporting quality of the published reports was poor. Also in a study by Dumbrigue et al.37 the quality of implant dentistry RCTs was assessed. Using a scheme developed through the Cochrane Collaboration, two examiners assessed the quality of 67 articles and concluded that accounting of randomisation procedures and blinding for most implant RCTs was inadequate. The quality of prosthodontic RCTs was evaluated by Jokstad et al.38. Randomisation and procedures for concealment allocation were not described in 70% of the articles. The random allocation sequence was not clear in 94% of papers. Blinding was not reported in 72% of cases.

The quality of published RCTs has been examined also in relation to Journal Impact Factor® in different areas of dental research by Sjogren and Halling5. The Jadad scale was used primary. The randomisation process was judged to be adequate in only 22% of cases, and blinding was reported in just 28% of papers. The accounting for withdrawals/dropouts was present in 35% of cases. Finally, a high quality score was reached by only one third of the RCTs examined and those on oral surgery had significantly higher median quality scores, compared with all other areas grouped together. No correlation between the journal impact factor and the quality of RCTs was found. This result has also been confirmed by Barbui et al.39.

So far, the reporting of orthodontic clinical trials has been judged to be insufficient to help readers to assess the validity of RCTs and the median quality of published trials has been estimated to be poor. In a survey by Harrison40, the accounting of withdrawals and allocation concealment was mostly evaded by research reports. Only 2.6% of the trials analysed presented double blinding. It must be noted that the reporting of blinding of trial personnel is frequently missing in clinical trials41 and a major concern arises when talking about blinding in certain disciplines such as orthodontics. Indeed, since orthodontic appliances have commonly different shapes or configurations, clinicians are often aware of which intervention patients are receiving, especially when testing multibracket fixed appliances. In some cases, for instance while performing trials concerning functional removable appliance, a second clinician blinded should take the measurements after patient has removed the appliance and the clinical chart has been hidden.

CONCLUSIONS

Since it is proved that RCTs provide the most reliable and valid estimates of the efficacy of the treatments tested, clinicians should focus their daily clinical practice on RCTs outcomes. Before trusting the outcomes resulting from published RCTs it is necessary to assess their quality. The number of dental RCTs has increased enormously during the last decade but, unfortunately, the quality of the trials published is variable and further efforts to promote high quality studies are needed.

Clinicians should be able to recognise high quality trials among those published every year by using the systems available. At the present time, the tools available for quality assessment have improved the quality of RCTs published and have contributed to the standardisation process in RCT quality assessment. Assessing quality of research can be quite challenging, particularly for readers who are unacquainted with research methodology. Further to the quality tools reviewed in this report, it is always important to evaluate an article by thorough and careful reading as well as by critical appraisal of all of its parts, especially the materials and methods sections. This may be difficult and time-consuming but remains probably the best way to make an in-depth evaluation of the quality of clinical research.

REFERENCES

  • 1.Rosenberg WC, Donald A. Evidence based Medicine: an approach to clinical problem solving. Br Med J. 1995;310:1122–1126. doi: 10.1136/bmj.310.6987.1122. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 2.Pocock SJ. John Wiley & Sons Ltd; Chichester: 1996. Clinical Trials – A Practical Approach. [Google Scholar]
  • 3.Verhagen AP, de Vet HCW, de Bie RA, et al. The art of quality assessment of RCTs included in systematic reviews. J Clin Epidemiol. 2001;54:651–654. doi: 10.1016/s0895-4356(00)00360-7. [DOI] [PubMed] [Google Scholar]
  • 4.Sackett DL, Straus SE, Richardson S, et al. EBM 2nd edition. Churchill-Livingstone; London: 2000. Evidence-Based Medicine: How to Practice and Teach. [Google Scholar]
  • 5.Sjogren P, Halling A. Quality of reporting randomised clinical trials in dental and medical research. Br Dent J. 2002;192:100–103. doi: 10.1038/sj.bdj.4801304. [DOI] [PubMed] [Google Scholar]
  • 6.Jadad AR. BMJ Books; London: 1998. Randomised controlled trials: a user’s guide. [Google Scholar]
  • 7.Jadad AR, Moore RA, Carroll D, et al. Assessing the quality of reports of randomized clinical trials: is blinding necessary? Control Clin Trials. 1996;17:1–12. doi: 10.1016/0197-2456(95)00134-4. [DOI] [PubMed] [Google Scholar]
  • 8.Moher D, Pham B, Jones A, et al. Does quality of reports of randomised trials affect estimates of intervention efficacy reported in meta-analyses? Lancet. 1998;352:609–613. doi: 10.1016/S0140-6736(98)01085-X. [DOI] [PubMed] [Google Scholar]
  • 9.Schultz KF, Chalmers A, Haynes RJ, et al. Empirical evidence of bias. J Am Med Assoc. 1995;273:408–412. doi: 10.1001/jama.273.5.408. [DOI] [PubMed] [Google Scholar]
  • 10.Juni P, Witschi A, Bloch R, et al. The hazards of scoring the quality of clinical trials for meta-analysis. J Am Med Assoc. 1999;282:1054–1060. doi: 10.1001/jama.282.11.1054. [DOI] [PubMed] [Google Scholar]
  • 11.Moher D, Jadad AR, Nichol G, et al. Assessing the quality of randomized controlled trials: an annotated bibliography of scales and checklists. Control Clin Trials. 1996;16:62–73. doi: 10.1016/0197-2456(94)00031-w. [DOI] [PubMed] [Google Scholar]
  • 12.Olivo SA, Macedo LG, Gadotti IC, et al. Scales to assess the quality of randomized controlled trials: a systematic review. Phys Ther. 2008;88:156–175. doi: 10.2522/ptj.20070147. [DOI] [PubMed] [Google Scholar]
  • 13.Jadad AR, Cook DJ, Jones AL et al. The quality of randomised controlled trials included in meta-analyses and systematic reviews: how often and how is it assessed? Abstract presented at the 4th Cochrane Colloquium, Adelaide, Australia, October, 1996.
  • 14.Bhandari M, Richards RR, Sprague S, et al. Quality in the reporting of randomized trials in surgery: is the Jadad scale reliable? Control Clin Trials. 2001;22:687–688. doi: 10.1016/s0197-2456(01)00147-7. [DOI] [PubMed] [Google Scholar]
  • 15.Boutron I, Tubach F, Giraudeau B, et al. Methodological differences in clinical trials evaluating nonpharmacological and pharmacological treatments of hip and knee osteoarthritis. JAMA. 2003;290:1062–1070. doi: 10.1001/jama.290.8.1062. [DOI] [PubMed] [Google Scholar]
  • 16.Baron G, Boutron I, Giraudeau B, et al. Violation of the intent-to-treat principle and rate of missing data in superiority trials assessing structural outcomes in rheumatic diseases. Arthritis Rheum. 2005;52:1858–1865. doi: 10.1002/art.21116. [DOI] [PubMed] [Google Scholar]
  • 17.Shakespeare TP, Thiagarajan A, Gebski V. Evaluation of the quality of radiotherapy randomized trials for painful bone metastases: implications for future research design and reporting. Cancer. 2005;103:1976–1981. doi: 10.1002/cncr.20993. [DOI] [PubMed] [Google Scholar]
  • 18.Seferiadis A, Rosenfeld M, Gunnarsson R. A review of treatment interventions in whiplash-associated disorders. Eur Spine J. 2004;13:387–397. doi: 10.1007/s00586-004-0709-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19.Clark HD, Wells GA, Huet C, et al. Assessing the quality of randomized trials: reliability of the Jadad scale. Control Clin Trials. 1999;20:448–452. doi: 10.1016/s0197-2456(99)00026-4. [DOI] [PubMed] [Google Scholar]
  • 20.Latronico N, Botteri M, Minelli C, et al. Quality of reporting of randomised controlled trials in the intensive care literature: a systematic analysis of papers published in Intensive Care Medicine over 26 years. Intensive Care Med. 2002;28:1316–1323. doi: 10.1007/s00134-002-1339-x. [DOI] [PubMed] [Google Scholar]
  • 21.Kjaergard LL, Villumsen J, Gluud C. Reported methodological quality and discrepancies between large and small randomized trials in meta-analyses. Ann Intern Med. 2001;135:982–989. doi: 10.7326/0003-4819-135-11-200112040-00010. [DOI] [PubMed] [Google Scholar]
  • 22.Verhagen AP, de Vet HC, de Bie RA, et al. The Delphi list: a criteria list for quality assessment of randomized clinical trials for conducting systematic reviews developed by Delphi consensus. J Clin Epidemiol. 1998;51:1235–1241. doi: 10.1016/s0895-4356(98)00131-0. [DOI] [PubMed] [Google Scholar]
  • 23.Brouwer RW, Jakma TS, Bierma-Zeinstra SM, et al. Osteotomy for treating knee osteoarthritis. Cochrane Database Syst Rev. 2005;1:CD00401. doi: 10.1002/14651858.CD004019.pub2. [DOI] [PubMed] [Google Scholar]
  • 24.Verhagen AP, Damen L, Berger MY, et al. Is any one analgesic superior for episodic tension-type headache? J Fam Pract. 2006;55:1064–1072. [PubMed] [Google Scholar]
  • 25.Available from PEDro Physiotherapy Evidence Database: http://www.pedro.org.au/
  • 26.Brandt C, Sole G, Krause MW, et al. An evidence-based review on the validity of the Kaltenborn rule as applied to the glenohumeral joint. Man Ther. 2007;12:3–11. doi: 10.1016/j.math.2006.02.011. [DOI] [PubMed] [Google Scholar]
  • 27.Sherrington C, Herbert RD, Maher CG, et al. PEDro: a database of randomized trials and systematic reviews in physiotherapy. Man Ther. 2000;5:223–226. doi: 10.1054/math.2000.0372. [DOI] [PubMed] [Google Scholar]
  • 28.Moher D, Jadad AR, Tugwell P. Assessing the quality of randomized controlled trials. Current issues and future directions. Int J Technol Assess Health Care. 1996;12:195–208. doi: 10.1017/s0266462300009570. [DOI] [PubMed] [Google Scholar]
  • 29.Juni P, Altman DG, Egger M. Systematic reviews in health care: assessing the quality of controlled clinical trials. Br Med J. 2001;323:42–46. doi: 10.1136/bmj.323.7303.42. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 30.Available from CONSORT transparent reporting of trials website: http://www.consort-statement.org/
  • 31.Moher D, Jones A, Lepage L. Use of CONSORT statement and quality of reports of randomized trials: a comparative before-and-after evaluation. J Am Med Assoc. 2001;285:1992–1995. doi: 10.1001/jama.285.15.1992. [DOI] [PubMed] [Google Scholar]
  • 32.Parent N, Hanley JA. Assessing quality of reports on randomized clinical trials in nursing journals. Can J Cardiovasc Nurs. 2009;19:25–39. [PubMed] [Google Scholar]
  • 33.Andersson D, Samuelsson K, Karlsson J. Treatment of anterior cruciate ligament injuries with special reference to surgical technique and rehabilitation: an assessment of randomized controlled trials. Arthroscopy. 2009;25:653–685. doi: 10.1016/j.arthro.2009.04.066. [DOI] [PubMed] [Google Scholar]
  • 34.Karri V. Randomised clinical trials in plastic surgery: survey of output and quality of reporting. J Plast Reconstr Aesthet Surg. 2006;59:787–796. doi: 10.1016/j.bjps.2005.11.027. [DOI] [PubMed] [Google Scholar]
  • 35.Xu L, Li J, Zhang M, et al. Chinese authors do need CONSORT: reporting quality assessment for five leading Chinese medical journals. Contemp Clin Trials. 2008;29:727–731. doi: 10.1016/j.cct.2008.05.003. [DOI] [PubMed] [Google Scholar]
  • 36.Al-Namankany AA, Ashley P, Moles DR, et al. Assessment of the quality of reporting of randomized clinical trials in paediatric dentistry journals. Int J Paediatr Dent. 2009;19:318–324. doi: 10.1111/j.1365-263X.2009.00974.x. [DOI] [PubMed] [Google Scholar]
  • 37.Dumbrigue HB, Al-Bayat MI, Ng CC, et al. Assessment of bias in methodology for randomized controlled trials published on implant dentistry. J Prosthodont. 2006;15:257–263. doi: 10.1111/j.1532-849X.2006.00115.x. [DOI] [PubMed] [Google Scholar]
  • 38.Jokstad A, Esposito M, Coulthard P, et al. The reporting of randomized controlled trials in prosthodontics. Int J Prosthodont. 2002;15:230–242. [PubMed] [Google Scholar]
  • 39.Barbui C, Cipriani A, Malvini L, et al. Validity of the impact factor of journals as a measure of randomized controlled trial quality. J Clin Psychiatry. 2006;67:37–40. doi: 10.4088/jcp.v67n0106. [DOI] [PubMed] [Google Scholar]
  • 40.Harrison JE. Clinical trials in orthodontics II: assessment of the quality of reporting of clinical trials published in three orthodontic journals between 1989 and 1998. J Orthod. 2003;30:309–315. doi: 10.1093/ortho/30.4.309. [DOI] [PubMed] [Google Scholar]
  • 41.Haahr MT, Hróbjartsson A. Who is blinded in randomized clinical trials? A study of 200 trials and a survey of authors. Clin Trials. 2006;3:360–365. doi: 10.1177/1740774506069153. [DOI] [PubMed] [Google Scholar]

Articles from International Dental Journal are provided here courtesy of Elsevier

RESOURCES