Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2016 May 1.
Published in final edited form as: Osteoarthritis Cartilage. 2015 May;23(5):677–685. doi: 10.1016/j.joca.2015.03.011

Key analytic considerations in design and analysis of randomized controlled trials in osteoarthritis

Elena Losina 1,2,3,*, Jonas Ranstam 4,*, Jamie Collins 1,2, Thomas J Schnitzer 5, Jeffrey N Katz 1,2
PMCID: PMC4772721  NIHMSID: NIHMS747799  PMID: 25952341

Abstract

Objective

To highlight methodologic challenges pertinent to design, analysis, and reporting of results of randomized clinical trials in OA and offer practical suggestions to overcome these challenges.

Design

The topics covered in this paper include subject selection, randomization, approaches to handling missing data, subgroup analysis, sample size, and issues related to changing design mid-way through the study. Special attention is given to standardizing the reporting of results and economic analyses.

Results

Key findings include the importance of blinding and concealment, the distinction between superiority and non-inferiority trials, the need to minimize missing data, and appropriate analysis and interpretation of subgroup effects.

Conclusion

Investigators may use the findings and recommendations advanced in this paper to guide design and conduct of randomized controlled trials of interventions for osteoarthritis.

Keywords: randomized controlled trials, design, outcomes, reporting, economic evaluation, missing data

Introduction

Researchers who design and conduct randomized controlled trials to establish the efficacy of treatments for OA should pay special attention to several design features highlighted in this paper: changes in trial design, blinding, defining placebo, choice of primary outcome and optimal time of outcome assessment, and prevention of informative censoring. We provide recommendations to address these issues in the design and analytic stages (Table I). We also highlight standardized reporting, economic evaluation alongside clinical trials, and value of information analysis for prioritization of research.

Table I.

Recommendations to address methodologic challenges in osteoarthritis clinical trials.

Domain Recommendation
Design Analysis
Selection of primary endpoint Clearly define primary outcome (endpoint), means of measuring, timing of assessment. Ensure that selected outcome is acceptable by regulatory organizations and scientific community. Do not select outcomes that did not undergo a thorough evaluation regarding validity and reliability Plan and execute primary analysis focused on selected a priori primary outcome, which should be clearly and unambiguously defined. Always report results on primary outcome.
Choice of study design Decision between superiority and non-inferiority designs should be governed by the novelty of treatment modality and choice of control group. If the treatment modality under investigation is compared to placebo – superiority design is the design of choice.
In situations where the treatment under investigation is compared to another active treatment non- inferiority design are chosen if new treatment is likely to have similar efficacy but offers better tolerability and safety profile
Reporting should be consistent with chosen study design. The results of superiority and, in particular, non-inferiority trials are best presented using two-sided 95% confidence intervals.
Blinding and Allocation Concealment When possible, both participants and the research team should be blinded to treatment assignment. When this is not possible, those ascertaining the outcome and analyzing the data should be blinded. Design features such as a varying block randomization scheme and sequentially numbered, opaque, sealed envelopes should be utilized to ensure allocation concealment. Unblinding of data should not take place until the trial is terminated and data are cleaned to an acceptable level of quality. If an interim analysis is performed, this must be a completely confidential process by independent statisticians and conducted with blinded data.
Randomization Randomization should be performed using a computer random number generator, incorporating strategies to ensure concealment (e.g. varying random blocks, sequentially numbered, opaque sealed envelopes). Primary analysis should use intention to treat principal, at least in superiority trials. The intention-to-treat population plays a slightly different role in non-inferiority trials. It is usually a good recommendation to present results for both the intention-to-treat population and the per-protocol population.
Sample size Careful sample size calculations should be made before the start of the RCT, and should take into account the hypothesized difference in primary outcome between the treatment and control groups and the variability in outcome. It is important that investigators consider an effect size that is realistic, but also clinically important. Sample size calculations should also take into account secondary outcomes and subgroup analyses. In some cases analytic sample size calculations cannot be easily performed because of methodological complexity. It may then be a good alternative to use sample size simulation instead.
Missing data Strategies should be put in place at the design phase to minimize missing data. These include limited burden of data collection on study participants, incentives, and clear protocols for study staff to follow for contacting participants. Detailed reason for dropout should be recorded, and participants wishing to discontinue their assigned intervention should be given the opportunity to continue study assessments. Last observation carried forward and complete case analyses require strong assumptions and should be avoided unless there is strong justification. Likelihood based approaches such as mixed-effects models and multiple imputation are valid approaches when the missing data is MAR. Sensitivity analyses should be utilized to assess the robustness of the results to this assumption.
When to consider design changes Design changes should be kept to a minimum and be pre-specified in the study protocol. Such pre-specified design changes are usually initiated by the results of an interim analysis, especially regarding the sample size and termination of a trial. Interim analyses may be logistically complicated as they usually require independent statistical analysis. Avoid interim analyses, especially if they are not pre-specified.
Reporting The study protocol can be published in a journal, and a short version of it should be registered at a public trial registry accepted by the ICMJE and WHO prior to the first patient’s randomization. The statistical analysis and results reporting should comply with the CONSORT statement’s requirements and methodological recommendations provided by regulatory authorities. When submitting a manuscript to a scientific journal, include a copy of the study protocol and a completed CONSORT statement checklist and flow chart.

Osteoarthritis trials

Pharmacological treatment for OA consists both of symptom modifying and structure modifying drugs. Non-pharmacologic treatments generally focus on symptomatic relief and functional restoration, but can also be evaluated for structural modification, using biochemical or imaging markers. For example, the IDEA trial focused on efficacy of weight management and exercise for pain reduction, but also showed impact on inflammatory and MRI biomarkers1.

For structure modifying interventions, clinical endpoints are also important. For example, an endpoint of the structure modifying intervention may be time to joint replacement. The bona fide clinical endpoint (e.g. reduction in joint space) may be substituted by a biomarker if the treatment’s effect on this biomarker correlates with the treatment’s effect on the bona fide endpoint. This correlation, the surrogacy value of the biomarker, must be shown before the biomarker is used as a primary endpoint2.

Trials of symptom and structure modifying treatments typically have a randomized, double blind, parallel group design. These trials are recommended to be placebo controlled. Symptom modifying treatments are also recommended to have an established active comparator as a ‘usual care’ control3. Due to ethical considerations the comparison with placebo in the presence of another active treatment that has been shown to be more efficacious than placebo is not advisable. The goals of the equivalence or non-inferiority trials are to show that the new treatment is no worse than accepted active comparator, or, in the case where new treatment is compared with placebo, that all treatment effect is just due to placebo. The planning and conducting of such trials may include discussions about interim analyses and design changes. To ensure success of the trial, it is critical to define the key clinical hypothesis of interest and design the trial to provide conclusive results regarding this hypothesis.

Allocation concealment and blinding

The purpose of concealed allocation is to avoid selection bias. Investigators should provide sufficient detail on how allocation concealment was achieved so the reader can determine the likelihood of success. The assessment of treatment concealment is especially relevant in RCTs with subjective patient reported outcomes. Allocation concealment means that the person implementing randomization does not know or can’t predict the next treatment allocation. It reinforces the value of randomization in reducing selection bias.

Blinding refers to the fact that the study participant does not know what treatment he/she is getting. Double blinding means that participants and investigators are unaware of treatment assignment; this may be challenging to achieve in studies of physical therapy, behavioral therapy, weight management, and exercise. When it is hard to blind study participants or intervention providers to the treatment assignments, several steps can be taken: 1) those who ascertain outcomes should be blinded to treatment arm allocation; 2) study participants should be advised not to discuss treatment details with outcome assessors; 3) providers delivering the intervention should not participate in outcome assessment; and 4) in studies with active and control interventions, both arms should be portrayed to participants as ‘active’ and intended to reduce symptoms or improve function.

We recommend that a plan for allocation concealment should be an integral part of the manual of operating procedures (MOOP). The minimum methodological standards of ensuring allocation concealment include randomization schema using varying block size implemented centrally; and, sequentially numbered, opaque, sealed envelopes (SNOSE). Additional means of ensuring allocation concealment include: protocols that clearly state that the informed consent and baseline assessment are done prior to randomization, inclusion of reports describing means of allocation concealment in the manuscripts and report of baseline comparisons between arms on the established prognostic factors.

Randomization and stratification

Another rationale for randomization is to avoid possible dependencies between consecutive patients and obtain valid variance estimates. Randomization prevents systematic imbalance of measured and unmeasured baseline characteristics. To avoid random imbalance in important prognostic factors, the randomization should be stratified on these factors. This design issue should then also be taken into account in the statistical analysis. We recommend, depending on sample size, to stratify the randomization on a few known prognostic factors, taking into consideration that stratification without blocking is ineffective.

Placebo effect vs. trial effect

Often, individuals seek trial participation when their symptoms flare. Baseline flare can affect interpretation of study results since OA pain often is episodic. Building in an observation period of 1–2 weeks could help to distinguish the ‘trial’ effect from placebo or attention control effect. During this initial observation period, all participants could be exposed to attention control.

Regression to the mean occurs when the inclusion criteria are broad and some participants have an extreme initial value of a variable that is imperfectly correlated with subsequent measurements due to measurement error or biological variation. If these participants are measured again we may observe that the mean of this “extreme” subgroup is closer to the mean of the original sample. For example, a first high blood pressure measurement on a subject will tend to be closer to the average at the next measurement of that subject. The same can occur with a serum biomarker. The more the value deviates from the population mean and the lower the correlation between the initial and subsequent measurement, the larger the regression4. As a consequence, within-group change could be exaggerated and difficult to interpret. We recommend that the evaluation of treatment effects should thus not be based on within-group change but on between-group comparisons. In addition, the pre-treatment observation period can help distinguish true treatment effects from regression to the mean.

It is imperative to distinguish achieving symptomatic relief vs. sustaining pain relief. The initial observation period is helpful to reduce the impact of flare on efficacy in achieving relief. There is presently no accepted standard for minimal duration of effect required to establish sustainability; this issue is worthy of discussion in the clinical trial community. We recommend that trial planning activity should include careful collection of preliminary data or secondary analysis of prior studies to establish a reliable pattern of pain trajectories over time, in studies where pain is the primary outcome.

Choice of outcomes

It is important to center the study on outcomes that are meaningful to the patient and lead to appreciable improvements in quality of life. One lively debate regarding the choice of primary outcome centers on observed measures of functional performance vs. self-reported measures of pain, function, and other domains. We summarize key elements of performance-based and self-report outcomes to help guide such deliberations.

Performance-based or objective outcomes

These measures involve an observer-based assessment of specific tasks such as walking a set time or distance, arising from a chair, and the like. A recent meta-analysis showed that several of these measures including 40m self-paced walking test and timed up and go tests had the best psychometric properties among unidimensional performance based measures5.

Patient-oriented or subjective outcomes

More recently, greater attention has been given to ascertainment of outcomes that are meaningful and important to patients with OA. In contrast to performance-based outcomes, the patient reported measures assess patient perceptions of functional status, pain, and other domains. Several established self report measures used frequently in OA research include Western Ontario and McMaster Universities Osteoarthritis Index (WOMAC) for lower extremity arthritis, the Knee Injury and Osteoarthritis Score (KOOS) and Hip Injury and Osteoarthritis Score (HOOS), and more recently the PROMIS-29 (and other aggregations of PROMIS measures)68. PROMIS-29 is a composite score that encompasses overarching domains that OA is likely to affect, including functional limitations, pain interference and intensity, ability to fulfill desired social roles, anxiety and depression, sleep disturbance and fatigue. It is noteworthy that the PROMIS initiative also includes a suite of assessments conducted using computer adaptive testing and invoking item response theory. Each participant is given a limited number of items, with the degree of difficulty of the item based upon responses to prior items. We recommend that the rationale for primary endpoint be clearly delineated in study protocol alongside with the validated means of measuring the primary endpoint.

Planning trials/sample size calculations

Sample size refers to the minimal number of participants required to establish efficacy with a specified level of confidence that the observed difference between intervention and control cannot be ascribed to chance. Careful sample size considerations require clear understanding of the magnitude of the effect that is perceived as clinically important. This ensures that an unnecessarily large sample will not lead investigators to overstate the importance of small effects that are unlikely to be clinically meaningful, even with p-values below a critical threshold (e.g. 0.05)9, 10. Conversely, ensuring that the trial is appropriately powered will allow investigators to distinguish between a negative trial, meaning that the data did not suggest that the outcome differs between the control and intervention arms, and an underpowered trial, in which clinically meaningful differences were observed but did not achieve statistical significance. That is, “Absence of evidence is not evidence of absence.” The adequacy of the sample size has of course also important economical and ethical implications.

Technically, the assessment of sample size is usually based on a two-sided hypothesis test of the primary endpoint (or on a two-sided confidence interval). If a one-sided test (or confidence interval) will be used in the analysis, the ICH recommendation is to use a significance level of half the one that would have been used for a two-sided test (and equivalently for a one-sided confidence interval). The sample size is then calculated from the defined Type I error rate (false positive; usually 5%) and type 2 rate (false negative; usually 10–20%), an estimate of the endpoint’s variability, and the smallest treatment effect to be detected in the trial.

As several different patient number scenarios are required to evaluate the consequences of the uncertainty in the used estimates, to achieve a robust study design, a computer program for sample size calculation can be recommended for the calculation. Such programs are included in most statistical software packages.

Feasibility assessment

Sample size estimation should be accompanied by parallel assessment of study feasibility. The feasibility assessment should take into consideration the number of eligible patients seen in a particular clinic or other setting over the course of a specified unit of time (e.g. week, month, year), the intensity of the study (time commitment, intensity of the intervention), and the willingness of eligible participants to take part in an RCT11, 12. The feasibility assessment offers critical insight into the anticipated duration of enrollment and necessity to recruit additional centers.

Subgroup analysis and multiplicity issues

A subgroup analysis addresses the hypothesis that the efficacy of the regimen under study may be greater (or worse) in particular subgroups. Such a hypothesis implies that treatment success is greater in the presence of certain factors than in the absence of those factors. Examples might include a specific trait such as severe malalignment or obesity in persons with knee OA or number of affected joints in persons with hand OA1315.

As recommended by the CONSORT guidelines (Consolidated Standards of Reporting Trials checklist to improve the quality of reporting of RCTs)16, only pre-planned subgroup analyses should be undertaken. Multiple testing increases the chances of a false positive result (Type I error), but the Type I error rate can be controlled by adjusting the statistical significance level of each analysis according to the number of tests conducted. It is necessary to have a pre-specified description of confirmatory analyses documented in the study protocol or the statistical analysis plan (SAP). A simple Bonferroni correction for multiple testing requires that the threshold for significance be divided by the number of tests. For example, a study with 5 pre-planned subgroup analyses would adjust the significance threshold to 0.05/5 = 0.01 for each analysis in order to maintain an overall Type I error of 0.0517. The Bonferroni-Holm procedure is a less conservative option for multiple testing adjustment18. All confirmatory analyses should be stated in the original registration of the RCT on ClinicalTrials.gov or other trial registries.

Planning for subgroup analyses alongside an RCT often requires a larger sample size. Studies powered to detect a main effect are unlikely to have sufficient sample size to detect interaction effects or establish the presence of a prognostic factor with sufficient levels of statistical certainty. Caution should be exercised in interpreting and reporting the results of subgroup analyses when the study has not been powered to detect an interaction effect. In these circumstances, lack of statistical significance due to low statistical power can be misinterpreted as evidence of the absence of the differential efficacy among subgroups19. Additional non-pre-specified subgroup analyses can also be performed, but should be clearly reported as exploratory.

We recommend that the decision to conduct a pre-specified subgroup analysis in RCTs should be justified by prior data suggesting that a potential prognostic factor either promotes or inhibits the hypothesized mechanism of effect of the intervention. Confirmation that a potential prognostic factor indeed is associated with differential efficacy should be supported by a formal statistical test for interaction20.

Missing data prevention and handling

To minimize dropouts, investigators may discuss during the design phase the minimum follow-up duration that would be clinically meaningful without straining both study participants and personnel. Ways to minimize study dropouts include offering study participants the opportunity to complete visits in person or by phone, accommodating participant schedules, offering reimbursement for travel and/or parking, and modest incentive stipends for participation. It is also important to establish protocols to ensure that research staff exert the maximal effort permitted by the governing ethics body to obtain follow up data from participants. Clear protocols for research staff to follow in contacting study participants, such as the number of contact attempts, should be outlined at the beginning of the study21.

Detailed reasons for dropout should be recorded if a participant requests withdrawal. Reasons may include a participant’s assessment that he/she is too busy to continue participating, lack of efficacy, side effects (e.g., injury sustained during exercise), etc. This level of detail is especially important if reasons for dropout differ between treatment arms. Collecting detailed reasons for dropout will allow investigators to better understand the dropout mechanism and assess whether or not dropout is related to outcome22. Participants wishing to discontinue their assigned intervention should be offered the opportunity to continue study assessments, facilitating intention-to-treat inferences. Intention-to-treat implies that subjects are analyzed according to the arm randomized, irrespective of the actual treatment they received. The number of participants dropping out and reasons for dropout should be reported for each arm.

Missing data could lead to reduction of the sample size which in turn may affect statistical power of the study. Further, missing observations, if available, could have more extreme values that would, in turn affect the estimation of the variability of the effect and could artificially narrow the confidence interval for the treatment effect.

While last observation carried forward (LOCF) and complete case (CC) analytic approaches are frequently used in the analysis of RCTs, both require strong, and often unrealistic, assumptions about the missing data mechanism. In particular, both methods require that the missing data be missing completely at random (MCAR) -- that is, that the missing data are completely independent of observed or unobserved measurements (dependent or independent variables). This implies that that study completers are a random sample of the original study cohort. The less restrictive missing at random (MAR) assumption is often more reasonable. MAR assumes that dropout may depend on observed outcomes or covariates, but does not depend on unobserved data. For example, in a study of physical function in patients with knee OA, older participants may be more likely to miss follow-up visits. Older patients may also have more functional limitations. Thus, simply examining observed function could lead to a biased estimate – observed function would be an overestimate of physical function in the cohort, since older patients, with lower function, are more likely to be missing. However, if given a participant’s age, missingness is independent of function, the data are said to be missing at random. That is, the missingness depends on observed data (age) but not the missing data (function).

Under the MAR assumption, likelihood-based approaches such as mixed-effects models will produce valid inferences. Multiple imputation (MI) is another strategy that is valid under MAR. In this case, a range of plausible values is imputed for the missing outcomes using a prediction equation. This equation can incorporate ancillary information about missing data, including observed outcomes up until time of dropout, and reason for dropout. One of the reasons MI is superior compared to simple imputation methods, such as Last Observation Carried Forward, is that LOCF leads to underestimated variance and inflated statistical significance.

While there is no statistical test to verify the MAR assumption, its appropriateness should be investigated by reviewing the dropout reasons22.22. It is possible that data will be missing not at random (MNAR), meaning that missingness is related to unobserved outcomes. This is also termed informative censoring. Sensitivity analyses should be utilized to assess the robustness of the results to this assumption.

Since there is no simple statistical solution to handling data missing not at random, we underscore the importance of minimizing missing data at the design and implementation phases of RCTs.

Interim analysis

Interim analyses may compromise the scientific value of an RCT by inflating the Type I error rate. Also, conditioning the effect estimation on interim observations may bias both the effect estimate and its variance if the trial is terminated early. Interim analyses should therefore generally be avoided. If an interim analysis is planned, it should be described in the study protocol and performed in a way that protects the Type I error rate.

We recommend that the performance of the interim analysis must be a confidential process authorized and conducted by an independent statistician and decision making committee, such as Data Safety and Monitoring Board. It also means that multiplicity adjustments are likely to be necessary.

Stopping a trial early for reasons of efficacy or safety concerns

Interim analyses are used to interpret accumulating information during a trial, often to investigate whether to terminate the trial at an early stage. The analysis can reveal whether the treatments already have convincingly established different - or similar - treatment effects, or that they have too severe side effects.

The group sequential trial design allows repeated testing of treatment effect, but trials designed with an interim analysis as part of a conventional design are often planned with only one interim analysis. The decision to stop a trial early should focus both on the primary endpoint as well as consistency among secondary endpoints and safety data.

Some participants may have incomplete follow up at the time of the interim analysis and thus not be included in the interim analysis. When their follow up is completed, and they are included in a new analysis, the outcome of the trial may differ. When an interim analysis leads to early stopping, the final results that include these additional subjects are considered more important than the results from the interim analysis, which excluded these subjects.

Futility stopping

An interim analysis may also include an evaluation of whether or not the trial can achieve its objectives. If the interim results indicate that the trial is unlikely to achieve statistical significance, stopping it can save important resources23. Such futility testing is often performed on the primary or a suitable intermediate endpoint and based on forecasting the outcome of a reference test using conditional or predictive power. It is important to recognize that Type I error usually is not protected if a trial is continued when the interim analysis suggests stopping the trial for futility24.

Change in Trial Design

Changes that occur after the first patient is randomized can compromise the scientific integrity of a trial. In some cases, however, such changes can lead to better use of resources or shorter study time. An adaptive trial is based on a study protocol that includes detailed description of one or more pre-planned interim analyses and corresponding design modifications that can be made with full control of the Type I error rate24. Adaptive trials can increase the effectiveness and reduce the costs of new treatments, but they have limitations2. When considering modifying an ongoing trial the complete trial recommendations of the International Conference on Harmonization of Technical Requirements for Registration of Pharmaceuticals for Human Use (ICH) recommendations should be taken into account2.

Reassessing sample size

When an interim analysis neither shows superiority with regard to efficacy, nor suggests that a continuation of the trial would be futile, a sample size reassessment may be useful. Sample size calculation performed with an empirical based variance estimate is likely to be more informative than one performed with a hypothetical variance. However, a sample size reassessment that is based on the result of an ongoing trial may inflate the Type I error rate. Adjustment to account for the inflated Type I error will increase the necessary sample size. Thus, a procedure that does not affect the Type I error rate should be used, such as one that does not require unblinding of the data. Sample size re-estimation should be done by a study statistician, with the rest of the investigative team blinded

Since unblinding trial data in an interim analysis has several disadvantages, we recommend that it should be avoided when reassessing sample size. Several methods exist, the simplest one just ignoring treatment allocation and using a one-sample variance estimate2527. This variance estimate may be biased when a treatment effect exists, but a bias adjustment can be made with the treatment effect assumed (under alternative hypothesis) in the sample size calculation when calculating the type 2 error rate27.

Superiority and non-inferiority

Statistically nonsignificant outcomes in trials designed to show that one treatment is superior to another are often wrongly interpreted as a proof of no difference between treatments. Equality of treatment effects can never be proven with statistical methods. Instead, equivalence and non-inferiority trials are designed to show that two treatments are not “too different”, or that a new treatment is not “unacceptably worse” than a standard one28. While the null hypothesis in a superiority trial is that treatment effects are identical, the null hypothesis in an equivalence or non-inferiority trial is defined with reference to a specific difference in treatment effects, the equivalence, or non-inferiority, margin. The thresholds for not “too different” or not “unacceptably worse” are defined using clinical criteria. When the research team envisions that both superiority and non-inferiority approaches might be appropriate, the trial should be powered as a non-inferiority trial, and the switching from non-inferiority to superiority should be planned and described in the protocol a priori. The superiority and non-inferiority approaches also differ with regard to the role of the study population. As the ITT population tends to be less conservative than the per-protocol population in non-inferiority trials, the latter is often chosen as the primary study population.

We recommend that if, after a successful interim analysis, a non-inferiority trial is continued in order to show superiority, the final conclusion of the trial should be based on all collected data, even if this result happens to be less supportive than the one from the interim analysis.

Changing primary endpoint

If important scientific knowledge has been gained after the initiation of a trial, it may be appropriate to incorporate this knowledge into the trial29. For example, consider a pharmaceutical trial in OA patients in which the primary outcome is the change in a serum biomarker associated with radiographic progression. If a valid, and more responsive biomarker became available during the course of the trial, it might be reasonable to consider adding this new biomarker to the trial endpoints.

The main principle when considering changing the primary endpoint is whether the change is independent of the data collected in the trial. In some large trials with long duration, it may be appropriate to change an endpoint even if the change depends on data. The trial would then have to be divided into two stages, the first being hypothesis generating (endpoint identifying), the second hypothesis testing (confirmatory), and the outcome of the trial relying entirely on the second stage. The design change, like other major changes, should be documented in an amendment to the study protocol.

Other major design changes

Other changes, such as treatment duration, co-medications, changes of inclusion or exclusion criteria, may be necessary in an ongoing trial. In some cases, the changes affect the sample size. If such changes are made, a formal protocol amendment should be made, and the primary analysis should be stratified according to whether the patients were randomized before or after the change24. Homogeneity of the results should be carefully investigated, but a combination of the results would still require careful argumentation.

Analyzing and reporting RCT data

Analysis and reporting of results from a RCT should be conducted according to the guidelines proposed by the CONSORT group for pharmacologic studies and CLEAR NPT (CheckList to Evaluate A Report of a NonPharmacological Trial) group for non-pharmacologic studies using the intention-to-treat (ITT) analysis as the primary analysis12, 16, 3038. In the intention- to- treat analysis, all study participants are analyzed according to treatment arm to which they were randomized, taking advantage of the balance in measured and unmeasured factors.

Study teams have an ethical obligation to report the results of their research activities39. The manner in which study data are reported has evolved significantly over the past 20 years, and there is general consensus today of the basic requirements. All studies should be prospectively entered into an appropriate clinical trials registry. There are numerous country and region-specific registries, and the WHO maintains the International Clinical Trials Registry Platform (ICTRP), whose main aim is “to facilitate the prospective registration of the WHO Trial Registration Data Set on all clinical trials, and the public accessibility of that information.”26 In the United States, registration is required for trials that meet the Food and Drug Administration Amendments Act (FDAAA) definition of an “applicable clinical trial”40. “Applicable Clinical Trials” include trials of drugs, biologics and devices, other than phase 1 studies, of agents that are subject to FDA regulation and that have one or more sites in the United States, are conducted under an FDA Investigational New Drug (IND) or Investigational Device Exemption (IDE) or involves a drug, biologic or device manufactured in the United States and exported for research. Trials must be registered within 21 days of the first participant being enrolled. Clinicaltrials.gov is a web-based registry resource25 developed as a consequence of the Food and Drug Administration Modernization Act of 1997 (FDAMA)41 and maintained by the National Library of Medicine, a part of the National Institutes of Health. The data entered are publicly available. The scope of clinical trials that must be registered and the information provided, including study results, has been significantly expanded since its initiation. Timely registration of clinical trials has become a requirement from the majority of scientific journals for acceptance of manuscripts reporting study results and is a long-standing recommendation of the International Committee of Medical Journal Editors (ICMJE)42.

Less consensus exists at this time regarding what specific clinical trials information should be made publicly available. Internationally, the WHO Registry Platform Working Group on the Reporting of Findings of Clinical Trials supports that “the findings of all clinical trials must be made publicly available” but has provided no specific template or recommendations43. In the United States, Section 801 of the FDAAA requires the submission of “basic results” for certain clinical trials, generally not later than one year after study completion44. As a consequence, the clinicaltrials.gov results database was implemented and allows the submission of data in a standard, tabular format consisting of sections dealing with participant flow, baseline characteristics, outcome measures and statistical analyses and adverse events. The sponsor or principal investigator is responsible for submitting these data. Details are provided at http://prsinfo.clinicaltrials.gov/ and the article by Tse et al44.

The information content to be reported in publications continues to elicit discussion, with the CONSORT statement45 currently serving as a widely accepted standard for the reporting of clinical trials. The statement includes an evidence-based set of recommendations, presented as a 25-item checklist and flow diagram. An accompanying “explanation and elaboration document”46 provides the principles upon which the recommendations were generated and offers valuable information to assist in ensuring the quality of reporting of clinical trials in the literature. Extensions of the CONSORT statement to randomized controlled trials that differ from the standard parallel-group design have recently been published6. Many scientific journals now strongly suggest or require that submitted manuscripts adhere to the CONSORT principles, checklist and flow diagram. Details regarding the CONSORT group and its activities are available from its website47. Guidelines for data content and the structure of reports required for filings to regulatory authorities are beyond the scope of this update but can be accessed by those interested (guideline document of ICH48 and guidance document of FDA18).

Economic evaluations alongside RCTs in osteoarthritis

When trial results indicate that an intervention is efficacious, policy makers must decide whether the treatment offers good value for the additional resources spent4953. To address this question, investigators may consider a formal cost-effectiveness or cost-utility analysis alongside clinical trials or beyond the timeframe of the trial54, 55.

OA treatments generally affect quality of life, not longevity. Therefore, investigators considering economic evaluation alongside an RCT should include appropriate measures of quality of life in the data collection instruments. In economic evaluations, quality of life is quantified by utilities ranging from 0 to 1, with 1 corresponding to perfect health and 0 to death56, 57. Investigators should decide whether assessment of utilities will be done directly, using the Standard Gamble or Time-Trade-Off method58, or derived indirectly using instruments such as EuroQol or short form health surveys (SF36 or SF12), WOMAC, with cross-mapping of health states assessed with these measures to population-based repositories of preference-based utilities5963.

Investigators should collect data related to subjects’ health care utilization and select a time frame over which patients can reliably recall their utilization of health care. It is of paramount importance to decide on the perspective (societal, government, payer, provider) that informs the collection of the cost data. Questions related to health care utilization should encompass visits to primary doctors and specialists, medication use, hospitalizations, and emergency department visits. It is important to account for direct medical and indirect costs. Direct medical costs include pharmacological and non-pharmacological regimens, ambulatory and emergency department visits and inpatient stays. Indirect costs capture productivity losses due to a medical condition.

The Panel on Cost-Effectiveness in Health and Medicine, recommends that cost-effectiveness analyses (CEA) be done over the long term, potentially the lifetime, to capture long-term consequences of treatment strategies. For example in the case of OA, a key economic and quality of life outcome is total knee replacement. Often, to conduct a CEA over a long time duration, data from an RCT need to be augmented by data from other sources. Decision analysis modeling should be employed in addition to commonly used statistical methods to conduct a CEA over the lifetime horizon.

Careful attention should be given to the investigation of uncertainty in parameter estimates. Depending on the type of economic evaluation (stochastic evaluation alongside of RCT or model based evaluation, beyond the time frame of the RCT) different types of uncertainty should be considered. For the evaluation alongside of the RCT, there are four main types of uncertainty: methodological, sampling variation, extrapolation, and generalizability/transferability. The main methods to handle methodological and generalizability/transferability concerns include the one- and multi-way sensitivity analyses. Parameter uncertainty is more relevant to model-based evaluations. The reduction of uncertainty through additional research may improve decisions but comes at a cost. Probabilistic sensitivity analyses followed by formal value of information (VOI) analyses, a formal methodology that is designed to facilitate establishing research priorities, evaluate the impact of uncertainty, and the value of future research to reduce uncertainty64.

The traditional cost-effectiveness analysis is designed to address whether a treatment offers good value but does not provide insight into affordability. This issue can be addressed with budget impact analysis (BIA). The goal of BIA is to quantify the financial consequences of adopting the specific treatment strategy (such as a weight loss program) by various payers, including insurance organizations, health care systems, and government, given real life resource constraints. The BIA permits projection of how adoption of the treatment could impact spending for specific payers. The results of BIA are often used for budget planning and changes in health insurance premiums.

Acknowledgments

Role of the funding source

This work was supported in part by the National Institutes of Health, National Institute of Arthritis and Musculoskeletal and Skin Diseases grants K24AR057827, R01AR064320, P60 AR047782.

OARSI gratefully acknowledges support to defer in part the cost of printing of the OARSI Recommendations for Conducting Clinical Trials in Osteoarthritis from: Abbvie, BioClinica, Boston Imaging Core Lab, and Flexion.

Footnotes

Conflict of Interest

Elena Losina –none

Jonas Ranstam –none

Jamie Collins – none

Jeffrey N. Katz –none

Author Contributions

All authors contributed to the writing and revision of the manuscript and approved the final version.

Publisher's Disclaimer: This is a PDF file of an unedited manuscript that has been accepted for publication. As a service to our customers we are providing this early version of the manuscript. The manuscript will undergo copyediting, typesetting, and review of the resulting proof before it is published in its final citable form. Please note that during the production process errors may be discovered which could affect the content, and all legal disclaimers that apply to the journal pertain.

Contributor Information

Elena Losina, Email: elosina@partners.org.

Jonas Ranstam, Email: jonas.ranstam@gmail.com.

Jamie Collins, Email: jcollins13@partners.org.

Thomas J Schnitzer, Email: tjs@northwestern.edu.

Jeffrey N. Katz, Email: jnkatz@partners.org.

References

  • 1.Beavers KM, Beavers DP, Newman JJ, Anderson AM, Loeser RF, Jr, Nicklas BJ, et al. Effects of total and regional fat loss on plasma CRP and IL-6 in overweight and obese, older adults with knee osteoarthritis. Osteoarthritis and cartilage. 2015 Feb;23(2):249–56. doi: 10.1016/j.joca.2014.11.005. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 2.FDA Document: Guidelines for industry. Statistical principles for clinical trial E9; International conference on harmonisation of technical requirements for registration of pharmaceuticals for human use; 16 September 1998. [Google Scholar]
  • 3.Miller F, Bjornsson M, Svensson O, Karlsten R. Experiences with an adaptive design for a dose-finding study in patients with osteoarthritis. Contemporary clinical trials. 2014 Mar;37(2):189–99. doi: 10.1016/j.cct.2013.12.007. [DOI] [PubMed] [Google Scholar]
  • 4.Bland JM, Altman DG. Some examples of regression towards the mean. BMJ. 1994 Sep 24;309(6957):780. doi: 10.1136/bmj.309.6957.780. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5.Dobson F, Bennell K, Hinman R, Roos E, Abbott H, Stratford P, et al. OARSI recommended performance-based tests to assess physical function in osteoarthritis of the hip or knee: authors’ reply. Osteoarthritis and cartilage. 2013 Oct;21(10):1625–6. doi: 10.1016/j.joca.2013.07.011. [DOI] [PubMed] [Google Scholar]
  • 6.Bellamy N, Buchanan WW, Goldsmith CH, Campbell J, Stitt LW. Validation study of WOMAC: a health status instrument for measuring clinically important patient relevant outcomes to antirheumatic drug therapy in patients with osteoarthritis of the hip or knee. J Rheumatol. 1988 Dec;15(12):1833–40. [PubMed] [Google Scholar]
  • 7.Hinchcliff M, Beaumont JL, Thavarajah K, Varga J, Chung A, Podlusky S, et al. Validity of two new patient-reported outcome measures in systemic sclerosis: Patient-Reported Outcomes Measurement Information System 29-item Health Profile and Functional Assessment of Chronic Illness Therapy-Dyspnea short form. Arthritis Care Res (Hoboken) 2011 Nov;63(11):1620–8. doi: 10.1002/acr.20591. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.Roos EM, Lohmander LS. The Knee injury and Osteoarthritis Outcome Score (KOOS): from joint injury to osteoarthritis. Health Qual Life Outcomes. 2003;1:64. doi: 10.1186/1477-7525-1-64. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9.Jaeschke R, Singer J, Guyatt GH. Measurement of health status. Ascertaining the minimal clinically important difference. Control Clin Trials. 1989 Dec;10(4):407–15. doi: 10.1016/0197-2456(89)90005-6. [DOI] [PubMed] [Google Scholar]
  • 10.Wells G, Beaton D, Shea B, Boers M, Simon L, Strand V, et al. Minimal clinically important differences: review of methods. J Rheumatol. 2001 Feb;28(2):406–12. [PubMed] [Google Scholar]
  • 11.Meinert CL. Beyond CONSORT: need for improved reporting standards for clinical trials. Consolidated Standards of Reporting Trials. JAMA. 1998 May 13;279(18):1487–9. doi: 10.1001/jama.279.18.1487. [DOI] [PubMed] [Google Scholar]
  • 12.Moher D, Schulz KF, Altman DG. The CONSORT statement: revised recommendations for improving the quality of reports of parallel-group randomized trials. Ann Intern Med. 2001 Apr 17;134(8):657–62. doi: 10.7326/0003-4819-134-8-200104170-00011. [DOI] [PubMed] [Google Scholar]
  • 13.Bolland MJ, Grey A, Reid IR. Evidence from randomized controlled trials, meta-analyses, and subgroup analyses. JAMA. 2010 Apr 7;303(13):1254. doi: 10.1001/jama.2010.366. author reply -5. [DOI] [PubMed] [Google Scholar]
  • 14.Sacristan JA. Evidence from randomized controlled trials, meta-analyses, and subgroup analyses. JAMA. 2010 Apr 7;303(13):1253–4. doi: 10.1001/jama.2010.365. author reply 4–5. [DOI] [PubMed] [Google Scholar]
  • 15.Vosk A. Evidence from randomized controlled trials, meta-analyses, and subgroup analyses. JAMA. 2010 Apr 7;303(13):1253. doi: 10.1001/jama.2010.364. author reply 4–5. [DOI] [PubMed] [Google Scholar]
  • 16.Altman DG, Schulz KF, Moher D, Egger M, Davidoff F, Elbourne D, et al. The revised CONSORT statement for reporting randomized trials: explanation and elaboration. Ann Intern Med. 2001 Apr 17;134(8):663–94. doi: 10.7326/0003-4819-134-8-200104170-00012. [DOI] [PubMed] [Google Scholar]
  • 17.Bland JM, Altman DG. Multiple significance tests: the Bonferroni method. BMJ. 1995 Jan 21;310(6973):170. doi: 10.1136/bmj.310.6973.170. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18.Holm S. A simple sequential rejective multiple test procedure. Scandinavian Journal of Statistics. 1979;6:65–70. [Google Scholar]
  • 19.Brookes ST, Whitely E, Egger M, Smith GD, Mulheran PA, Peters TJ. Subgroup analyses in randomized trials: risks of subgroup-specific analyses; power and sample size for the interaction test. Journal of clinical epidemiology. 2004 Mar;57(3):229–36. doi: 10.1016/j.jclinepi.2003.08.009. [DOI] [PubMed] [Google Scholar]
  • 20.Lagakos SW. The challenge of subgroup analyses--reporting without distorting. N Engl J Med. 2006 Apr 20;354(16):1667–9. doi: 10.1056/NEJMp068070. [DOI] [PubMed] [Google Scholar]
  • 21.Scharfstein DO, Hogan J, Herman A. On the prevention and analysis of missing data in randomized clinical trials: the state of the art. The Journal of bone and joint surgery American volume. 2012 Jul 18;94( Suppl 1):80–4. doi: 10.2106/JBJS.L.00273. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22.Little RJ, D’Agostino R, Cohen ML, Dickersin K, Emerson SS, Farrar JT, et al. The prevention and treatment of missing data in clinical trials. N Engl J Med. 2012 Oct 4;367(14):1355–60. doi: 10.1056/NEJMsr1203730. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 23.He P, Lai T, Liao O. Futility stopping in clinical trials. Statist Interface. 2012;5:415–23. [Google Scholar]
  • 24.Committee for Medicinal Products for Human Use. Reflection paper on methodological issues in confirmatory clinical trials planned with an adaptive design. London: EMEA; Oct, 2007. [Google Scholar]
  • 25.Broberg P. Sample size re-assessment leading to a raised sample size does not inflate type I error rate under mild conditions. BMC medical research methodology. 2013;13:94. doi: 10.1186/1471-2288-13-94. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 26.Chen YH, DeMets DL, Lan KK. Increasing the sample size when the unblinded interim result is promising. Statistics in medicine. 2004 Apr 15;23(7):1023–38. doi: 10.1002/sim.1688. [DOI] [PubMed] [Google Scholar]
  • 27.Kieser M, Friede T. Simple procedures for blinded sample size adjustment that do not affect the type I error rate. Statistics in medicine. 2003 Dec 15;22(23):3571–81. doi: 10.1002/sim.1585. [DOI] [PubMed] [Google Scholar]
  • 28.Lesaffre E. Superiority, equivalence, and non-inferiority trials. Bulletin of the NYU hospital for joint diseases. 2008;66(2):150–4. [PubMed] [Google Scholar]
  • 29.Evans S. When and how can endpoints be changed after initiation of a randomized clinical trial? PLoS clinical trials. 2007;2(4):e18. doi: 10.1371/journal.pctr.0020018. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 30.Boutron I, Moher D, Tugwell P, Giraudeau B, Poiraudeau S, Nizard R, et al. A checklist to evaluate a report of a nonpharmacological trial (CLEAR NPT) was developed using consensus. Journal of clinical epidemiology. 2005 Dec;58(12):1233–40. doi: 10.1016/j.jclinepi.2005.05.004. [DOI] [PubMed] [Google Scholar]
  • 31.Bromhead HJ, Goodman NW. CONSORT statement on the reporting standards of clinical trials. Reporting of refusal of consent to take part in clinical trials is still poor. BMJ. 1997 Apr 12;314(7087):1126–7. [PMC free article] [PubMed] [Google Scholar]
  • 32.Freemantle N, Mason JM, Haines A, Eccles MP. CONSORT: an important step toward evidence-based health care. Consolidated Standards of Reporting Trials. Ann Intern Med. 1997 Jan 1;126(1):81–3. doi: 10.7326/0003-4819-126-1-199701010-00011. [DOI] [PubMed] [Google Scholar]
  • 33.Junker C, Egger M, Schneider M, Zellweger T, Antes G. The CONSORT statement. JAMA. 1996 Dec 18;276(23):1876–7. author reply 7. [PubMed] [Google Scholar]
  • 34.Meade TW, Wald N, Collins R. CONSORT statement on the reporting standards in clinical trials. Recommendations are inappropriate for trial reports. BMJ. 1997 Apr 12;314(7087):1126. author reply 7. [PMC free article] [PubMed] [Google Scholar]
  • 35.Moher D, Schulz KF, Altman D. The CONSORT statement: revised recommendations for improving the quality of reports of parallel-group randomized trials. JAMA. 2001 Apr 18;285(15):1987–91. doi: 10.1001/jama.285.15.1987. [DOI] [PubMed] [Google Scholar]
  • 36.O’Toole LB. Consort statement on the reporting standards of clinical trials. MRC uses checklist similar to CONSORT’s. BMJ. 1997 Apr 12;314(7087):1127. [PMC free article] [PubMed] [Google Scholar]
  • 37.Ross SD. The CONSORT statement. JAMA. 1996 Dec 18;276(23):1877. doi: 10.1001/jama.276.23.1877b. author reply. [DOI] [PubMed] [Google Scholar]
  • 38.Schulz KF. The quest for unbiased research: randomized clinical trials and the CONSORT reporting guidelines. Ann Neurol. 1997 May;41(5):569–73. doi: 10.1002/ana.410410504. [DOI] [PubMed] [Google Scholar]
  • 39.Iannone F, De BC, Dell’Accio F, Covelli M, Cantatore FP, Patella V, et al. Interleukin-10 and interleukin-10 receptor in human osteoarthritic and healthy chondrocytes. Clinical and Experimental Rheumatology. 2001;19(2):139–45. [PubMed] [Google Scholar]
  • 40.Iannone F, De BC, Dell’Accio F, Covelli M, Patella V, Lo BG, et al. Increased expression of nerve growth factor (NGF) and high affinity NGF receptor (p140 TrkA) in human osteoarthritic chondrocytes. Rheumatology. 2002;41(12):1413–8. doi: 10.1093/rheumatology/41.12.1413. [DOI] [PubMed] [Google Scholar]
  • 41.Ichihara K, Haga N, Abiko Y. Is ischemia-induced pH decrease of dog myocardium respiratory or metabolic acidosis? The American Journal of Physiology. 1984;246:H652–H7. doi: 10.1152/ajpheart.1984.246.5.H652. [DOI] [PubMed] [Google Scholar]
  • 42.Illingworth CM, Barker AT. Measurement of electrical currents emerging during the regeneration of amputated fingertips in children. Clinical Physics and Physiological Measurements. 1980;1(1):87–9. [Google Scholar]
  • 43.Ghersi D, Clarke M, Berlin J, Gulmezoglu A, Kush R, Lumbiganon P, et al. Reporting the findings of clinical trials: a discussion paper. Bulletin of the World Health Organization. 2008 Jun;86(6):492–3. doi: 10.2471/BLT.08.053769. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 44.Tse T, Williams RJ, Zarin DA. Reporting “basic results” in ClinicalTrials.gov. Chest. 2009 Jul;136(1):295–303. doi: 10.1378/chest.08-3022. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 45.Schulz KF, Altman DG, Moher D. CONSORT 2010 statement: updated guidelines for reporting parallel group randomised trials. Int J Surg. 2011;9(8):672–7. doi: 10.1016/j.ijsu.2011.09.004. [DOI] [PubMed] [Google Scholar]
  • 46.Moher D, Hopewell S, Schulz KF, Montori V, Gotzsche PC, Devereaux PJ, et al. CONSORT 2010 explanation and elaboration: updated guidelines for reporting parallel group randomised trials. Int J Surg. 2012;10(1):28–55. doi: 10.1016/j.ijsu.2011.10.001. [DOI] [PubMed] [Google Scholar]
  • 47.Illingworth CM. Trapped fingers and amputated finger tips in children. Journal of Pediatric Surgery. 1974;9(6):853–8. doi: 10.1016/s0022-3468(74)80220-4. [DOI] [PubMed] [Google Scholar]
  • 48.Im HJ, Muddasani P, Natarajan V, Schmid TM, Block JA, Davis F, et al. Basic Fibroblast Growth Factor Stimulates Matrix Metalloproteinase-13 via the Molecular Cross-talk between the Mitogen-activated Protein Kinases and Protein Kinase C Pathways in Human Adult Articular Chondrocytes. Journal of Biological Chemistry. 2007;282(15):11110–21. doi: 10.1074/jbc.M609040200. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 49.Gillespie WJ, Pekarsky B, O’Connell DL. Evaluation of new technologies for total hip replacement. Economic modelling and clinical trials. J Bone Joint Surg Br. 1995 Jul;77(4):528–33. [PubMed] [Google Scholar]
  • 50.Jonsson B, Weinstein MC. Economic evaluation alongside multinational clinical trials. Study considerations for GUSTO IIb. Int J Technol Assess Health Care. 1997 Winter;13(1):49–58. doi: 10.1017/s0266462300010229. [DOI] [PubMed] [Google Scholar]
  • 51.Morris S. A comparison of economic modelling and clinical trials in the economic evaluation of cholesterol-modifying pharmacotherapy. Health Econ. 1997 Nov-Dec;6(6):589–601. doi: 10.1002/(sici)1099-1050(199711)6:6<589::aid-hec286>3.0.co;2-d. [DOI] [PubMed] [Google Scholar]
  • 52.Siegel JE, Weinstein MC, Russell LB, Gold MR. Recommendations for reporting cost-effectiveness analyses. Panel on Cost-Effectiveness in Health and Medicine. JAMA. 1996 Oct 23–30;276(16):1339–41. doi: 10.1001/jama.276.16.1339. [DOI] [PubMed] [Google Scholar]
  • 53.Weinstein MC, Siegel JE, Gold MR, Kamlet MS, Russell LB. Recommendations of the Panel on Cost-effectiveness in Health and Medicine. JAMA. 1996 Oct 16;276(15):1253–8. [PubMed] [Google Scholar]
  • 54.Edwards RT, Hounsome B, Linck P, Russell IT. Economic evaluation alongside pragmatic randomised trials: developing a standard operating procedure for clinical trials units. Trials. 2008;9:64. doi: 10.1186/1745-6215-9-64. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 55.Haycox A, Drummond M, Walley T. Pharmacoeconomics: integrating economic evaluation into clinical trials. Br J Clin Pharmacol. 1997 Jun;43(6):559–62. doi: 10.1046/j.1365-2125.1997.00576.x. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 56.Tsevat J, Weeks JC, Guadagnoli E, Tosteson AN, Mangione CM, Pliskin JS, et al. Using health-related quality-of-life information: clinical encounters, clinical trials, and health policy. J Gen Intern Med. 1994 Oct;9(10):576–82. doi: 10.1007/BF02599287. [DOI] [PubMed] [Google Scholar]
  • 57.Weinstein MC. A QALY is a QALY--or is it? J Health Econ. 1988 Sep;7(3):289–90. doi: 10.1016/0167-6296(88)90030-6. [DOI] [PubMed] [Google Scholar]
  • 58.Torrance GW. Measurement of health state utilities for economic appraisal. J Health Econ. 1986 Mar;5(1):1–30. doi: 10.1016/0167-6296(86)90020-2. [DOI] [PubMed] [Google Scholar]
  • 59.Brazier J, Jones N, Kind P. Testing the validity of the Euroqol and comparing it with the SF-36 health survey questionnaire. Qual Life Res. 1993 Jun;2(3):169–80. doi: 10.1007/BF00435221. [DOI] [PubMed] [Google Scholar]
  • 60.Brazier JE, Walters SJ, Nicholl JP, Kohler B. Using the SF-36 and Euroqol on an elderly population. Qual Life Res. 1996 Apr;5(2):195–204. doi: 10.1007/BF00434741. [DOI] [PubMed] [Google Scholar]
  • 61.Selai C, Rosser R. Eliciting EuroQol descriptive data and utility scale values from inpatients. A feasibility study. Pharmacoeconomics. 1995 Aug;8(2):147–58. doi: 10.2165/00019053-199508020-00006. [DOI] [PubMed] [Google Scholar]
  • 62.Wolfe F, Hawley DJ. Measurement of the quality of life in rheumatic disorders using the EuroQol. Br J Rheumatol. 1997 Jul;36(7):786–93. doi: 10.1093/rheumatology/36.7.786. [DOI] [PubMed] [Google Scholar]
  • 63.Dolan P. Modeling valuations for EuroQol health states. Med Care. 1997 Nov;35(11):1095–108. doi: 10.1097/00005650-199711000-00002. [DOI] [PubMed] [Google Scholar]
  • 64.Claxton KP, Sculpher MJ. Using value of information analysis to prioritise health research: some lessons from recent UK experience. Pharmacoeconomics. 2006;24(11):1055–68. doi: 10.2165/00019053-200624110-00003. [DOI] [PubMed] [Google Scholar]

RESOURCES