Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2020 Jul 1.
Published in final edited form as: Pharmacoeconomics. 2019 Jul;37(7):871–877. doi: 10.1007/s40273-019-00770-z

The Curve of Optimal Sample Size (COSS): a Graphical Representation of the Optimal Sample Size from a Value of Information Analysis

Eric Jutkowitz 1, Fernando Alarid-Escudero 2, Karen M Kuntz 3, Hawre Jalal 4
PMCID: PMC6556417  NIHMSID: NIHMS1521626  PMID: 30761461

Abstract

Value of information (VOI) analysis quantifies the opportunity cost associated with decision uncertainty, and thus informs the value of collecting further information to avoid this cost. VOI can inform study design, optimal sample size selection, and research prioritization. Recent methodological advances have reduced the computational burden of conducting VOI and have made it easier to evaluate the expected value of sample information, the expected net benefit of sampling, and the optimal sample size of a study design (n*). The volume of VOI analyses being published is increasing, and there is now a need for VOI studies to conduct sensitivity analyses on VOI-specific parameters. In this practical application, we introduce the curve of optimal sample size (COSS), which is a graphical representation of n* over a range of willingness-to-pay thresholds and VOI parameters (example data and R code are provided). In a single figure, the COSS presents summary data for decision makers to determine the sample size that optimizes research funding given their operating characteristics. The COSS also presents variation in the optimal sample size given variability or uncertainty in VOI parameters. The COSS represents an efficient and additional approach for summarizing results from a VOI analysis.

Keywords: value of information, optimal sample size, visualization

Introduction

Value of information (VOI) analysis quantifies the value of reducing decision uncertainty regarding the choice between competing strategies and can inform research prioritization and inform future data collection efforts [1]. Recent methodological advances have reduced computational burdens associated with conducting VOI analyses [27]. With improved efficiency, the use of VOI to inform decision-making and research prioritization has increased [813].

Findings from a VOI analysis are generally reported via several summary measures described in detail below. The expected value of perfect information (EVPI) and the expected value of partial perfect information (EVPPI) are normally graphically presented over a range of willingness-to-pay (WTP) thresholds. In contrast, VOI analyses typically only report point estimates of measures of the expected value of sample information (EVSI), the expected net benefit of sampling (ENBS), and the optimal sample size (n*), which is the sample size that maximizes the ENBS of a data collection process (i.e., research study) that aims to reduce decision uncertainty (i.e., n* at a single WTP) [1417]. However, as we illustrate below, n* can be sensitive to a decision maker’s WTP threshold. To increase the utility of VOI analyses, researchers should not only report point estimates for all VOI measures but should report n* over a range of WTP thresholds and variation in n* in VOI parameters (e.g., cost of research).

Our objective is to help analysts report n* over a relevant set of sensitivity analyses on VOI-specific parameters that can help decision makers better understand findings from VOI analyses. We introduce the curve of optimal sample size (COSS), which is a graphical representation of n* over a range of WTP thresholds and can also be used to show how n* changes based on VOI-specific parameters. The COSS shows n* in a single figure for a range of decision maker operating characteristics and it represents an additional graphical approach to summarize results from a VOI.

Review of VOI Measures

VOI analysis typically follows an economic evaluation and the core data used to inform a VOI are derived from a probabilistic sensitivity analysis (PSA) of an economic evaluation [2, 7, 3]. Detailed methodological [1820] and applied [8, 21, 22] VOI studies have been published, so we present only a brief overview of five key VOI measures that are typically used to inform a research prioritization framework. In addition, VOI analyses require many assumptions. As our objective is to introduce the COSS as a method for summarizing findings, we refer the reader to relevant literature that discuss in detail the analytical assumptions that must be made.

The first VOI measure is EVPI, which represents the maximum amount a decision maker should be willing to pay to reduce all sources of uncertainty in the model’s parameters.

The second measure is EVPPI, which estimates the value of eliminating uncertainty (within the context of the decision problem) for specific model inputs (e.g., effectiveness of treatment) or sets of inputs [2, 23, 24]. Results from an EVPPI analysis can be used to inform the upper threshold that a decision maker should be willing to pay for a data collection process (e.g., randomized trial) to collect data to eliminate uncertainty for a single input or sets of inputs. In practical terms, the actual data collection process is driven by the nature of the input(s). For example, inputs related to treatment effect typically imply a randomized trial. A key assumption of VOI is that any data collection process will result in an unbiased sample estimate of the input of interest. Thus far, the only requirements to the VOI analysis are the PSA dataset and the WTP threshold.

Up to now, we have discussed reporting EVPI and EVPPI on an individual level. To extrapolate results to a population level, several assumptions need to be made regarding the discount rate that should be applied to the future benefit of information, the decision lifetime, the number of current prevalent cases that could benefit from the decision and the number of future incidence cases that could benefit from the decision during the decision lifetime [18, 19, 25].

The third and fourth VOI measures are the population expected value of sample information (pEVSI) and the expected net benefit of sampling (ENBS). pEVSI represents the value of reducing uncertainty by a data collection process for a single input or sets of inputs by collecting data from a finite sample size of n individuals for a given WTP threshold [2, 26, 27]. Recent methodological advances support the computation of pEVSI [2, 7, 3, 6] for different sets of inputs, which can represent various data collection processes, using economic models of high complexity [28] and over different WTP threshold values [8].

There is a cost to collecting data. ENBS provides an estimate of the marginal population benefit of an additional sample collected in a specific data collection process accounting for the cost of collecting the data. To calculate ENBS, a study cost function is required. Cost functions can be simple or complex and depend on the study design and setting (e.g., single site or multicenter study). Basic cost functions typically include a fixed cost of conducting the research study and a variable or per-person cost of conducting the research study [18, 19]. For randomized trials, the variable cost of research can also include the loss of benefit from randomizing a portion of the patients to an inferior arm, the time it takes to conduct the trial, and imperfect implementation of the intervention into practice [29, 30].

A fifth measure in a VOI analysis is the optimal sample size (n*), which is the sample size that maximizes the ENBS of pEVSI. ENBS can be computed for multiple data collection processes. From a research prioritization perspective, the data collection process with the highest ENBS is optimal. Again, selecting the WTP threshold and making VOI analytical assumptions are not trivial and detailed methodological approaches are extensively described elsewhere [1820].

VOI Parameters

In summary, six main parameters inform a VOI analysis that can be categorized into three groups: decision maker parameters: 1) WTP threshold, 2) discount rate, 3) decision lifetime; epidemiological parameters: 4) number of prevalent cases, 5) number of incident cases; study design parameters: 6) costs function of study design separated into relevant components (e.g., fixed costs and variable costs) (Table 1). Parameters 1 and 2 affect the weight placed on the effective population’s health gains. Parameters 3–5 directly affect the size of the effective population or number of individuals who can benefit from the information. Parameter 6 is a cost function.

Table 1.

VOI Inputs, Potential Data Sources, and Ranges for Sensitivity Analysis

VOI Parameter Potential Data Sources Examples of Potential Ranges to Test in Sensitivity Analysis
Decision Maker Parameters
  1. Willingness-to-pay threshold Cost-effectiveness acceptability curve and frontier, or decision makers revealed threshold ranges Based on decision makers’ revealed threshold ranges or identified from the cost-effectiveness acceptability curves and frontier where there is competition regarding optimal strategy.
  2. Discount rate Expert opinion or literature (gray or published) 3% based on recommendation in Cost-Effectiveness in Health and Medicine 2nd Ed [32].
  3. Decision lifetime Expert opinion or patent data High/low estimates from experts or external sources such as patent data.
Epidemiological Parameters
  4. Prevalent cases Epidemiological data High/low ranges from data or from the literature.
  5. Incident cases Epidemiological data High/low ranges from data or from the literature.
Study Design Parametersa
  6. Cost functionb Expert opinion or literature (gray or published) High/low estimates from experts, the literature, or data.
a

Proposed study designs can be hypothesized from a review of the expected value of partial perfect information for individual or groups of parameters.

b

costs function of study design should be separated into relevant components (e.g., fixed costs and variable costs).

VOI value of information

Decision makers often have a range of values over which they evaluate VOI parameters (e.g., range of WTP thresholds). In addition, there is often heterogeneity between decision makers and contexts that impact the values selected for VOI parameters. Finally, for some VOI parameters there is uncertainty in the value of the parameter (e.g., number of prevalent cases).

Curve of Optimal Sample Size (COSS)

To determine variability in n* over a range of WTP values and VOI parameters, results can be visualized. The curve of optimal sample size (COSS) graphically presents the change in n* (y-axis) for each data collection process over a range of WTP thresholds (x-axis). Results are graphed over a range of WTP values and can also be presented for variation in select VOI parameters (e.g., cost of research), multiple select VOI parameters, or as best-case and worst-case scenarios where all VOI parameters are set to their extreme values. By graphing n* over a range of WTP values and VOI parameters, decision makers can identify if further data collection is warranted given their operating characteristics (Table 2 for COSS algorithm).

Table 2.

Curve of Optimal Sample Size (COSS) Algorithm

Step 1: Generate a probabilistic sensitivity analysis dataset from an economic evaluation.
Step 2: For each willingness-to-pay threshold, calculate the expected value of sample information for a set of parameters or a proposed study design.
Step 3: For each willingness-to-pay threshold, calculate the expected net benefit of sampling as the difference between the expected value of sample information and cost of research for a study of size n.
Step 4: Compute the optimal sample size (n*) that maximizes the expected net benefit of sampling.
Step 5: Plot n* across willingness-to-pay thresholds.

Applied COSS Example

Below we provide a simplified VOI example and corresponding R code (eStatistical Code; eTable 1 data dictionary) to generate the COSS. The data and R code are based on our prior cost-effectiveness analysis and VOI evaluation of nine urate-lowering treatment strategies for the management of gout [8]. From these analyses, we obtained the data from the PSA (eData 1). The emphasis of this example is on the application of the COSS; as such, the original study and data have been simplified for pedagogical purposes. Three strategies (allopurinol only dose-escalation, allopurinol-febuxostat sequential therapy dose-escalation, and febuxostat-allopurinol sequential therapy dose-escalation) were optimal across a range of WTP thresholds. We report the VOI analyses based on these three optimal strategies and select model inputs related to treatment effectiveness of allopurinol dose-escalation and health utilities.

We first identified values for the VOI parameters for the WTP threshold (input #1), discount rate (input #2), decision lifetime (input #3), number of prevalent gout cases that could benefit from the decision (input #4), and number of future incident gout cases that could benefit from the decision (input #5). We adopted a WTP range of $US0-$US150,000 per QALY based on recommended decision thresholds (eTable 2 assumptions of VOI analysis). Without an a-priori decision threshold, we could also identify the range of WTP values for use in the VOI by identifying regions on the cost-effectiveness acceptability curve and frontier derived from the PSA where there is meaningful uncertainty regarding the optimal strategy (eFigure 1). We chose a discount rate of 3% which corresponds with guidelines for discounting in cost-effectiveness analyses and a decision lifetime of 5 years, which corresponds with the remaining lifetime of the patent for febuxostat [31, 32]. The number of prevalent and incident gout patients that could benefit from the decision regarding the cost-effectiveness of urate-lowering treatment strategies were obtained from epidemiological data [3336].

Using the PSA data (eData 1) generated from the cost-effectiveness analysis, we calculated population EVPI (eFigure 2). We then calculated population EVPPI (eFigure 3) separately for model parameters related to the effectiveness of allopurinol dose-escalation and health utilities. Population EVPPI indicates there is potential value in reducing uncertainty on the parameters informing the effectiveness of allopurinol dose-escalation and health utilities.

To inform the cost-effectiveness of model inputs related to treatment effectiveness and health utilities, we hypothesized two distinct data collection processes (i.e., study designs) could be conducted. First, a randomized controlled trial in which allopurinol dose-escalation is compared against placebo. Second, an observational study to elicit health utilities of gout patients. Utilities would be obtained that correspond with the model states of controlled on treatment, uncontrolled on treatment, and uncontrolled off treatment. In a more detailed VOI evaluation, additional study designs could be hypothesized.

For pedagogical purposes, we assumed our cost function only had a fixed and variable cost (input #6). The cost of conducting a randomized trial also included the loss of benefit for patients randomized to placebo compared to allopurinol dose-escalation (i.e., incremental net benefit). In a more detailed evaluation we could also include costs that capture the time it takes to conduct the study and implement findings [29, 30]. We obtained data on the cost of research from studies reporting the total and average per participant cost of conducting a randomized trial [37, 38]. We separately estimated the cost of conducting an observational study. To be consistent with our prior analyses, we report all costs in 2013 US dollars.

We used a published method to calculate pEVSI that is based on adopting a Gaussian approximation of the Bayesian updating and a linear regression meta-modeling of the relations between the model inputs and the expected opportunity cost [2].

We graphically reported n* over a range of WTP using the COSS (Figure 1), and we graphically reported n* for a randomized controlled trial comparing allopurinol dose-escalation to placebo over a range of WTP given variation in the cost of research (Figure 2).

Figure 1.

Figure 1

Curve of Optimal Sample Size n* is reported as the total sample size. For a two-arm randomized trial, n*/2 represents the optimal sample size per study arm.

Figure 2.

Figure 2

Curve of Optimal Sample Size (COSS) for Study Design 1 (RCT) with Sensitivity Analysis on Variable Cost of Research n* is reported as the total sample size. For a two-arm randomized trial, n*/2 represents the optimal sample size per study arm.

From the COSS, a decision maker is able to determine the optimal sample size for a study design given their a-priori WTP threshold. For example, given a decision maker’s WTP is $US50,000 per QALY, an observational study collecting data on health utilities should enroll 4,300 subjects (rounded to nearest 25th) and a randomized trial on the effectiveness of allopurinol dose-escalation should enroll 100 subjects or 50 per arm. From a research prioritization perspective, the decision maker must also evaluate the ENBS of the proposed study designs to determine which study generates the largest benefit (eFigure 4). While the optimal sample size of the randomized trial is less than the observational study, the ENBS of the randomized trial is greater indicating it will generate a larger reduction in uncertainty.

DISCUSSION/CONCLUSION

VOI should be used beyond just informing value of eliminating uncertainty and can and should inform future research funding and study design. Recent methodological developments allow researchers to quickly and efficiently calculate VOI measures. Similar to cost-effectiveness analyses, sensitivity analyses in VOI analysis should become the norm. To support the reporting of sensitivity analysis of parameters and assumptions, we propose the COSS as a standard for graphically representing n* over a range of WTP values and VOI parameters.

The COSS graphically displays n* for all study designs over a range of WTP threshold values and VOI parameters as these vary by agency and setting. The COSS can also show variation and uncertainty in multiple or select VOI parameters (e.g., cost of research) on n*. To facilitate research prioritization, ENBS and n* should be evaluated concurrently. As our example illustrated, the study design with the highest optimal sample size is not necessarily the design that yields the largest benefit (eFigure 4).

Since a decision maker’s WTP is given a-priori, the COSS should not be used to determine the cost-effectiveness threshold required to provide an optimal sample size. That is, the COSS should be used to determine the optimal study design and n* given an a-priori WTP threshold.

Graphically presenting n* in the COSS efficiently summarizes a key measure of a VOI analysis and allows decision makers with different WTP thresholds and operating characteristics to identify if further research is in fact needed and the size of such research studies.

Supplementary Material

40273_2019_770_MOESM1_ESM
40273_2019_770_MOESM2_ESM
40273_2019_770_MOESM3_ESM
40273_2019_770_MOESM4_ESM

Key Points for Decision Makers.

  • We introduce the curve of optimal sample size (COSS), which is a graphical representation of the optimal sample size of a study design over a range of willingness-to-pay thresholds.

  • The COSS presents summary data for decision makers to determine the sample size that optimizes research funding given their operating characteristics (e.g., willingness-to-pay threshold).

Acknowledgments

Funding/Support: Financial support for this study was provided in part by a Doctoral Dissertation Fellowship from the Graduate School of the University of Minnesota as part of Dr. Alarid-Escudero's doctoral program. Drs. Kuntz and Alarid-Escudero were supported by a grant from the National Cancer Institute (U01-CA-199335) as part of the Cancer Intervention and Surveillance Modeling Network (CISNET). Dr. Jutkowitz was supported by a grant from the National Institute on Aging (1R21AG059623-01) and a grant from the Brown School of Public Health. The funding agencies had no role in the design of the study, interpretation of results, or writing of the manuscript. The content is solely the responsibility of the authors and does not necessarily represent the official views of the National Institutes of Health. The funding agreement ensured the authors' independence in designing the study, interpreting the data, writing, and publishing the report.

Footnotes

Data Availability Statement: Data and statistical code are provided in the online appendix.

Compliance with Ethical Standards

Conflicts of Interest: EJ reports no conflicts of interest. FAE reports no conflicts of interest. KMK reports no conflicts of interests. HL reports no conflicts of interest.

References

  • 1.Claxton K, Posnett J. An economic approach to clinical trial design and research priority-setting. Health Econ 1996;5(6):513–24. [DOI] [PubMed] [Google Scholar]
  • 2.Jalal H, Alarid-Escudero F. A Gaussian Approximation Approach for Value of Information Analysis. Med Decis Making 2018;38(2):174–88. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3.Menzies NA. An Efficient Estimator for the Expected Value of Sample Information Med Decis Making 2016;36(3):308–20. [DOI] [PubMed] [Google Scholar]
  • 4.Jalal H, Goldhaber-Fiebert JD, Kuntz KM. Computing Expected Value of Partial Sample Information from Probabilistic Sensitivity Analysis Using Linear Regression Metamodeling. Med Decis Making 2015;35(5):584–95. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5.Baio G, Berardi A, Heath A. Bayesian cost-effectiveness analysis with the R package BCEA London: Springer; 2017. [Google Scholar]
  • 6.Heath A, Manolopoulou I, Baio G. Efficient Monte Carlo Estimation of the Expected Value of Sample Information Using Moment Matching. Med Decis Making 2018;38(2):163–73. [DOI] [PubMed] [Google Scholar]
  • 7.Strong M, Oakley JE, Brennan A, Breeze P. Estimating the expected value of sample information using the probabilistic sensitivity analysis sample: a fast, nonparametric regression-based method. Med Decis Making 2015;35(5):570–83. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.Jutkowitz E, Alarid-Escudero F, Choi HK, Kuntz KM, Jalal H. Prioritizing Future Research on Allopurinol and Febuxostat for the Management of Gout: Value of Information Analysis. Pharmacoeconomics 2017;35(10):1073–85. [DOI] [PubMed] [Google Scholar]
  • 9.Tuffaha HW, Gordon LG, Scuffham PA. Value of Information Analysis Informing Adoption and Research Decisions in a Portfolio of Health Care Interventions. MDM Policy Pract 2016;1(1):1–11. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10.Tuffaha HW, Reynolds H, Gordon LG, Rickard CM, Scuffham PA. Value of information analysis optimizing future trial design from a pilot study on catheter securement devices. Clin Trials 2014;11(6):648–56. [DOI] [PubMed] [Google Scholar]
  • 11.Kearns B, Chilcott J, Whyte S, Preston L, Sadler S. Cost-effectiveness of screening for ovarian cancer amongst postmenopausal women: a model-based economic evaluation. BMC Med 2016. 6;14(1):200. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12.Rabideau DJ, Pei PP, Walensky RP, Zheng A, Parker RA. Implementing Generalized Additive Models to Estimate the Expected Value of Sample Information in a Microsimulation Model: Results of Three Case Studies. Med Decis Making 2018;38(2):189–99. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.Steuten L, van de Wetering G, Groothuis-Oudshoorn K, Retèl V. A systematic and critical review of the evolving methods and applications of value of information in academia and practice. Pharmacoeconomics 2013;31(1):25–48. [DOI] [PubMed] [Google Scholar]
  • 14.Wilson EC. A Practical Guide to Value of Information Analysis. Pharmacoeconomics 2015;33(2)105–21. [DOI] [PubMed] [Google Scholar]
  • 15.Willan A, Kowgier M. Determining optimal sample sizes for multi-stage randomized clinical trials using value of information methods. Clin Trials 2008;5(4):289–300. [DOI] [PubMed] [Google Scholar]
  • 16.Willan AR. Optimal sample size determinations from an industry perspective based on the expected value of information. Clin Trials 2008;5(6):587–94. [DOI] [PubMed] [Google Scholar]
  • 17.Willan AR, Pinto EM. The value of information and optimal clinical trial design. Stat Med 2005;24(12):1791–806. [DOI] [PubMed] [Google Scholar]
  • 18.Heath A, Manolopoulou I, Baio G. A Review of Methods for Analysis of the Expected Value of Information. Med Decis Making 2017; 37(7):747–758. [DOI] [PubMed] [Google Scholar]
  • 19.Briggs A, Sculpher M, Claxton K. Decision modelling for health economic evaluation Oxford: Oxford University Press; 2006. [Google Scholar]
  • 20.Eckermann S, Willan AR. Expected value of information and decision making in HTA. Health Econ 2007;16(2):195–209. [DOI] [PubMed] [Google Scholar]
  • 21.Tuffaha HW, Gordon LG, Scuffham PA. Value of information analysis in healthcare: a review of principles and applications. J Med Econ 2014;17(6):377–83. [DOI] [PubMed] [Google Scholar]
  • 22.Tuffaha HW, Gordon LG, Scuffham PA. Value of information analysis in oncology: the value of evidence and evidence of value. J Oncol Pract 2014;10(2):e55–62. [DOI] [PubMed] [Google Scholar]
  • 23.Strong M, Oakley JE, Brennan A. Estimating multiparameter partial expected value of perfect information from a probabilistic sensitivity analysis sample: a nonparametric regression approach. Med Decis Making 2014;34(3):311–26. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 24.Madan J, Ades AE, Price M, Maitland K, Jemutai J, Revill P, Welton NJ. Strategies for efficient computation of the expected value of partial perfect information. Med Decis Making 2014;34(3):327–42. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25.Philips Z, Claxton K, Palmer S. The half-life of truth: what are appropriate time horizons for research decisions? Med Decis Making 2008;28(3):287–99. [DOI] [PubMed] [Google Scholar]
  • 26.Ades AE, Lu G, Claxton K. Expected value of sample information calculations in medical decision modeling. Med Decis Making 2004;24(2):207–27. [DOI] [PubMed] [Google Scholar]
  • 27.Meltzer DO, Hoomans T, Chung JW, Basu A. Minimal Modeling Approaches to Value of Information Analysis for Health Research. Med Decis Making 2011;31(6):E1–22. [DOI] [PubMed] [Google Scholar]
  • 28.Jutkowitz E, Choi HK, Pizzi LT, Kuntz KM. Cost-Effectiveness of Allopurinol and Febuxostat for the Management of Gout. Ann Intern Med 2014;161(9):617–26. [DOI] [PubMed] [Google Scholar]
  • 29.Eckermann S, Willan AR. Time and expected value of sample information wait for no patient. Value Health 2008;11(3):522–6. [DOI] [PubMed] [Google Scholar]
  • 30.Willan AR, Eckermann S. Optimal clinical trial design using value of information methods with imperfect implementation. Health Econ 2010;19(5):549–61. [DOI] [PubMed] [Google Scholar]
  • 31.U.S. Department of Health and Human Services; U.S. Food & Drug Administration. Orange Book: Approved Drug Products with Therapeutic Equivalence Evaluations 2016. https://www.accessdata.fda.gov/scripts/cder/ob/patent_info.cfm?Product_No=001&Appl_No=021856&Appl_type=N. Accessed 15 March 2018.
  • 32.Neumann PJ, Sanders GD, Russell LB, Siegel JE, Ganiats TG. Cost-effectiveness in health and medicine 2nd ed. New York: Oxford University Press; 2017. [Google Scholar]
  • 33.Zhu Y, Pandya BJ, Choi HK. Prevalence of gout and hyperuricemia in the US general population: The National Health and Nutrition Examination Survey 2007–2008. Arthritis Rheum 2011;63(10):3136–41. [DOI] [PubMed] [Google Scholar]
  • 34.Wallace KL, Riedel AA, Joseph-Ridge N, Wortmann R. Increasing prevalence of gout and hyperuricemia over 10 years among older adults in a managed care population. J Rheumatol 2004;31(8):1582–7. [PubMed] [Google Scholar]
  • 35.Arromdee E, Michet CJ, Crowson CS, O’Fallon WM, Gabriel SE. Epidemiology of gout: is the incidence rising? J Rheumatol 2002;29(11):2403–6. [PubMed] [Google Scholar]
  • 36.Mikuls TR, Farrar JT, Bilker WB, Fernandes S, Schumacher HR Jr, Saag KG. Gout epidemiology: results from the UK General Practice Research Database, 1990–1999 Ann Rheum Dis 2005;64(2):267–72. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 37.Johnston SC, Rootenberg JD, Katrak S, Smith WS, Elkins JS. Effect of a US National Institutes of Health programme of clinical trials on public health and costs. Lancet 2006;367(9519):1319–27. [DOI] [PubMed] [Google Scholar]
  • 38.Emanuel EJ, Schnipper LE, Kamin DY, Levinson J, Lichter AS. The costs of conducting clinical research. J Clin Oncol 2003;21(22):4145–50. [DOI] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

40273_2019_770_MOESM1_ESM
40273_2019_770_MOESM2_ESM
40273_2019_770_MOESM3_ESM
40273_2019_770_MOESM4_ESM

RESOURCES