Skip to main content
Implementation Science : IS logoLink to Implementation Science : IS
. 2020 Dec 14;15:109. doi: 10.1186/s13012-020-01069-w

Cost-effectiveness of the Adaptive Implementation of Effective Programs Trial (ADEPT): approaches to adopting implementation strategies

Andria B Eisman 1,, David W Hutton 2, Lisa A Prosser 2,3, Shawna N Smith 2,4, Amy M Kilbourne 5,6
PMCID: PMC7734829  PMID: 33317593

Abstract

Background

Theory-based methods to support the uptake of evidence-based practices (EBPs) are critical to improving mental health outcomes. Implementation strategy costs can be substantial, and few have been rigorously evaluated. The purpose of this study is to conduct a cost-effectiveness analysis to identify the most cost-effective approach to deploying implementation strategies to enhance the uptake of Life Goals, a mental health EBP.

Methods

We used data from a previously conducted randomized trial to compare the cost-effectiveness of Replicating Effective Programs (REP) combined with external and/or internal facilitation among sites non-responsive to REP. REP is a low-level strategy that includes EBP packaging, training, and technical assistance. External facilitation (EF) involves external expert support, and internal facilitation (IF) augments EF with protected time for internal staff to support EBP implementation. We developed a decision tree to assess 1-year costs and outcomes for four implementation strategies: (1) REP only, (2) REP+EF, (3) REP+EF add IF if needed, (4) REP+EF/IF. The analysis used a 1-year time horizon and assumed a health payer perspective. Our outcome was quality-adjusted life years (QALYs). The economic outcome was the incremental cost-effectiveness ratio (ICER). We conducted deterministic and probabilistic sensitivity analysis (PSA).

Results

Our results indicate that REP+EF add IF is the most cost-effective option with an ICER of $593/QALY. The REP+EF/IF and REP+EF only conditions are dominated (i.e., more expensive and less effective than comparators). One-way sensitivity analyses indicate that results are sensitive to utilities for REP+EF and REP+EF add IF. The PSA results indicate that REP+EF, add IF is the optimal strategy in 30% of iterations at the threshold of $100,000/QALY.

Conclusions

Our results suggest that the most cost-effective implementation support begins with a less intensive, less costly strategy initially and increases as needed to enhance EBP uptake. Using this approach, implementation support resources can be judiciously allocated to those clinics that would most benefit. Our results were not robust to changes in the utility measure. Research is needed that incorporates robust and relevant utilities in implementation studies to determine the most cost-effective strategies. This study advances economic evaluation of implementation by assessing costs and utilities across multiple implementation strategy combinations.

Trial registration

ClinicalTrials.gov Identifier: NCT02151331, 05/30/2014.

Keywords: Implementation science; Costs and cost analysis; Mental health services, community; Cost-effectiveness analysis


Contributions to the literature.

  • Researchers to date have focused primarily on quantifying intervention costs; few have focused on implementation strategy costs and cost-effectiveness.

  • This research focuses on advancing approaches for evaluating cost and cost-effectiveness of implementation strategies, which are provider tools/strategies to promote intervention uptake and have been understudied.

  • This study is one of the first to conduct a comparative economic analysis of an adaptive implementation strategy trial, to provide useful, accessible information for communities to make well-informed decisions about resourcing implementation investments.

Background

Evidence-based treatments for mental health conditions, including depression, are essential to improving the public’s health [1]. Mental health conditions frequently co-occur with substance use disorders, and other co-occurring conditions, inciting sequelae of short- and long-term consequences [2]. Mental health conditions have a significant financial toll: researchers estimated in 2008 that the annual earnings loss for serious mental illness in 2008 was $193.2 billion [3]. Collaborative care models (CCMs) have demonstrated effectiveness in improving outcomes among patients with mental disorders; collaborative care models such as Life Goals are designed to improve medical and psychiatric outcomes for persons with mood disorders through personal goal-setting aligned with wellness and symptom coping strategies and supported through collaborative care [46]. Life Goals is an evidence-based CCM that focuses on three components recognized as central to effective CCMs: patient self-management, care management, and provider decision support [7, 8]. Several randomized trials have shown Life Goals to be effective in improving mental and physical health outcomes for patients with unipolar and bipolar depression [46, 9]. The Life Goals self-management component comprises six psychosocial sessions for patients, to be delivered in either individual or group format. While all Life Goals patients complete core introduction and conclusion modules, the four intermediary sessions can be chosen by patients and providers from among several mental health and wellness subjects, including depression, mania, physical activity, or substance abuse. Life Goals also provides manualized support for care management and provider decision support, including templates for tracking patient progress and guides to common medications for unipolar/bipolar depression patients. Most individuals suffering from depression and other mental health conditions are not receiving evidence-based practices (EBPs) such as Life Goals in community settings, resulting in poor and costly health outcomes and millions of research dollars wasted when EBPs fail to reach those most in need [1012]. Researchers increasingly recognize that EBPs must be complemented by effective implementation strategies (i.e., implementation interventions) to achieve desired public health outcomes [13]. Replicating Effective Programs (REP) is an implementation strategy focused on maximizing flexibility and fidelity in EBP delivery [14]. REP, based on the CDC’s research-to-practice framework [15], is guided by Social Learning [16] and Diffusion of Innovations Theories [17]. Standard REP includes three primary components: program packaging, provider training, and facilitation. Standard REP is a low intensity, minimal cost intervention that is akin to standard implementation for many evidence-based programs and practices; standard REP has improved uptake of brief HIV-focused interventions but has been less successful with the uptake of more complex behavioral interventions [18]. Researchers have also developed enhanced REP for more complex clinical behavioral interventions, which include added customization for program packaging and training, and implementation facilitation [19]. Implementation facilitation (i.e., facilitation) is a promising implementation strategy from the integrating Promoting Action on Research Implementation in Health Services (iPARIHS) framework that provides ongoing, individualized assistance for program delivery that can help enhance uptake of EBPs such as Life Goals in community clinics [19, 20]. Facilitation applies principles of interactive problem solving with practice-based knowledge to support providers as they engage in program delivery [21, 22]. Individuals within (internal facilitator, IF) and outside of (external facilitator, EF) the organization can provide ongoing support for EBP implementation [19]. External facilitators (EF) provide expertise, active guidance, and support for intervention delivery. Internal facilitators (IF) work in tandem with EFs to support providers in program delivery and communicate with organizational leadership and the external facilitator.

The costs associated with implementation strategies, especially multicomponent strategies such as REP+facilitation, can be substantial. Cost is a key consideration from an organizational or system perspective when implementing new innovations [11]. Understanding the resources needed to achieve desired behavioral outcomes (e.g., improved mental health) is essential to implementing and sustaining EBPs in communities [23]. Most economic evaluation of implementation, however, has focused on intervention costs and not the costs of implementation strategies required to deploy and sustain them [24]. Economic evaluation of implementation refers to the systematic evaluation of what outcomes a specific implementation strategy or set of competing strategies achieves and the costs of achieving them [25]. Economic evaluation provides key information for decision makers regarding implementation strategies to support and sustain EBP delivery. Organizations benefit from evidence that supports (or refutes) investment in specific strategies as an efficient use of resources, and this can help prioritize implementation efforts [11, 24, 26]. Despite this need for practical economic information that will provide decision makers with information on whether the cost of deploying an implementation strategy is worth the added cost (versus standard implementation or an alternative strategy), less than 10% of implementation studies include cost information, and even fewer conduct comparative economic analyses [25, 27]. Thus, additional research is needed to advance economic evaluation of implementation as this will be instrumental in demonstrating if investment in implementation strategies is worth the additional costs [28].

Many types of cost evaluation exist, but one well suited to implementation science is cost-effectiveness analysis. Cost-effectiveness analysis (CEA) assesses whether incremental benefits of one strategy versus another are sufficient to justify additional costs and has been used to support mental health treatment-focused EBPs for clinical settings [29]. CEA can inform decisions about resource allocation for program selection and delivery [30].

The objective of this study is to estimate the costs and conduct a CEA as part of an adaptive implementation trial comparing different implementation strategies. The goal of Adaptive Implementation of Effective Programs Trial (ADEPT) is to use a sequential multiple assignment randomized trial (SMART) design to compare the effectiveness of different augmentations to REP using EF or a combination of EF + IF on mental health outcomes among patients diagnosed with depression or bipolar disorders in community-based practices; details of the ADEPT trial are described in more detail elsewhere [19]. A secondary ADEPT aim was to assess the costs for different scenarios of combining REP+facilitation (see Fig. 1 and Fig. 4 in the Appendix) to identify the most cost-effective implementation strategy approach. We compare four different implementation strategy combinations and evaluate relative cost-effectiveness to identify which implementation strategies are most cost-effective in achievi program goals: Strategy 0: REP only, Strategy 1: REP+EF, Strategy 2: REP+EF add IF if needed, and Strategy 3: REP+EF/IF. Clinics responding to their respective implementation strategy (e.g., > 10 patients receiving Life Goals) discontinued the implementation strategy. Among those that did not respond during the second phase of the trial, for the final phase Strategy 1 continued with EF, Strategy 2 added IF, and Strategy 3 continued with EF/IF.

Fig. 1.

Fig. 1

Decision tree of the ADEPT trial. aSites that responded to the implementation strategy after the initial 6 months of the Trial Phase: either < 10 patients receiving Life Goals or > 50% of patients receiving Life Goals had ≤ 3 sessions, min dose for clinically significant results. Sites that responded to the implementation strategy discontinued the strategy during the second 6 months/Phase III of the trial

Methods

This study will use a simulation modeling approach using data from a previously conducted clinical trial [19, 31]. Our results are reported using the Consolidated Health Economic Evaluation Reporting (CHEERS) guidelines [32]. Implementation strategies included in the model reflect implementation strategies that could be developed using data from the trial. In this study, we focus on the ADEPT community-based mental health or primary care clinics who were non-responsive after 6 months of Replicating Effective Programs (REP) and would receive additional implementation support (i.e., facilitation) to enhance uptake of Life Goals. Non-responsive to REP was defined as 10 or fewer patients receiving Life Goals or < 50% of patients receiving a clinically significant dose of Life Goals, fewer than three Life Goals sessions (< 3 out of 6), after 6 months [3335]. Eligible sites had at least 100 unique patients diagnosed with depression and could designate at least 1 mental health provider to administer individual or group collaborative care sessions for patients. The study was approved by local institutional review boards (IRBs) and registered under clinicaltrials.gov (identifier: NCT02151331).

Modeling approach

Using data from the ADEPT trial, we designed a cost-effectiveness study to evaluate three strategies that could be implemented to support the uptake and clinical effectiveness of Life Goals. These strategies do not exactly match the arms in the clinical trial because our goal was to evaluate the optimal implementation strategy approach among non-responders. We developed a decision tree to assess 1-year costs and outcomes for different intervention strategies following 6 months of REP (baseline) among non-responsive sites (i.e., slow adopter sites). Implementation strategies included in the model (see Fig. 1) were as follows: Strategy 0: REP only, Strategy 1: REP+EF, Strategy 2: REP+EF, ADD IF if needed, and Strategy 3: REP+EF/IF. The probability of non-response to the implementation strategies in the model was based on observed response rates in the study, which remained consistent across each phase at approximately .09 (that is, 9% were responders). Sites who responded to their assigned implementation strategy after 6 months of the trial (Phase II) discontinued the strategy. Sites who did not respond at after 6 months proceeded with Phase III as follows: for Strategy 1: continued REP+EF, for Strategy 2, added IF and for Strategy 3 continued with REP+EF/IF. The analysis uses a 1-year time horizon and assumes a health sector perspective. Parameter inputs were derived using primary data from ADEPT.

Costs

Implementation strategy costs for baseline REP were the same for all participants and include costs for training program providers, training compensation (e.g., pay during non-work hours), time costs for assessing organizational needs, and pre-implementation meetings. Non-labor costs included costs of the curriculum (manual and materials) and travel costs [24, 36]. Facilitation costs were based on the facilitation logs. The study EF and site IFs logged their tasks, categorizing mode, personnel interaction, duration, and the primary focus of each task. These tasks included coaching, developing an implementation plan, education, linking to outside resources, and consultation. We calculated costs based on time spent by hourly wage plus fringe rates for facilitators. As there was one EF employed by the study team, we used the EF hourly wage + fringe. For the IFs, training, and background (and thus costs) varied. We based the IF salary and fringe rates on current rates for Licensed Masters of Social Work (LMSW) professional using Bureau of Labor Statistics data, as many of the IFs were LMSWs. As we anticipated differences in uptake, that is the number of patients receiving Life Goals by condition, we calculated the total site-level cost per strategy (the level of randomization) and divided by the number of patients in that implementation strategy condition. The number of patients per condition was obtained from site-level records. Costs were collected in 2014 and adjusted to US 2018 dollars using the Consumer Price Index [37]. A summary of cost parameters is provided in Table 1. We report summary statistics for implementation costs with 95% confidence intervals. We estimated the costs of REP using the available cost data to obtain a comprehensive assessment of total implementation intervention costs, plus the costs of facilitation activities in each condition (EF and EF/IF).

Table 1.

Model inputs

Parameter Base Low High Distributiona Source
Costs
 Cost of REP (Phase I)b 588.95 0 558.95 Normal Time and resource tracking, study staff
 Additional cost of EF (Phase II) 32.70 32.39 33.01 Normal Time logs
 Additional cost of EF (Phase III) 17.55 1.22 30.84 Normal Time logs
 Additional cost of EF and IF (Phase II) 31.23 28.51 30.49 Normal Time logs
 Additional cost of EF and IF (Phase III) 6.35 3.15 9.27 Normal Time logs
Probabilities
 Probability of response after Phase II with REP+EF 0.095 0 0.095 Site response data
 Probability of response after Phase II with REP+EF/IF 0.091 0 0.091 Site response data
Utilitiesc,d
 REP only .475 0.43 0.521 Beta Patient survey
 EF (Phase II) .497 0.42 0.573 Beta Patient survey
 EF non-responding site (Phase III) .446 0.306 0.586 Beta Patient survey
 EF responding site (Phase III) .721 0.533 0.909 Beta Patient survey
 EF and IF (Phase II) .463 0.362 0.564 Beta Patient survey
 EF add IF (Phase III) .568 0.392 0.566 Beta Patient survey
 EF and IF non-responding site (Phase III) .479 0.392 0.566 Beta Patient survey
 EF and IF responding site (Phase III) .372 0.184 0.559 Beta Patient survey

aDistributions are parameterized such that the mean is the base value and the standard deviation is ¼ of the difference between the low and high values

bPhases refer to values calculated within phases of the original trial: Phase I: baseline/initial 6-month period prior to the trial phase of the study with REP only, Phase II: second 6 months of the study, Phase III: final 6 months of the trial

cInverse probability weighting to account for missing data with weights estimated from a logistic regression model for predicting non-response

dEQ-5D calculated using mapping algorithm from components of the SF-12

Health outcomes

Quality-adjusted life years (QALYs)

To develop a preference-based health utility measure for the current study, we mapped the SF-12 (which was collected as part of the patient-level evaluation in the ADEPT trial) to the EQ-5D, a multi-attribute utility instrument, using an established algorithm developed by Franks and colleagues [38]. The EQ-5D yields interval-level scores ranging from 0 (dead) to 1 (perfect health). This mapping provides a health utility measure for each health state experienced by patients in the study and can be used to calculate quality-adjusted life years, the preferred measure for health benefits used in cost-effectiveness analysis.

Data analytic approach

We used a decision-tree model to compare the cost-effectiveness across different scenarios for combining REP+facilitation for the Life Goals EBP (see Fig. 1 and Fig. 4 in the Appendix). The time horizon for this analysis was 12 months as this is the duration of the trial phase of the study. In this analysis, we adopted a health system/payer perspective. This narrower perspective stands in contrast to the full, societal perspective, which incorporates all relevant costs and benefits and is recommended for most economic evaluations [39]. While this narrower perspective can potentially ignore important costs or benefits from the broad societal standpoint, it has the practical value of explicitly addressing the budgetary concerns of payers. Thus, this approach fits well with implementation science contexts where financial factors are often central to whether programs and services are adopted and sustained [40].

Assumptions were made on the psychometric properties of the outcome measures, the effectiveness of the Life Goals intervention, and the reliability of time reporting by the facilitators. We test these assumptions in the sensitivity analyses by varying the costs and outcomes related to each intervention condition at low and high values (95% confidence interval). To address missing data on our utility (outcome) measures, we employed an inverse probability weighting (IPW) approach [41].

We estimated per-patient costs and QALYs for each implementation strategy sequence. We calculated the per-patient cost by dividing the total costs per condition by the number of patients in each condition. To compare interventions, we divided net incremental costs (net increase in costs from REP+EF/IF versus REP+EF, for example) by incremental effectiveness (net increase in QALYs in REP+EF/IF versus REP+EF groups, for example) to calculate the incremental cost-effectiveness ratio for patient-level outcomes across the conditions. We conducted a one-way sensitivity analysis on all input parameters listed in Table 1 to create a tornado diagram using net monetary benefits (NMB). We used NMB as this facilitates multiple comparisons, as in the current study, and incremental cost-effectiveness ratios (ICERs) are less suitable with more than 2 comparators [42]. The sensitivity analysis evaluated how costs and incremental cost-effectiveness are affected by variations in key parameters [30]. When available, we based upper/lower bound estimates on the 95% confidence intervals. We also conducted a probabilistic sensitivity analysis (PSA). PSA characterizes uncertainty in all parameters simultaneously, reflecting the likelihood that each model parameter takes on a specific value and provides information on overall decision uncertainty based on parameter uncertainty [43]. We conducted 1000 model simulations to quantify the probability that the implementation strategy is cost-effective for a range of thresholds of willingness-to-pay [44]. We conducted a scenario analysis to evaluate results for longer analytic time horizons, from 2 to 10 years. In this additional analysis, the effects of the intervention were assumed to remain constant over time, consistent with values estimated during the final phase for each condition of the trial.

Results

Results of base case analysis

Base case results are presented in Table 2, and a plot of cost-effectiveness across implementation strategies is depicted in Fig. 2. Our base case analysis results indicate REP ONLY is the least expensive. REP+EF, ADD IF has an ICER of $593/QALY. REP+EF had higher QALYs than REP alone, but the QALYs were not as high as REP+EF, ADD IF, and it was higher cost than REP+EF, ADD IF. REP+EF/IF had higher costs and lower QALYs than REP ONLY.

Table 2.

Base case analysis results

Condition Cost per patient Effectiveness ICERa NMBb
Utility: QALYS
 Baseline (REP only) 588.95 0.47 0 46911.05
 REP+EF only 637.53 0.48 Dominatedc 47822.00
 REP+EF, add IF if needed 627.40 0.54 593.42d 53351.18
 REP+EF/IF 625.95 0.47 Dominatedc 45987.68

aICER: incremental cost-effectiveness ratio

bNMB: net monetary benefit. Willingness-to-pay threshold $100,000/QALY [45]

cDominated: more expensive and less effective than comparators [46]

dSince “REP+EF/IF” is dominated, “REP+EF, add IF if needed” is compared to REP only

Fig. 2.

Fig. 2

Cost-effectiveness plane, organization/payer perspective

Sensitivity analysis

To test our model assumptions, we conducted sensitivity analyses on all parameters (Appendix Table 3). In the tornado diagram (see Fig. 3), we found the results were most sensitive to the following variables (in order): utility of individuals in the REP+EF, ADD IF arm at Phase III, the utility of individuals in the REP+EF arm at Phase II, the utility of individuals in the REP+EF arm at Phase III for responders, and utility of individuals in the REP+EF only arm at Phase III. We then conducted threshold analyses for each of the most sensitive parameters. We found that at utility values below .44 for REP+EF, ADD IF at Phase III, the value of REP+EF, ADD IF is no longer cost-effective and REP+EF becomes the most cost-effective choice. We also found that at utility values above .57 for REP+EF at Phase III, REP+EF ADD IF is no longer the most cost-effective option and REP+EF becomes the most cost-effective choice.

Fig. 3.

Fig. 3

Tornado diagram showing one-way sensitivity analyses for the base case with the most sensitive parameters. All parameters were evaluated and data are provided in the appendix. Thick vertical black lines on the ends of the bars indicate values at which the initial preferred option is no longer cost-effective

In addition to the deterministic sensitivity analyses, we also conducted probabilistic sensitivity analysis. The results indicate that the intervention with the best results in terms of utility would be most preferred. The willingness-to-pay threshold was not important (unless using a $0 willingness-to-pay threshold). REP+EF, ADD IF is the optimal strategy in about 30% of iterations, REP ONLY is the optimal strategy in 31% of the iterations, and REP+EF/IF the optimal strategy in 22% of the iterations.

We also conducted sensitivity analyses to explore an extended time horizon. In this analysis, we investigated the effects of extending the utilities during the final 6-month period in each condition from the current 12-month time horizon to 10 years. We found that if there are no additional benefits beyond 1 year, our cost-effectiveness ratio of REP+EF, ADD IF is $593.42/QALY. If benefits continued to 2 years, the ICER was $223.06/QALY; at 3 years, the ICER was $137.34/QALY; and at 10 years, patients gain 1.14 QALYs and the cost-effectiveness ratio is $33.71/QALY. Full results are provided in Appendix Table 4. REP+EF, ADD IF remained the most cost-effective option with an extended time horizon.

Discussion

Effective implementation of EBPs for mental health treatment in communities is critical to improving the public’s health. Most individuals suffering from depression and other mental health conditions are not receiving evidence-based practices (EBPs) such as Life Goals (LG) in community settings, resulting in poor and costly health outcomes and millions of research dollars wasted when EBPs fail to reach those most in need [1012]. Implementation strategies are key to improving uptake of EBPs in communities and achieving public health objectives of evidence-based treatments such as Life Goals. Implementation strategies, however, vary in intensity and cost. More research is needed on applying these strategies with consideration of the economic impact, given that community clinics often have limited—and carefully allocated—resources to promote EBP uptake [47]. This research is vital to bridging the research-to-practice gap, but economic evaluation of implementation strategies remains understudied [47]. This study is one of the first to investigate the cost-effectiveness of implementation strategies as part of an adaptive trial. Adaptive trials are an effective way to accelerate research-to-practice translation by simultaneously evaluating multiple strategies and combinations of strategies, based on clinics’ needs.

We found that, overall, REP+facilitation in its various permutations is a relatively low-cost implementation strategy. Identifying the costs and potential utilities for each REP+facilitation combination can help decision makers with resource allocation for implementation. Understanding the resources needed to achieve desired behavioral outcomes (e.g., reduced ATOD use) is essential to implementing and sustaining EBIs [23]. Also, we found that REP+EF, ADD IF may be the most cost-effective implementation strategy. But these results are still uncertain based on highly variable quality-of-life assessments by participants. Although researchers have debated if a step-up versus step-down approach to evidence-based clinical treatment is most effective, the optimal approach for implementation strategies to enhance the uptake of these treatments is also unclear. Our results are consistent with other clinical research that suggests a step-up strategy is a more cost-effective approach [48]. This information will support organizations in making informed decisions by providing evidence that supports (or refutes) investment in specific implementation strategies as an efficient use of resources and thus can help prioritize implementation efforts.

We also found that stepping up non-responsive sites immediately to REP+EF/IF, the most intensive and costly strategy (at the site level), was not cost-effective. This may be for several reasons. First, EF alone may be sufficient for community clinics to effectively implement the Life Goals intervention and, thus, in most cases IF may not be necessary [31]. Second, many sites had difficulty identifying an internal facilitator. Subsequent analyses into time data indicate that the mean time to identify an IF was 69 days. This suggests that many sites assigned to the IF condition did not have one for the first 2 months of the evaluation period. These results also indicate that community clinics may have a limited capacity to identify and effectively utilize an IF. Finally, we may have had more favorable results with the REP+EF, ADD IF condition during Phase II as the EF was able to work with the clinic on their barriers to uptake immediately and may have been working with several versus a single staff member.

Our results were highly dependent on the assessment of utility. The utilities were variable and uncertain across the different intervention arms. This had a strong influence on the overall assessment of cost-effectiveness. Further research is needed that incorporates robust and relevant utilities in implementation research to identify the most cost-effective strategies. Although the trial only evaluated patients up until 1 year, our results did not change if we simulated a longer time horizon of benefits. Extending benefits out from the current trial (12 months) to 10 years, the cost-effectiveness of REP+EF, ADD IF improved to $33.71/QALY. As the clinical benefit from engaging in evidence-based practices for mental health treatment may extend beyond the time horizon of the trial itself, studies that only observe outcomes over a short time horizon may report artificially high CE ratios [49]. We have found that extending the time horizon does reduce the CE ratio.

Limitations

We adopted a health payer perspective, which may not account for other relevant costs if considering the societal perspective. This may include indirect costs such as patient time, costs of hospitalization, or other treatments or lost productivity. Yet, this narrower perspective has the practical value of explicitly addressing the budgetary concerns of payers and fits well with implementation science contexts where financial factors are often central to whether programs and services are adopted and sustained [40]. We did not have additional information on estimates of REP costs to vary parameters and these cost estimates primarily relied on research team recall. There may be additional costs not included in the estimates that may have implications on the CEA results. Also, additional information to vary specific parameters may also help inform those parameters that are most influential on our estimates. In our CEA analyses, however, all groups had REP costs incorporated into total costs, so this is unlikely to influence the CEA results across the REP+facilitation permutations. We did not have a direct measure of QALYs and thus our utility estimates may be especially susceptible to measurement error. A notable amount of research has been done, however, on mapping the SF-12 components to the EQ-5D thus reducing the likelihood of error as a result of the mapping process. Next, we had a notable amount of missing patient-level survey data, increasing the likelihood of biased estimates. We did, however, attempt to reduce this bias using inverse probability weighting.

The current study would benefit from a mixed methods approach, specifically a sequential design, to obtain qualitative data following the quantitative data collection to better understand challenges to utilizing IF/EF. Finally, our study has a limited time horizon. Using incremental QALYs gained based on the survey results and running sensitivity analysis to evaluate potential effects of a longer time horizon showed REP+EF, ADD IF was still highly cost-effective. However, a longer time horizon within the RCT would provide additional information for a longer-term return-on-investment and could provide more confidence about which adaptive implementation strategy is best.

Conclusions

Our study has several practice implications. First, our results support using a step-up strategy for implementation support for sites that are slow to implement as a cost-effective approach to enhancing uptake and clinical outcomes. Second, our results provide information for decision makers and community health clinic leadership on the costs and relative benefits of using various implementation strategies to improve clinical outcomes. Third, our results support the need for further cost-effectiveness research and incorporating robust utility assessments in community health clinics to provide evidence that will support or refute investments in specific strategies. Finally, our results point to the need for mid-program utility evaluation for both effectiveness and cost-effectiveness to make accurate determinations of the most efficient implementation strategy approach.

Abbreviations

EBPs

Evidence-based practices

REP

Replicating Effective Programs

CDC

Centers for Disease Control and Prevention

iPARIHS

Integrating Promoting Action on Research Implementation in Health Services

IF

Internal facilitation/facilitator

EF

External facilitation/facilitator

VA

Veteran’s Administration

CEA

Cost-effectiveness analysis

ADEPT

Adaptive Implementation of Effective Programs Trial

SMART

Sequential multiple assignment randomized trial

LMSW

Licensed Masters of Social Work

QALYs

Quality-adjusted life years

IPW

Inverse probability weighting

NMB

Net monetary benefits

ICER

Incremental cost-effectiveness ratio

PSA

Probabilistic sensitivity analysis

Appendix

Fig. 4.

Fig. 4

The original study design to evaluate effectiveness (a) and decision tree model to evaluate cost-effectiveness (b). This cost-effectiveness analysis focuses on implementation strategies for sites not responding to the REP alone intervention (the “sites not responding to REP alone” portion of the tree in 2a). In the original study, baseline data were gathered prior to initiation of the trial phase (Phase I). In this study, we sought to determine the most cost-effective option for deploying an implementation strategy with multiple components across its all possible permutations (e.g., REP+EF/IF) and comparing this to usual implementation (baseline REP). To accomplish this, we created the decision tree to represent all the decision options and their subsequent steps and estimate their respective costs and consequences to allow for comparison. This modeling approach represents the possible implementation strategy decision options for decision makers, quantifies the uncertainty, and allows for evaluation of alternatives. a In the original trial, non-responding sites were randomized following Phase I to REP+EF or REP+EF/IF. b Following Phase II, non-responding sites in the REP+EF condition were randomized again to either continue REP+EF or add IF (REP+EF/IF). Details of the trial are published elsewhere (see Kilbourne et. al., 2014). c Sites that responded to the implementation strategy after the initial 6 months of the Trial Phase: either < 10 patients receiving Life Goals or > 50% of patients receiving Life Goals had ≤ 3 sessions, min dose for clinically significant results. Sites that responded to the implementation strategy discontinued the strategy during the second 6 months/Phase III of the trial

Table 3.

Tornado text report including results for all model input parameters

Variable Low Variable Base Variable High Impact Low High Spread Spread2 Risk % Cum Risk % Spread
Utility REPEFaddIF 12mo 0.442 0.568 0.694 Increase 47821.99524 59051.17619 11229.18095 126094504.9 0.666061436 0.666061436 11229.18095
Utility REPEFonly 6mo 0.42 0.497 0.573 Increase 49501.17619 57151.17619 7650 58522500 0.30912989 0.975191325 7650
Utility REPEF responder 12mo 0.533 0.721 0.909 Increase 52455.9381 54246.41429 1790.47619 3205804.989 0.016933831 0.992125157 1790.47619
Utility REPEFonly 12mo 0.306 0.446 0.586 Increase 53351.17619 54155.32857 804.152381 646661.0518 0.003415819 0.995540976 804.152381
Probability respond REPEF 0 0.095238095 0.095238095 Increase 52622 53351.17619 729.1761905 531697.9168 0.002808556 0.998349532 729.1761905
Cost baseline rep 0 588.95 558.95 Decrease 53381.17619 53940.12619 558.95 312425.1025 0.001650304 0.999999836 558.95
Cost REPEFIF 12mo 3.15 6.35 9.27 Decrease 53348.53429 53354.07143 5.537142857 30.65995102 1.61953E−07 0.999999998 5.537142857
Cost REPEF 6mo 32.39 32.7 33.01 Decrease 53350.86619 53351.48619 0.62 0.3844 2.03049E 09 1 0.62
Probability respond REPEFIF 0 0.090909091 0.090909091 Increase 53351.17619 53351.17619 0 0 0 1 0
Utility REPEFIF 6mo 0.362 0.463 0.564 Increase 53351.17619 53351.17619 0 0 0 1 0
Cost REPEF 12mo 1.22 17.55 30.84 Increase 53351.17619 53351.17619 0 0 0 1 0
Utility REPEFIF 12mo 0.392 0.479 0.566 Increase 53351.17619 53351.17619 0 0 0 1 0
Utility REPbaseline 0.43 0.475 0.521 Increase 53351.17619 53351.17619 0 0 0 1 0
Utility REPEFIF responder 12mo 0.184 0.372 0.559 Increase 53351.17619 53351.17619 0 0 0 1 0
Cost REPEFIF 6mo 28.51 31.23 32.22 Increase 53351.17619 53351.17619 0 0 0 1 0

SPREAD2 spread value, squared

Table 4.

One-way sensitivity analysis results for extended time horizon

Year Strategy Strategyindex Cost Incr costa Effb Incr effc ICERd NMBe C/Ef Dominance
0 Baseline(REP only) 2 588.95 0 0.475 0 0 46911.05 1239.894737
0 REP+EF/IF 1 625.9527273 37.00272727 0.466136364 − 0.008863636 − 4174.666667 45987.68364 1342.853242 (Dominated)
0 REP+EF, addIF if needed 0 627.3952381 38.4452381 0.539785714 0.064785714 593.4215362 53351.17619 1162.304265
0 REP+EF only 3 637.5285714 10.13333333 0.484595238 − 0.055190476 − 183.6065574 47821.99524 1315.589839 (Dominated)
1 Baseline(REP only) 2 588.95 0 0.95 0 0 94411.05 619.9473684
1 REP+EF/IF 1 625.9527273 37.00272727 0.935409091 − 0.014590909 − 2536.012461 92914.95636 669.175373 (Dominated)
1 REP+EF, addIF if needed 0 627.3952381 38.4452381 1.122357143 0.172357143 223.0556707 111608.319 558.9978574
1 REP+EF only 3 637.5285714 10.13333333 0.956785714 − 0.165571429 − 61.20218579 95041.04286 666.3232549 (Dominated)
2 Baseline(REP only) 2 588.95 0 1.425 0 0 141911.05 413.2982456
2 REP+EF/IF 1 625.9527273 37.00272727 1.404681818 − 0.020318182 − 1821.163311 139842.2291 445.618872 (Dominated)
2 REP+EF, addIF if needed 0 627.3952381 38.4452381 1.704928571 0.279928571 137.3394573 169865.4619 367.9891631
2 REP+EF only 3 637.5285714 10.13333333 1.42897619 − 0.275952381 − 36.72131148 142260.0905 446.1435926 (Dominated)
3 Baseline(REP only) 2 588.95 0 1.9 0 0 189411.05 309.9736842
3 REP+EF/IF 1 625.9527273 37.00272727 1.873954545 − 0.026045455 − 1420.69808 186769.5018 334.0277003 (Dominated)
3 REP+EF, addIF if needed 0 627.3952381 38.4452381 2.2875 0.3875 99.21351767 228122.6048 274.2711423
3 REP+EF only 3 637.5285714 10.13333333 1.901166667 − 0.386333333 − 26.2295082 189479.1381 335.3354457 (Dominated)
4 Baseline(REP only) 2 588.95 0 2.375 0 0 236911.05 247.9789474
4 REP+EF/IF 1 625.9527273 37.00272727 2.343227273 − 0.031772727 − 1164.606581 233696.7745 267.1327423 (Dominated)
4 REP+EF, addIF if needed 0 627.3952381 38.4452381 2.870071429 0.495071429 77.6559419 286379.7476 218.5991721
4 REP+EF only 3 637.5285714 10.13333333 2.373357143 − 0.496714286 − 20.4007286 236698.1857 268.6188943 (Dominated)
5 Baseline(REP only) 2 588.95 0 2.85 0 0 284411.05 206.6491228
5 REP+EF/IF 1 625.9527273 37.00272727 2.8125 − 0.0375 − 986.7393939 280624.0473 222.5609697 (Dominated)
5 REP+EF, addIF if needed 0 627.3952381 38.4452381 3.452642857 0.602642857 63.79439769 344636.8905 181.7144906
5 REP+EF only 3 637.5285714 10.13333333 2.845547619 − 0.607095238 − 16.69150522 283917.2333 224.0442462 (Dominated)
6 Baseline(REP only) 2 588.95 0 3.325 0 0 331911.05 177.1278195
6 REP+EF/IF 1 625.9527273 37.00272727 3.281772727 − 0.043227273 − 856.0042061 327551.32 190.7361598 (Dominated)
6 REP+EF, addIF if needed 0 627.3952381 38.4452381 4.035214286 0.710214286 54.13188508 402894.0333 155.48003
6 REP+EF only 3 637.5285714 10.13333333 3.317738095 − 0.71747619 − 14.12358134 331136.281 192.1575945 (Dominated)
7 Baseline(REP only) 2 588.95 0 3.8 0 0 379411.05 154.9868421
7 REP+EF/IF 1 625.9527273 37.00272727 3.751045455 − 0.048954545 − 755.8588672 374478.5927 166.8742048 (Dominated)
7 REP+EF, addIF if needed 0 627.3952381 38.4452381 4.617785714 0.817785714 47.01138382 461151.1762 135.8649528
7 REP+EF only 3 637.5285714 10.13333333 3.789928571 − 0.827857143 − 12.24043716 378355.3286 168.2165137 (Dominated)
8 Baseline(REP only) 2 588.95 0 4.275 0 0 426911.05 137.7660819
8 REP+EF/IF 1 625.9527273 37.00272727 4.220318182 − 0.054681818 − 676.6916043 421405.8655 148.3188471 (Dominated)
8 REP+EF, addIF if needed 0 627.3952381 38.4452381 5.200357143 0.925357143 41.54637849 519408.319 120.6446444
8 REP+EF only 3 637.5285714 10.13333333 4.262119048 − 0.938238095 − 10.80038573 425574.3762 149.5801887 (Dominated)
9 Baseline(REP only) 2 588.95 0 4.75 0 0 474411.05 123.9894737
9 REP+EF/IF 1 625.9527273 37.00272727 4.689590909 − 0.060409091 − 612.5357412 468333.1382 133.477043 (Dominated)
9 REP+EF, addIF if needed 0 627.3952381 38.4452381 5.782928571 1.032928571 37.21964825 577665.4619 108.4909195
9 REP+EF only 3 637.5285714 10.13333333 4.734309524 − 1.048619048 − 9.66350302 472793.4238 134.6613626 (Dominated)
10 Baseline(REP only) 2 588.95 0 5.225 0 0 521911.05 112.7177033
10 REP+EF/IF 1 625.9527273 37.00272727 5.158863636 − 0.066136364 − 559.4914089 515260.4109 121.3353892 (Dominated)
10 REP+EF, addIF if needed 0 627.3952381 38.4452381 6.3655 1.1405 33.70910837 635922.6048 98.56181574
10 REP+EF only 3 637.5285714 10.13333333 5.2065 − 1.159 − 8.743169399 520012.4714 122.4485876 (Dominated)

aIncremental cost

bEffectiveness

cIncremental effectiveness

dIncremental cost-effectiveness ratio

eNet monetary benefit

fCost-effectiveness

Authors’ contributions

AE conceived of the study, conducted the analyses, and drafted the manuscript. DH and LP aided with design and interpretations of cost-effectiveness models and results provided critical input on the full manuscript. SS oversaw data collection and cleaning, aided with interpretations of models and results, drafted sections of the manuscript, and provided critical input on the full manuscript. AK led the data collection and project conception and provided input on study conception. All authors worked on the interpretation of data and critical review and approval of the final manuscript.

Funding

National Institute of Mental Health (R01 MH 099898) and National Institute on Drug Abuse (5K01DA044279-02).

Availability of data and materials

Deidentified data are available on request.

Ethics approval and consent to participate

The study recruited primary care and community mental health sites across the state of Michigan. Before the first randomization, sites that agreed to participate were asked to provide the names of at least 20 patients suitable for Life Goals. These patients were then contacted by the study survey coordinator to confirm eligibility and obtain patient consent. The study was approved by local institutional review boards and registered under clinicaltrials.gov.

Trial registration: ClinicalTrials.gov Identifier: NCT02151331, registered 05/30/2014, https://clinicaltrials.gov/ct2/show/NCT02151331.

Consent for publication

Not applicable.

Competing interests

None.

Footnotes

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

References

  • 1.Onken L, Carroll K, Shoham V, Cuthbert B, Riddle M. Reenvisioning clinical science: unifying the discipline to improve the public health. Clin Psychol Sci. 2014;2(1):22–34. doi: 10.1177/2167702613497932. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 2.Bauer M, Altshuler L, Evans D, Beresford T, Williford W, Hauger R. Prevalence and distinct correlates of anxiety, substance, and combined comorbidity in a multi-site public sector sample with bipolar disorder. J Affect Disord. 2005;85(3):301–315. doi: 10.1016/j.jad.2004.11.009. [DOI] [PubMed] [Google Scholar]
  • 3.Kessler R, Heeringa S, Lakoma M, Petukhova M, Rupp AE, Schoenbaum M, et al. Individual and societal effects of mental disorders on earnings in the United States: results from the national comorbidity survey replication. Am J Psychiatry. 2008;165(6):703–711. doi: 10.1176/appi.ajp.2008.08010126. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4.Kilbourne AM, Li D, Lai Z, Waxmonsky J, Ketter T. Pilot randomized trial of a cross-diagnosis collaborative care program for patients with mood disorders. Depress Anxiety. 2013;30(2):116–122. doi: 10.1002/da.22003. [DOI] [PubMed] [Google Scholar]
  • 5.Kilbourne A, Goodrich D, Nord K, Van Poppelen C, Kyle J, Bauer M, et al. Long-term clinical outcomes from a randomized controlled trial of two implementation strategies to promote collaborative care attendance in community practices. Adm Policy Ment Health. 2015;42(5):642–653. doi: 10.1007/s10488-014-0598-5. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6.McBride BM, Williford W, Glick H, Kinosian B, Altshuler L, et al. Collaborative care for bipolar disorder: part II. Impact on Clinical Outcome, Function, and Costs. Psychiatr Serv. 2006;57(7):937–945. doi: 10.1176/ps.2006.57.7.937. [DOI] [PubMed] [Google Scholar]
  • 7.Woltmann E, Grogan-Kaylor A, Perron B, Georges H, Kilbourne A, Bauer M. Comparative effectiveness of collaborative chronic care models for mental health conditions across primary, specialty, and behavioral health care settings: systematic review and meta-analysis. Am J Psychiatry. 2012;169(8):790–804. doi: 10.1176/appi.ajp.2012.11111616. [DOI] [PubMed] [Google Scholar]
  • 8.Miller C, Grogan-Kaylor A, Perron B, Kilbourne A, Woltmann E, Bauer M. Collaborative chronic care models for mental health conditions: cumulative meta-analysis and metaregression to guide future research and implementation. Med Care. 2013;51(10):922–930. doi: 10.1097/MLR.0b013e3182a3e4c4. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9.Kilbourne A, Goodrich D, Lai Z, Clogston J, Waxmonsky J, Bauer M. Life goals collaborative care for patients with bipolar disorder and cardiovascular disease risk. Psychiatr Serv. 2012;63(12):1234–1238. doi: 10.1176/appi.ps.201100528. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10.Proctor E, Landsverk J, Aarons G, Chambers D, Glisson C, Mittman B. Implementation research in mental health services: an emerging science with conceptual, methodological, and training challenges. Adm Policy Ment Health Ment Health Serv Res. 2009;36(1):24–34. doi: 10.1007/s10488-008-0197-4. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11.Proctor E, Silmere H, Raghavan R, Hovmand P, Aarons G, Bunger A, et al. Outcomes for implementation research: conceptual distinctions, measurement challenges, and research agenda. Adm Policy Ment Health Ment Health Serv Res. 2011;38(2):65–76. doi: 10.1007/s10488-010-0319-7. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12.Institute of Medicine. Improving the quality of health care for mental and substance-use conditions [Internet]. Improving the quality of health care for mental and substance-use conditions: quality chasm series. Washington, DC: National Academies Press; 2006. Available from: http://www.ncbi.nlm.nih.gov/pubmed/20669433. [PubMed]
  • 13.Kirchner J, Waltz T, Powell B, Smith J, Proctor E. Implementation Strategies. In: Brownson R, Colditz G, Proctor E, editors. Dissemination and implementation research in health: translating science to practice. 2. New York, NY: Oxford University Press; 2018. [Google Scholar]
  • 14.Kilbourne A, Neumann M, Pincus H, Bauer M, Stall R. Implementing evidence-based interventions in health care: application of the replicating effective programs framework. Implement Sci IS. 2007;2:42. doi: 10.1186/1748-5908-2-42. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15.Neumann M, Sogolow E. Replicating effective programs: HIV/AIDS prevention technology transfer. AIDS Educ Prev Off Publ Int Soc AIDS Educ. 2000;12(5 Suppl):35–48. [PubMed] [Google Scholar]
  • 16.Bandura A. Social learning theory. Englewood Cliffs, N.J: Prentice Hall; 1977. [Google Scholar]
  • 17.Rogers E. Diffusion of innovations. New York: Free Press; 2003. [Google Scholar]
  • 18.Tones K, Green J. Health promotion: planning and strategies. London ; Thousand Oaks, Calif: SAGE Publications; 2004. p. xv, 376. [Google Scholar]
  • 19.Kilbourne A, Almirall D, Eisenberg D, Waxmonsky J, Goodrich D, Fortney J, et al. Adaptive Implementation of Effective Programs Trial (ADEPT): Cluster randomized SMART trial comparing a standard versus enhanced implementation strategy to improve outcomes of a mood disorders program. Implement Sci. 2014;9(1):132. doi: 10.1186/s13012-014-0132-x. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 20.Harvey G, Kitson A. PARIHS revisited: from heuristic to integrated framework for the successful implementation of knowledge into practice. Implement Sci. 2016;11(1):1–13. [DOI] [PMC free article] [PubMed]
  • 21.Ritchie M, Dollar K, Miller C, Oliver K, Smith J, Lindsay J, et al. Using implementation facilitation to improve care in the Veterans Health Administration (Version 2) 2017. [Google Scholar]
  • 22.Ritchie M, Dollar K, Kearney L, Kirchner J. Research and services partnerships: responding to needs of clinical operations partners: transferring implementation facilitation knowledge and skills. Psychiatr Serv. 2014;65(2):141–143. doi: 10.1176/appi.ps.201300468. [DOI] [PubMed] [Google Scholar]
  • 23.Saldana L, Chamberlain P, Bradford W, Campbell M, Landsverk J. The cost of implementing new strategies (COINS): a method for mapping implementation resources using the stages of implementation completion. Child Youth Serv Rev. 2014;39:177–182. doi: 10.1016/j.childyouth.2013.10.006. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 24.Raghavan R. The role of economic evaluation in dissemination and implementation research. In: Brownson R, Colditz G, Proctor E, editors. Dissemination and implementation research in health: translating science to practice. Oxford; New York: Oxford University Press; 2012. p. xxiii, 536 p.
  • 25.Raghavan R. The role of economic evaluation in dissemination and implementation research. In: Brownson R, Colditz G, Proctor E, editors. Dissemination and implementation research in health: translating science to practice. Oxford; New York: Oxford University Press; 2018. p. 89–106.
  • 26.Bauer M, Damschroder L, Hagedorn H, Smith J, Kilbourne A. An introduction to implementation science for the non-specialist. TT -. BMC Psychol. 2015;3(32):12. doi: 10.1186/s40359-015-0089-9. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27.Vale L, Thomas R, MacLennan G, Grimshaw J. Systematic review of economic evaluations and cost analyses of guideline implementation strategies. Eur J Health Econ. 2007;8(2):111–121. doi: 10.1007/s10198-007-0043-8. [DOI] [PubMed] [Google Scholar]
  • 28.Powell B, Fernandez M, Williams N, Aarons G, Beidas RS, Lewis C, et al. Enhancing the impact of implementation strategies in healthcare: a research agenda. Front Public Health [Internet]. 2019 [cited 2019 Feb 14];7. Available from: https://www.frontiersin.org/articles/10.3389/fpubh.2019.00003/full. [DOI] [PMC free article] [PubMed]
  • 29.Saldana L. The stages of implementation completion for evidence-based practice: protocol for a mixed methods study. Implement Sci IS. 2014;9(1):43. doi: 10.1186/1748-5908-9-43. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 30.Drummond M, Sculpher M, Torrance G, O’Brien B, Stoddart G. Methods for the economic evaluation of health care programmes. Oxford ; New York: Oxford University Press; 2005. p. 379. [Google Scholar]
  • 31.Smith S, Almirall D, Prenovost K, Liebrecht C, Kyle J, Eisenberg D, et al. Change in patient outcomes after augmenting a low-level implementation strategy in community practices that are slow to adopt a collaborative chronic care model: a cluster randomized implementation trial. Med Care. 2019;57(7):503–511. doi: 10.1097/MLR.0000000000001138. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 32.Husereau D, Drummond M, Petrou S, Carswell C, Moher D, Greenberg D, et al. Consolidated Health Economic Evaluation Reporting Standards (CHEERS) statement. BMJ [Internet]. 2013 Mar 25 [cited 2020 Jul 2];346. Available from: https://www.bmj.com/content/346/bmj.f1049. [DOI] [PubMed]
  • 33.Spitzer R, Kroenke K, Williams J, Löwe B. A brief measure for assessing generalized anxiety disorder: The GAD-7. Arch Intern Med. 2006;166(10):1092–1097. doi: 10.1001/archinte.166.10.1092. [DOI] [PubMed] [Google Scholar]
  • 34.Bauer M, Vojta C, Kinosian B, Altshuler L, Glick H. The Internal State Scale: replication of its discriminating abilities in a multisite, public sector sample. Bipolar Disord. 2000;2(4):340–346. doi: 10.1034/j.1399-5618.2000.020409.x. [DOI] [PubMed] [Google Scholar]
  • 35.Glick H, McBride L, Bauer M. A manic-depressive symptom self-report in optical scanable format. Bipolar Disord. 2003;5(5):366–369. doi: 10.1034/j.1399-5618.2003.00043.x. [DOI] [PubMed] [Google Scholar]
  • 36.Zarkin GA, Dunlap LJ, Homsi G. The substance abuse services cost analysis program (SASCAP): a new method for estimating drug treatment services costs. Eval Program Plann. 2004;27(1):35–43. doi: 10.1016/j.evalprogplan.2003.09.002. [DOI] [Google Scholar]
  • 37.U.S. Bureau of Labor Statistics. Consumer Price Index [Internet]. Consumer Price Index. 2019 [cited 2020 Jan 23]. Available from: https://www.bls.gov/cpi/.
  • 38.Franks P, Lubetkin E, Gold M, Tancredi D, Jia H. Mapping the SF-12 to the EuroQol EQ-5D Index in a National US Sample. Med Decis Making. 2004;24(3):247–254. doi: 10.1177/0272989X04265477. [DOI] [PubMed] [Google Scholar]
  • 39.Sanders G, Neumann P, Basu A, Brock D, Feeny D, Krahn M, et al. Recommendations for conduct, methodological practices, and reporting of cost-effectiveness analyses: second panel on cost-effectiveness in health and medicine. JAMA - J Am Med Assoc. 2016;316(10):1093–1103. doi: 10.1001/jama.2016.12195. [DOI] [PubMed] [Google Scholar]
  • 40.Humphreys K, Wagner T, Gage M. If substance use disorder treatment more than offsets its costs, why don’t more medical centers want to provide it? A budget impact analysis in the Veterans Health Administration. J Subst Abuse Treat. 2011;41(3):243–251. doi: 10.1016/j.jsat.2011.04.006. [DOI] [PubMed] [Google Scholar]
  • 41.Li L, Shen C, Li X, Robins J. On weighting approaches for missing data. Stat Methods Med Res. 2013;22(1):14–30. doi: 10.1177/0962280211403597. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 42.Messori A, Trippoli S. Incremental cost-effectiveness ratio and net monetary benefit: promoting the application of value-based pricing to medical devices—a European perspective. Ther Innov Regul Sci. 2018;52(6):755–756. doi: 10.1177/2168479018769300. [DOI] [PubMed] [Google Scholar]
  • 43.Neumann P, Sanders G, Russell L, Siegel J, Ganiats T. Cost-effectiveness in health and medicine. Second Edition, New to this Edition: Oxford. New York: Oxford University Press; 2016. p. 536. [Google Scholar]
  • 44.Krishnan A, Finkelstein E, Levine E, Foley P, Askew S, Steinberg D, et al. A digital behavioral weight gain prevention intervention in primary care practice: cost and cost-effectiveness analysis. J Med Internet Res. 2019;21(5):e12201. doi: 10.2196/12201. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 45.Neumann P, Cohen J, Weinstein M. Updating cost-effectiveness--the curious resilience of the $50,000-per-QALY threshold. N Engl J Med. 2014;371(9):796–797. doi: 10.1056/NEJMp1405158. [DOI] [PubMed] [Google Scholar]
  • 46.Muennig P, Bounthavong M. Cost-effectiveness analyses in health: a practical approach. 3. San Francisco: Jossey-Bass; 2016. p. xvi, 266. [Google Scholar]
  • 47.Lee R, Gortmaker S. Health Dissemination and Implementation within Schools. In: Brownson R, Colditz G, Proctor E, editors. Dissemination and implementation research in health: translating science to practice. 2. New York, NY: Oxford University Press; 2018. pp. 401–416. [Google Scholar]
  • 48.van Marrewijk C, Mujakovic S, Fransen G, Numans M, de Wit NJ, Muris JWM, et al. Effect and cost-effectiveness of step-up versus step-down treatment with antacids, H2-receptor antagonists, and proton pump inhibitors in patients with new onset dyspepsia (DIAMOND study): a primary-care-based randomised controlled trial. Lancet Lond Engl. 2009;373(9659):215–225. doi: 10.1016/S0140-6736(09)60070-2. [DOI] [PubMed] [Google Scholar]
  • 49.Cohen D, Reynolds M. Interpreting the results of cost-effectiveness studies. J Am Coll Cardiol. 2008;52(25):2119–2126. doi: 10.1016/j.jacc.2008.09.018. [DOI] [PMC free article] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Data Availability Statement

Deidentified data are available on request.


Articles from Implementation Science : IS are provided here courtesy of BMC

RESOURCES