Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2015 May 2.
Published in final edited form as: Psychiatr Serv. 2015 Feb 17;66(5):484–490. doi: 10.1176/appi.ps.201400206

Multilevel Predictors of Clinic Adoption of State-Supported Trainings in Children’s Services

Su-chin Serene Olin 1, Ka Ho Brian Chor 1, James Weaver 1, Naihua Duan 1, Bonnie D Kerker 1, Lisa J Clark 1, Andrew F Cleek 1, Kimberly Eaton Hoagwood 1, Sarah McCue Horwitz 1
PMCID: PMC4417050  NIHMSID: NIHMS679553  PMID: 25686815

Abstract

Objective

Characteristics associated with participation in training in evidence-informed business and clinical practices by 346 outpatient mental health clinics licensed to treat youths in New York State were examined.

Methods

Clinic characteristics extracted from state administrative data were used as proxies for variables that have been linked with adoption of innovation (extraorganizational factors, agency factors, clinic provider-level profiles, and clinic client-level profiles). Multiple logistic regression models were used to assess the independent effects of theoretical variables on the clinics’ participation in state-supported business and clinical trainings between September 2011 and August 2013 and on the intensity of participation (low or high). Interaction effects between clinic characteristics and outcomes were explored.

Results

Clinic characteristics were predictive of any participation in trainings but were less useful in predicting intensity of participation. Clinics affiliated with larger (adjusted odds ratio [AOR]=.65, p<.01), more efficient agencies (AOR=.62, p<.05) and clinics that outsourced more clinical services (AOR=.60, p<.001) had lower odds of participating in any business-practice trainings. Participation in business trainings was associated with interaction effects between agency affiliation (hospital or community) and clinical staff capacity. Clinics with more full-time-equivalent clinical staff (AOR=1.52, p<.01) and a higher proportion of clients under age 18 (AOR=1.90, p<.001) had higher odds of participating in any clinical trainings. Participating clinics with larger proportions of youth clients had greater odds of being high adopters of clinical trainings (odds ratio=1.54, p<.01).

Conclusions

Clinic characteristics associated with uptake of business and clinical training could be used to target state technical assistance efforts.


The treatment of psychiatric disorders is among the five most costly medical conditions among children, constituting $8.9 billion or 9% of total annual medical spending for children (1). Research has identified a range of evidence-based practices for psychiatric conditions among youths (24). An ongoing challenge is the effective implementation of evidence-based practices in real-world settings; it takes 17 years for original research to change practice (5).

Over the past decade, states have led the way in promoting evidence-based practices to improve quality of care and reduce costs. Over 20 states are actively implementing evidence-based psychosocial therapies (6) or medication practices (7). Evidence-based practice implementation can be economically sound, with one study reporting a 56% rate of return on investment (8). The demand for states to implement evidence-based practices is matched by a demand for practical strategies and feasible measurement systems to track the impact of implementation (9,10).

New York State has been a leader in implementation of evidence-based practices within the child public mental health system. Propelled by the September 11, 2001, attacks and the expectation of trauma-related sequelae, New York State established the Child and Adolescent Trauma Treatments and Services Consortium with federal funds (11). The consortium’s success in conducting large-scale training in evidence-based trauma practices led to the establishment of the Evidence Based Treatment Dissemination Center (EBTDC) (12,13). The EBTDC provides in-depth technical assistance for clinicians around specific evidence-based treatments. However, data from the New York State Office of Mental Health (NYS OMH) indicate that providers at 52% of eligible clinics statewide have participated in any EBTDC initiatives, suggesting uneven adoption (14).

To expand training beyond traditional evidence-based treatments, New York State established the Clinic Technical Assistance Center (CTAC) in 2011. CTAC’s goals are to strengthen clinics’ capacity for meeting financial and regulatory challenges resulting from state and federal health care reforms. CTAC is thus focused on two core content areas: business and clinical evidence-informed practices. CTAC provides flexible training and consultation models that range from low-intensity, one-hour lunchtime Webinars to intensive learning collaboratives lasting six to 18 months.

Although substantial resources are invested in technical assistance, responses to such training have not been systematically examined. Nationally, strategies for scaling up evidence-based practices have not been guided by empirical knowledge about best practices for implementation (15), and the promising benefits of evidence-based practices have not been consistently realized (16). This article expands on an earlier article that described the training adoption behavior of clinics that are licensed by the NYS OMH to serve youths (17). In the first two years after the establishment of CTAC, 77% of clinics adopted a median of five trainings. Business-practice and clinical trainings were equally accessed, with lunchtime Webinars being the most popular modality. To date, no study has examined what factors, if any, are associated with clinics’ adoption of training. Understanding these factors has significant implications for state strategies to promote clinic training participation and reduce costs associated with training dropout.

Following a theoretical framework of adoption (18), we examined whether multilevel factors, ranging from extra-organizational to client-level factors, predicted clinics’ responses to training. Because predictors of clinic behavior may differ for various types of training, predictor variables for participation in business-practice and clinical trainings were examined separately. Using available administrative data and CTAC data on training participation, this study examined characteristics associated with any participation in business and clinical training and the intensity of participation among clinics that chose to enroll.

METHODS

Population

The study population included the 346 clinics licensed by the NYS OMH to treat youths in New York State. However, key fiscal and structural data of interest were not available for clinics affiliated with state-run facilities (N=17), which are structured differently, so these clinics were removed from our analyses. At these facilities, a higher proportion of clients were youth clients (<18 years old) (68.7%±45.2% versus 39.6%±33.4% at other clinics; t=−3.13, df=325, p<.01) and a higher proportion of youth clients had a serious emotional disturbance (76.3%±21.9% versus 35.0%±27.1% at other clinics; t=−4.77, df=306, p<.001). Thus our sample included the 329 outpatient clinics licensed by New York State to treat youths but not operated by the state.

Notably, clinics in New York State can be structured hierarchically, such that one or more clinics are licensed as part of a parent agency. The effects of clinic “nestedness” on each outcome were tested by using chi square tests. Except for one outcome (intensity of business-practice adoption), nestedness was considered ignorable, and agency-level variables were assumed to equally represent all clinics associated with an agency. For the outcome on intensity of business-practice adoption, no models were built, based on our model building criteria described below; none of the variables examined were associated with intensity of business-practice training participation.

Data Sources

Clinic predictor variables

Variables describing clinic characteristics that could be predictive of training adoption were extracted from four data sources. The U.S. Department of Health and Human Services Area Health Resources Files (AHRF) (ahrf.hrsa.gov) provided county demographic data. The NYS OMH directory of licensed clinics (bi.omh.ny.gov/ bridges/index) provided clinic demographic data. The 2011 NYS OMH Consolidated Fiscal Report (CFR) system (www.omh.ny.gov/omhweb/finance/main.htm) provided annual operational capacity and provider-level profiles. The 2011 NYS OMH Patient Characteristics Survey (PCS) (bi.omh.ny.gov/pcs/index) provided “snapshot” data collected during a one-week period in October 2011 on client populations served by the 329 clinics. Depending on the variable, 3.64% to 9.42% of clinics were missing data from the CFR system and 3.34% to 9.42% were missing data from the PCS. Observations with missing data were dropped from statistical analyses.

Outcome variables

Attendance data were prospectively tracked by training event (17). Because adoption of an evidence-based practice, defined as the decision to proceed with partial or complete implementation, is a distinct process that precedes implementation (1820), sending clinicians to be trained in an innovation represents an early adoption behavior by clinics. A clinic is counted as a training adopter if any individual provider affiliated with the clinic participates in any CTAC trainings. Among participating clinics, an average of 6.87±5.89 providers per clinic participated in any business-practice trainings; 7.09± 5.35 clinicians per clinic participated in any clinical trainings. For each outcome variable (business-practice or clinical training), clinics were categorized in terms of any training adoption (yes or no) as well as by intensity of participation among clinics that participated in at least one training (low or high). Low training adopters were defined by exclusive participation in hour-long Webinars, and high adopters were defined by participation in trainings that required in-person attendance, attendance at multiple events associated with a training, or both.

Measures of Clinic Variables

A theoretical model of innovation adoption developed by Wisdom and colleagues (18) provided a framework for clinic variable selection. Selection of the measures, described below, was guided by a review of constructs associated with four levels of predictors of innovation adoption (19).

Extraorganizational variables

Region-urbanicity served as a proxy for extraorganizational social networks (regions administratively linked by OMH) and training access (distance to urban locations where trainings are typically held). Clinics were initially categorized by region as downstate (New York City and Long Island) or upstate (Central, Hudson, and Western regions) and as urban or rural on the basis of AHRF county rural-urban continuum codes. Because downstate New York State is only urban, a new variable was used reflecting three categories of region-urbanicity: downstate-urban, upstate-urban, and upstate-rural.

Agency-level variables

Each NYS OMH–licensed clinic is associated with a parent agency. Relevant agency-level attributes included affiliation, operational size, fiscal efficiency, and operational structure. Affiliation refers to whether agencies are classified as community-based providers or hospitals; it served as a proxy for interorganizational social networks and operational structures. The agency’s annual clinic-related total expenses, used by NYS OMH as a measure of operational size, served as a proxy for a variety of theory-based variables, such as absorptive capacity and resource availability (19,20). The fiscal efficiency of clinics affiliated with an agency is a derived variable defined by NYS OMH as the annual average dollar gain or loss per unit of standardized service provided. This measure took into account the type and length of each service encounter. Operational structure, or the annual proportion of clinical workforce (providers who bill for direct services) to total workforce, was used as a proxy for staffing structure.

Clinic-provider profile variables

The CFR data provided two measures of clinic operational structure: clinical capacity, measured by the annual total clinical full-time-equivalents (FTEs), and outsourced services, measured by the proportion of all clinical staff whose services were contracted. Clinical capacity and degree of outsourcing are hypothesized to influence clinic investment in workforce training (strategic fit) (21).

Clinic-client profile variables

The PCS provided information that served as proxies for innovation values fit (22). These variables included the proportion of youth clients; proportion of Medicaid or Medicaid managed care (MMC) visits, estimated by the proportion of youth visits billed to Medicaid or MMC; and proportion of youth clients with a serious emotional disturbance, a measure of clinical severity among youths served.

Statistical Analyses

Analyses were performed by using Stata, version 11.2 (23). Descriptive statistics were used to characterize the clinic population according to the two outcome variables of interest: participation in business-practice trainings and clinical trainings. For each outcome variable, we examined clinic characteristics according to training adoption status (yes or no) and intensity of participation (low or high) among adopters. Chi square and t tests were used to compare categorical and continuous variables, respectively. Distributions of four continuous predictor variables (total expenses, gain or loss per unit of standardized service, total clinical FTEs, and proportion of clinical staff contracted out) were notably skewed; natural log-transformed versions of these variables were also tested for bivariate associations with the outcomes.

Multiple logistic regression models (adjusted odds ratios [AORs]) were used to assess the independent effects of predictor variables on clinic training participation. The same model-building procedure was used in four logistic regression models to predict any participation (yes or no) and intensity of participation (low or high) in each type of training (business and clinical).

The independent variables were entered into the models in the following ordered categories: extraorganizational variables, agency-level variables, clinic-provider profile variables, and clinic-client profile variables. Akaike Information Criterion values upon entry of each variable category into the model were compared to determine whether the information gained compensated for the increased model complexity and potential overfitting. Once the main effects for each model were determined, we explored the interactions of the following combinations of categorical and continuous variables, provided both pairs of variables were included among the main effects: region-urbanicity by total expenses, gain or loss per service unit, total clinical FTEs, and proportion of contracted clinical staff; and clinic affiliation by total expenses, gain or loss per service unit, and clinical capacity. Because none of the hospital-affiliated clinics outsourced any clinical services, interaction effects between clinic affiliation and proportion of services outsourced could not be examined. Once main effects and significant interaction terms were determined, reduced models created by removing statistically nonsignificant variables were tested to yield more parsimonious models with improved fit. For ease of interpretability and between-variable comparisons, all numeric variables were standardized for use in the models. Subsequently, model coefficients were interpreted as the change in odds of training adoption for an increase of one standard deviation in the respective predictor variable. All hypothesis tests were two-sided with alpha set at .05. Model fit of each final, reduced model was assessed with Hosmer and Lemeshow’s goodness-of-fit test by using eight predicted value groups. Specification errors were tested by checking for independent variable correlations with the model error term.

RESULTS

A majority of clinics were located in downstate, urban areas (59%) and were affiliated with community-based agencies (82%) (Table 1). The agencies with which the clinics were affiliated had average clinic-related total expenses of $6.04 million. Clinics operated at an average loss of $48.6 per unit of service provided and with a clinical workforce that represented 68.1% of the clinic’s overall workforce. Clinic-provider profiles indicated an average of 12.8 clinical FTEs on staff. On average, 8.3% of clinical services were provided by staff through outside contracts. Clinic-client profiles indicated that on average, youth clients constituted 39.7% of clients, of whom 49.8% were covered by Medicaid or MMC and 35.0% were classified as having a serious emotional disturbance.

TABLE 1.

Characteristics of 329 outpatient mental health clinics in New York State that serve youth clients (<18 years old))

Variable N %
Extraorganizational (region-urbanicity)
 Downstate-urban 194 59
 Upstate-urban 98 30
 Upstate-rural 37 11
Agency level
 Affiliation
  Community 269 82
  Hospital 60 18
 Total expenses, in millions (M±SD $) 6.04±6.91
 Gain or loss per service unit (M±SD $) −48.56±91.70
 Clinical staff as a percentage of total workforce (M±SD %) 68.14±16.02
Clinic-provider profile
 Clinical full-time-equivalents (M±SD) 12.75±10.11
 Proportion of clinical staff contracted out (M±SD %) 8.26±17.43
Clinic-client profile
 Youth clients (M±SD %) 39.65±33.35
 Visits by youth clients paid for by Medicaid or Medicaid managed care (M±SD %) 49.78±20.81
 Youth clients with a serious emotional disorder (M±SD %) 35.00±27.11

Table 2 reports clinic population characteristics by uptake of business-practice trainings. Compared with non-participants, clinics that participated in any business trainings were disproportionately located upstate (χ2=16.65, df=2, p<.001), had fewer total agency clinic-related expenses (t=2.21, df=315, p<.05), had more total clinical staff FTEs (t=−2.78, df=296, p<.01), had a smaller proportion of contracted clinical staff (t=3.09, df=296, p<.01), and served more youths (t=−2.31, df=311, p<.05) and more youths with a serious emotional disturbance (t=−2.36, df=296, p<.05). Among adopters of business-practice trainings, none of the variables examined were associated with intensity of training participation.

TABLE 2.

Characteristics of 329 outpatient mental health clinics in New York State that serve youth clients (<18 years old) and association with uptake of business-practice trainings

Variable No uptake (N=119, 36%)
Any uptake (N=210, 64%)
Any uptake vs. no uptakea
N % N % p AORb 95% CI p
Extraorganizational (region-urbanicity) <.001
 Downstate-urban 87 73 107 51
 Upstate-urban 26 22 72 34
 Upstate-rural 6 5 31 15
Agency level
 Affiliation ns
  Community (reference) 92 77 177 84
  Hospital 27 23 33 16 .50 .18–1.36 ns
 Total expenses, in millions (M±SE $) 7.19 ±.76 5.41 ±.44 <.05 .65 .50–.84 <.01
 Gain or loss per service unit (M±SE $) −45.1±7.5 −50.5±7.0 ns .62 .41–.94 <.05
 Clinical staff as a percentage of total workforce (M±SE %) 65.9±1.8 69.4±1.0 ns
Clinic-provider profile
 Clinical full-time-equivalents (FTEs) (M±SE) 10.6 ±.9 13.9 ±.7 <.01 1.33 .94–1.88 ns
 Proportion of clinical staff contracted out (M±SE %) 12.5±2.4 6.0 ±.9 <.01 .60 .46–.80 <.001
Clinic-client profile
 Youth clients (M±SE %) 33.9±3.2 42.9±2.3 <.05
 Visits by youth clients paid for by Medicaid or Medicaid managed care (M±SE %) 49.3±2.0 50.1±1.4 ns
 Youth clients with a serious emotional disorder (M±SE %) 30.0±2.6 37.7±2.0 <.05
Hospital (vs. community) affiliation × total clinical FTEs 4.89 1.31–18.28 <.05
a

Results are from a reduced logistic regression model for 287 clinics that removed statistically nonsignificant variables (p≥.05); pseudo R2=.1213, likelihood ratio χ2=44.71, df=6, p<.001

b

Adjusted odds ratio

Table 2 also displays results from reduced logistic regression models describing any business-practice training adoption as a linear function of the clinic characteristics. Clinics associated with agencies that had larger total expenses (AOR=.65, 95% confidence interval [CI]=.50–.84), had greater gains per unit of service provided (AOR=.62, CI=.41–94), outsourced more clinical care (AOR=.60, CI=.46–.80), and had lower odds of participating in any business-practice trainings. A statistically significant interaction effect of clinic affiliation and total clinical FTEs on the probability of uptake of any business-practice trainings was observed (AOR=4.89, CI=1.31–18.28). Among clinics with fewer total clinical FTEs, hospital-affiliated clinics were less likely to participate in any business trainings; among clinics with more total clinical FTEs, hospital-affiliated clinics were more likely to participate (Figure 1).

FIGURE 1.

FIGURE 1

Interaction effect of total number of clinical full-time-equivalents (FTEs) and affiliation of parent agency on probability of adoption of business-practice trainings by mental health clinics

Table 3 reports clinic population characteristics by clinical-training participation. Compared with nonparticipants, clinics that participated in any clinical trainings were more likely to be affiliated with agencies that had a larger proportion of clinical workforce (t=−2.08, df=316, p<.05). These clinics also had more total clinical staff FTEs (t=−2.84, df=296, p<.01) and served more youth clients (t=−4.93, df=311, p<.001) and more youth clients who were covered by Medicaid or MMC (t=−3.62, df=316, p<.001). Among participating clinics, only one variable was found to be associated with intensity of clinical-training uptake; on average, the percentage of youth clients was higher at clinics with high versus low participation (54.6%±3.7% versus 39.9%±3.0%, t=−3.14, df=203, p<.01).

TABLE 3.

Characteristics of 329 outpatient mental health clinics in New York State that serve youth clients (<18 years old) and association with uptake of clinical trainings

Variable No uptake (N=115, 35%)
Any uptake (N=214, 65%)
Any uptake vs. no uptakea
N % N % p AOR 95% CI p
Extraorganizational (region-urbanicity) ns
 Downstate-urban 77 67 117 55
 Upstate-urban 28 24 70 33
 Upstate-rural 10 9 27 13
Agency level
 Affiliation ns
  Community 89 77 180 84
  Hospital 26 23 34 16
 Total expenses, in millions (M±SE $) 5.77 ±.71 6.18 ±.46 ns
 Gain or loss per service unit (M±SE $) −44.2±8.1 −50.7±6.7 ns
 Clinical staff as a percentage of total workforce (M±SE %) 65.6±1.7 69.5±1.0 < .05
Clinic-provider profile
 Clinical full-time-equivalents (M±SE) 10.5 ±.8 13.9 ±.8 < .01 1.52 1.11–2.08 <.01
 Proportion of clinical staff contracted out (M±SE %) 9.7±1.9 7.5±1.2 ns
Clinic-client profile
 Youth clients (M±SE %) 27.3±2.8 46.1±2.4 < .001 1.90 1.42–2.55 <.001
 Visits by youth clients paid for by Medicaid or Medicaid managed care (M±SE %) 44.1±2.2 52.8±1.3 < .001
 Youth clients with a serious emotional disorder (M±SE %) 31.8±2.6 36.6±2.0 ns
a

Results are from a reduced logistic regression model for 294 clinics that removed statistically nonsignificant variables (p≥.05); pseudo R2=.079, likelihood ratio χ2=29.74, df=2, p<.001

b

Adjusted odds ratio

Reduced logistic regression models described any clinical-training participation as a linear function of the clinic characteristics (Table 3). Clinics with more clinical FTEs (AOR=1.52, CI=1.11–2.08) and a larger proportion of youth clients (AOR=1.90, CI=1.42–2.55) had higher odds of attending any clinical trainings. In the model predicting adoption intensity, clinics with a larger proportion of youth clients (OR=1.54, CI=1.17–2.05) had greater odds of high uptake (data not shown).

DISCUSSION

This study is the first, to our knowledge, to examine data on training participation collected prospectively from a population of state outpatient mental health clinics serving youths. To our knowledge, this is also the first attempt to test a multivariate model of adoption behavior involving multilevel characteristics proposed in current theoretical implementation models (18,20,24,25). Our results indicate that available administrative data can be valuable to states in predicting which clinics will participate in training. However, they were much less useful in terms of predicting the level of clinic participation; administrative data are primarily structural and lack more nuanced information—for example, provider attitudes—or local information—for example, details about the organizational culture or climate—that may influence clinic participation levels.

Various levels of clinic characteristics were found to be associated with adoption of different trainings. Business-practice training adoption was associated with agency (affiliation, size, and efficiency) and clinic-provider profile (outsourced clinical services) characteristics. By contrast, adoption of clinical trainings was associated with clinic-provider profile (clinical capacity) and clinic-client profile (proportion of youth clients) characteristics. These findings suggest that clinics make decisions on the basis of relevant agency, provider (strategic fit [21]), and client (innovation values fit [22]) factors.

We speculate that business-practice trainings were less attractive to clinics affiliated with larger agencies, which likely have in-house capacity for fiscal management or need different business-practice supports compared with clinics affiliated with smaller agencies. Clinics associated with more efficiently run agencies and those that outsource more clinical services were also less likely to participate in business-practice trainings. These agencies are likely to have less financial exposure and thus may be in less need of training in business practices. Our data suggest that clinical capacity may influence participation depending on agency affiliation. Among clinics with larger clinical capacity, hospital-affiliated clinics were more likely to participate in business-practice trainings. By contrast, among clinics with smaller clinical capacity, community-based clinics were more likely to adopt business-practice trainings. These findings suggest that operational structure or financial incentives may differ in different types of settings on the basis of clinical capacity, thus influencing business-practice adoption behavior.

Clinical trainings were more likely to be adopted by clinics with larger clinical capacity, which likely reflects clinics’ capacity to release clinical staff for training. Given that the trainings focused primarily on youth-related issues, it seems likely that clinics with a higher proportion of youth clients would consider the trainings more relevant and would be more likely to participate. In-depth interviews with a stratified random sample of these clinics are underway to test our hypotheses.

Implications for State Systems

Our findings suggest that state efforts to provide incentives or target training efforts should pay attention to specific clinic characteristics available through administrative data. Policy makers should understand factors that influence the types of training and the amount of training that clinics are willing or able to adopt. To help clinics move beyond training adoption to successful implementation, targeting training to address both clinic needs and clinics’ readiness for change is critical (21). Health care reform and new accountability standards make it imperative for states to understand and leverage factors that influence clinic decisions to embrace, implement, and sustain quality improvement efforts.

Limitations

When used as the only data source, administrative data have limited value in predicting adoption of training. It may be helpful to collect and examine data about other theoretical factors that are hypothesized to drive adoption behavior in advance of large-scale rollouts of training initiatives. Information about providers—such as attitudes, organizational context, and baseline competencies about business practices and clinical knowledge—must be integrated with administrative data to help policy makers determine appropriate levels of technical support for successful quality improvement efforts.

In addition, clinics may adopt training not associated with CTAC; we were unable to account for the relationship between adoption of CTAC trainings versus other training. National differences in relationships between clinics, counties, and states may limit the generalizability of our findings outside New York State. Nonetheless, our findings provide support for existing theoretical adoption or implementation models that explicate multilevel influences.

CONCLUSIONS

This study represents a first effort to understand multilevel characteristics that influence training adoption within a state outpatient child mental health system. Although this study focused on baseline adoption behaviors (training participation), how adoption translates into actual implementation of effective practices to improve care quality and youth outcomes is a critical question. The answer is likely to inform development of optimal models for rollout of state training in service practices.

Acknowledgments

This study was supported by grants P30 MH090322 and P20 MH085983 from the National Institute of Mental Health and by the Clinic Technical Assistance Center, funded by the New York State Office of Mental Health.

Footnotes

The authors report no financial relationships with commercial interests.

References

  • 1.Mental Disorders Among Most Costly Conditions in Children. Rockville, Md: Agency for Healthcare Research and Quality; 2009. Available at archive.ahrq.gov/news/newsroom/news-and-numbers/042209.html. [Google Scholar]
  • 2.Burns BJ, Hoagwood KE. Evidence-based practice, part I. a research update. Child and Adolescent Psychiatric Clinics of North America. 2004;13:xi–xiii. doi: 10.1016/j.chc.2004.11.001. [DOI] [PubMed] [Google Scholar]
  • 3.Kazdin AE, Weisz JR, editors. Evidence-Based Psychotherapies for Children and Adolescents. New York: Guilford; 2003. [Google Scholar]
  • 4.National Registry of Evidence-based Programs and Practices. Rockville, Md: Substance Abuse and Mental Health Services Administration; [Accessed May 13, 2014]. Available at nrepp.samhsa.gov/AboutNREPP.aspx. [Google Scholar]
  • 5.Green LW. From research to “best practices” in other settings and populations. American Journal of Health Behavior. 2001;25:165–178. doi: 10.5993/ajhb.25.3.2. [DOI] [PubMed] [Google Scholar]
  • 6.Bruns EJ, Hoagwood KE. State implementation of evidence-based practice for youths, part I. responses to the state of the evidence. Journal of the American Academy of Child and Adolescent Psychiatry. 2008;47:369–373. doi: 10.1097/CHI.0b013e31816485f4. [DOI] [PubMed] [Google Scholar]
  • 7.Essock SM, Covell NH, Leckman-Westin E, et al. Identifying clinically questionable psychotropic prescribing practices for Medicaid recipients in New York State. Psychiatric Services. 2009;60:1595–1602. doi: 10.1176/ps.2009.60.12.1595. [DOI] [PubMed] [Google Scholar]
  • 8.Aos S, Mayfield J, Miller M, et al. Washington State Evidence-Based Treatment of Alcohol, Drug, and Mental Health Disorders: Potential Benefits, Costs and Fiscal Impacts on Washington State. Olympia: Washington State Institute for Public Policy; 2006. [Google Scholar]
  • 9.Ganju V. Implementation of evidence-based practices in state mental health systems: implications for research and effectiveness studies. Schizophrenia Bulletin. 2003;29:125–131. doi: 10.1093/oxfordjournals.schbul.a006982. [DOI] [PubMed] [Google Scholar]
  • 10.Torrey WC, Drake RE, Dixon L, et al. Implementing evidence-based practices for persons with severe mental illnesses. Psychiatric Services. 2001;52:45–50. doi: 10.1176/appi.ps.52.1.45. [DOI] [PubMed] [Google Scholar]
  • 11.Hoagwood KE, Vogel JM, Levitt JM, et al. Implementing an evidence-based trauma treatment in a state system after September 11: the CATS project. Journal of the American Academy of Child and Adolescent Psychiatry. 2007;46:773–779. doi: 10.1097/chi.0b013e3180413def. [DOI] [PubMed] [Google Scholar]
  • 12.North MS, Gleacher AA, Radigan M, et al. Evidence-Based Treatment Dissemination Center (EBTDC): bridging the research-practice gap in New York State. Report on Emotional and Behavioral Disorders in Youth. 2008;8:9–17. [Google Scholar]
  • 13.Gleacher AA, Nadeem E, Moy AJ, et al. Statewide CBT training for clinicians and supervisors treating youth: the New York State Evidence Based Treatment Dissemination Center. Journal of Emotional and Behavioral Disorders. 2011;19:182–192. doi: 10.1177/1063426610367793. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14.OMH Clinic Reporting Participation—Internal Report Based on Analyses From SPARCS Data. Albany, New York: State Office of Mental Health; 2010. [Google Scholar]
  • 15.McHugh RK, Barlow DH. The dissemination and implementation of evidence-based psychological treatments: a review of current efforts. American Psychologist. 2010;65:73–84. doi: 10.1037/a0018121. [DOI] [PubMed] [Google Scholar]
  • 16.Fixsen D, Blase K, Metz A, et al. Statewide implementation of evidence-based programs. Exceptional Children. 2013;79:213–230. [Google Scholar]
  • 17.Chor KHB, Olin SS, Weaver J, et al. Adoption of clinical and business trainings by child mental health clinics in New York State. Psychiatric Services. 2014;65:1439–1444. doi: 10.1176/appi.ps.201300535. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18.Wisdom JP, Chor KHB, Hoagwood KE, et al. Innovation adoption: a review of theories and constructs. Administration and Policy in Mental Health. 2014;41:480–502. doi: 10.1007/s10488-013-0486-4. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19.Chor KHB, Wisdom JP, Olin SS, et al. Measures for predictors of innovation adoption. Administration and Policy in Mental Health. doi: 10.1007/s10488-014-0551-7. (Epub ahead of print, April 17, 2014) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 20.Aarons GA, Hurlburt M, Horwitz SM. Advancing a conceptual model of evidence-based practice implementation in public service sectors. Administration and Policy in Mental Health. 2011;38:4–23. doi: 10.1007/s10488-010-0327-7. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21.Panzano PC, Sweeney HA, Seffrin B, et al. The assimilation of evidence-based healthcare innovations: a management-based perspective. Journal of Behavioral Health Services and Research. 2012;39:397–416. doi: 10.1007/s11414-012-9294-y. [DOI] [PubMed] [Google Scholar]
  • 22.Klein KJ, Sorra JS. The challenge of innovation implementation. Academy of Management Review. 1996;21:1055–1080. [Google Scholar]
  • 23.Stata 11.2. College Station, Tex. StataCorp; [Accessed March 13, 2014]. Available at www.stata.com. [Google Scholar]
  • 24.Feldstein AC, Glasgow RE. A practical, robust implementation and sustainability model (PRISM) for integrating research findings into practice. Joint Commission Journal on Quality and Safety. 2008;34:228–243. doi: 10.1016/s1553-7250(08)34030-6. [DOI] [PubMed] [Google Scholar]
  • 25.Damschroder LJ, Aron DC, Keith RE, et al. Fostering implementation of health services research findings into practice: a consolidated framework for advancing implementation science. Implement Science. 2009;4:50. doi: 10.1186/1748-5908-4-50. [DOI] [PMC free article] [PubMed] [Google Scholar]

RESOURCES