Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2021 Jun 3.
Published in final edited form as: Clin Psychol (New York). 2020 Sep;27(3):e12330. doi: 10.1111/cpsp.12330

Therapist training in evidence-based interventions for mental health: A systematic review of training approaches and outcomes

Hannah E Frank 1, Emily M Becker-Haimes 2,3, Philip C Kendall 1
PMCID: PMC8174802  NIHMSID: NIHMS1583902  PMID: 34092941

Abstract

A lack of effective therapist training is a major barrier to evidence-based intervention (EBI) delivery in the community. Systematic reviews published nearly a decade ago suggested that traditional EBI training leads to higher knowledge but not more EBI use, indicating that more work is needed to optimize EBI training and implementation. This systematic review synthesizes the training literature published since 2010 to evaluate how different training models (workshop, workshop with consultation, online training, train-the-trainer, and intensive training) affect therapists’ knowledge, beliefs, and behaviors. Results and limitations for each approach are discussed. Findings show that training has advanced beyond provision of manuals and brief workshops; more intensive training models show promise for changing therapist behavior. However, methodological issues persist, limiting conclusions and pointing to important areas for future research.

Keywords: dissemination, evidence-based practice, implementation, mental health, training

1 |. INTRODUCTION

Psychological treatments have been developed and found to be efficacious for a range of mental health disorders (e.g., substance abuse, anxiety, depression), yet most therapists do not receive training in or use evidence-based interventions (EBIs; treatment techniques that have empirical support for improving mental health outcomes1) in routine practice (Garland, Bickman, & Chorpita, 2010; Gyani, Shafran, Myles, & Rose, 2014; Shiner et al., 2013). The field of implementation science (Eccles & Mittman, 2006) has identified strategies to facilitate the transfer of knowledge into settings where mental health services are most often delivered (e.g., community mental health; Ringel & Sturm, 2001). However, lack of EBI training remains a primary driver of poor access to effective mental health services, particularly for vulnerable populations (Kilbourne et al., 2018; Weissman et al., 2006). Conceptual models of implementation highlight that therapist use of EBIs is influenced by a complicated web of intersecting variables (e.g., organizational, fiscal, policy factors; Damschroder et al., 2009). Nonetheless, improving training remains a necessary step for successful EBI implementation and sustainment.

Many studies have examined the role of training to facilitate EBI implementation into practice. In 2010, three reviews of the effectiveness of training were published (Beidas & Kendall, 2010; Herschell, Kolko, Baumann, & Davis, 2010; Rakovshik & McManus, 2010). Each took a slightly different approach to evaluating the literature; however, all three highlighted the potential for optimizing training procedures as a strategy for increasing therapist EBI use. Overall, findings suggested that therapist self-study of treatment manuals was generally ineffective at increasing EBI use (Fixsen, Naoom, Blase, Friedman, & Wallace, 2005). In-person and online training courses demonstrated some effectiveness for increasing therapist knowledge and self-efficacy, but had limited effect on provider behavior and client outcomes (e.g., Dimeff et al., 2009; McHugh & Barlow, 2010). Prior reviews also indicated that using active training strategies, such as behavioral rehearsal (role play), led to higher adherence and skill on behavioral rehearsals after training, without consistently translating to greater use in practice (Beidas & Kendall, 2010). Given that increasing implementation of EBIs in clinical practice is the ultimate goal of therapist training, further examination of effective training strategies and their impact on practice is warranted.

The 2010 reviews also highlighted the importance of training that moves beyond the traditional workshop format (e.g., providing ongoing consultation or supervision2 following workshops) to increase therapist EBI use. In particular, consultation following workshops led to increased knowledge and skill and more use of the target EBI in practice (Beidas, Edmunds, Edmunds, Marcus, & Kendall, 2012; Schoenwald, Sheidow, & Letourneau, 2004). Multi-component trainings (i.e., using multiple of the following: in-person workshops, consultation, supervisor training, booster training) also demonstrated better outcomes than training approaches that relied on one or two components. These earlier reviews were consistent in calling for more research in areas related to therapist training strategies, including web-based training, consultation/supervision practices, and “train-the-trainer” (TT) methods, as well as the need for direct comparisons of different training methods. Such work has relevance in identifying specific implementation strategies to increase the uptake of EBIs in routine clinical care. All prior reviews also highlighted methodological issues that limited drawing firm conclusions, including limited use of theory to guide study design, lack of comprehensive measurement of target constructs (i.e., self-report only; no gold-standard outcome measures), and use of nonstandardized measures.

Despite these calls for more research and advances in implementation science (Williams & Beidas, 2019), EBI use in routine practice remains low (Becker, Smith, & Jensen-Doss, 2013; Beidas et al., 2019; Stirman et al., 2012). Active training strategies combined with post-training consultation are currently referred to as “gold-standard” packages, yet their dosage is not well defined, and they have yielded disappointing results (e.g., Beidas et al., 2019). Concerns about the feasibility and sustainability of strategies such as expert consultation and session feedback have been raised due to being resource-intensive and relatively short-term (McHugh & Barlow, 2010). Understanding how to optimize training strategies for EBI is key for improving access to EBIs for treatment-seeking individuals. A comprehensive review of the research conducted in this space since 2010 is needed to characterize the current state of the science for EBI training strategies to identify effective strategies and future directions for research on therapist training.

This review focused on five types of training: workshop only (an “in-person” time-limited training opportunity that focuses on a single intervention topic), workshop plus consultation (workshop plus additional expert consultation/supervision after conclusion of the initial training), online training (training conducted via an online platform with either asynchronous [self-paced] or synchronous learning, with or without consultation), TT (an expert trainer trains a set of local trainers who in turn will train local therapists, alternatively called a “cascade” model of training), and intensive training (defined as at least 20 hours of training plus two or more additional training components, such as consultation, direct observation of practice, or learning collaboratives). Outcomes of interest for this review represented a range of implementation and effectiveness constructs, including the following: (a) therapist attitudes, knowledge, self-efficacy, and training satisfaction; (b) EBI adherence, competence, and use; (c) implementation barriers; and (d) client outcomes. This review also examined the degree to which previously identified limitations (e.g., lack of theory, nonstandard measures) have been addressed. Finally, attention was paid to the extent to which therapist characteristics (e.g., age, experience) affected training outcomes, given work suggesting clinical outcomes may vary by therapist (Huppert et al., 2001). Although EBI training efforts might target therapists completing graduate education (i.e., “pre-service”), this review focused on strategies for training mental health therapists employed in community practice settings, including hospitals, private practice, and community mental health settings (i.e., “post-service” therapists).

2 |. METHOD

A systematic literature search identified all relevant articles published since the 2010 reviews (Beidas & Kendall, 2010; Herschell et al., 2010; Rakovshik & McManus, 2010) between March 2010 and May 2018. An initial search for articles published between 2010 and 2017 was conducted in PsycINFO and PubMed using the following key words: training, workshop, dissemination, implementation, TT, web-based training, online training, adoption, therapist, clinician, community, mental health, behavior health, evidence-based, empirically supported, intervention, evidence-based practice, empirically-supported treatment. Due to the length of the review process, Google alerts for the phrases “dissemination and implementation” and “clinician training evidence-based practice” identified articles from November 2017 to May 2018. Finally, consistent with Greenhalgh et al. (2004) and Herschell et al. (2010), a snowball method augmented the database search, whereby references of included articles were used to identify additional citations. Abstracts were screened by the first and second authors, supported by Rayyan software (Ouzzani, Hammady, Fedorowicz, & Elmagarmid, 2016). Regular meetings were used to discuss uncertainties about an abstract’s relevance with decisions made by consensus. Relevant articles were full-text reviewed to determine inclusion. Figure 1 illustrates the PRISMA study flow diagram.

FIGURE 1.

FIGURE 1

PRISMA flow diagram indicating study identification, screening, and selection

Studies were included if (a) the focus was on training mental health providers with an advanced degree (i.e., a medical degree, doctorate, or master’s degree) in a mental health field (e.g., social work, psychology, psychiatry, substance abuse counseling), and (b) the training focus was an evidence-based mental health or substance abuse intervention. Studies additionally had to report on at least one of the following training outcomes: post-training attitudes, knowledge, self-efficacy, satisfaction, adherence, competence, or EBI use, include at least some quantitative data, be a peer-reviewed study, and be in the English language. Inclusion criteria were intended to allow for the broadest possible sample of studies aimed at improving EBI use in mental health and thus allowed for inclusion of studies with different training foci, durations, and approaches. Exclusion criteria included the following: (a) more than half of participants were not mental health providers (e.g., teachers, occupational therapists, nurses) or were high-school or bachelor’s level therapists (e.g., peer specialists), (b) the target EBI was not a mental health intervention, (c) the study used an identical sample to a parent study included in the review, or (d) the training was geared toward graduate students (e.g., semester-long courses, graduate education initiatives) and not practicing therapists. This latter exclusion criterion was selected because implementation challenges are thought to differ for students relative to postgraduate, practicing therapists (Becker-Haimes et al., 2018). Included studies were classified as follows: workshop only, workshop with consultation, online training, TT, or intensive trainings. Studies that examined multiple training approaches were classified by the modality of the active comparison condition (e.g., if a study compared an online workshop to traditional in-person workshops, it was included in the online training section).

Consistent with Herschell et al. (2010), studies were classified according to Nathan and Gorman’s (2002, 2007) criteria for evaluating methodological rigor. All studies in this review were classified as Type 1, Type 2, or Type 3; as noted above, studies needed to include quantitative data analysis; as such, Type 4 (reviews with secondary data analysis), Type 5 (reviews without secondary data analysis), and Type 6 (case studies, essays, opinion papers) studies were excluded. Type 1 studies are randomized controlled trials (RCTs) with comparison conditions, random assignment, blind assessments, clearly defined inclusion and exclusion criteria, state-of-the-art diagnostic methods, and sufficient statistical power and description of statistical methods. Type 2 studies are clinical trials missing one or more of the criteria for Type 1 studies (e.g., lack of random assignment, underpowered), but not fatally flawed. Finally, Type 3 studies have clear methodological limitations and are typically uncontrolled studies using pre-post design and retrospective design. The first author coded all studies for classification as Type 1, Type 2, or Type 3. The last author reviewed a random subset of studies (n = 16; 21.1%) and discussed classification with the first author, obtaining consensus of classification for 100% of studies reviewed. A minority (12%; n = 9) of studies were Type 1, 14% (n = 11) were Type 2, and 74% (n = 56) were Type 3. Most Type 2 studies lacked clear inclusion criteria or random assignment to condition. Most Type 3 studies had a lack of comparison conditions (i.e., within-subjects design) or lack of randomization.

3 |. RESULTS

3.1 |. Summary of the literature

A total of 76 reports met inclusion criteria. Identified studies examined in-person workshop only (n = 9), in-person workshop and consultation (n = 21), online workshop (n = 20), TT (n = 4), and intensive trainings (n = 22). No identified studies examined treatment manuals or written materials alone. The following sections detail and review the approach, findings, and limitations of studies using each training modality, consistent with Herschell et al. (2010). Additional details for each study are in Tables 15, and an overview of findings for all studies is shown in Table 6.

TABLE 1.

Workshop only

Author Nathan & Gorman classification Sample size Training topic Amount of training Comparison groups Follow-up Role of consultation Outcome measure domains
Farrell et al. (2016) 2 N = 49 mental health clinicians Exposure therapy for anxiety disorders 8 hr 2: Standard versus Enhanced training None None A, K, S
Chin et al. (2018) 3 N = 53 community clinicians Prolonged exposure for PTSD 8 hr None None None A, Int
Crisanti et al. (2016) 3 n = 20 licensed behavioral health practitioners + n = 15 peer support workers Seeking Safety for trauma 6 hr None None None K, S, Sat
Deacon et al. (2013) 3 N = 162 community mental health professionals Exposure therapy for anxiety disorders 7 hr None None None A
Lim et al. (2012) 3 N = 268 public sector youth mental health providers EBIs for internalizing and externalizing disorders Not specified None None None A, K
Mirick et al. (2016) 3 N = 543 participants Suicide assessment and intervention 6 hr None None None C, K
Richards et al. (2011) 3 N = 73 mental health clinicians Acceptance and Commitment Therapy 1 day None 1 year None K, P, PF
Scott et al. (2016) 3 N = 30 therapists CBT 1 day None None None K
Waller et al. (2016) 3 N = 34 clinicians Exposure therapy for eating disorders 1.5 hr None None None A

Abbreviations: A, attitudes; C, confidence (self-efficacy); CBT, cognitive behavioral therapy; Cl, clinical outcome; F, treatment fidelity or adherence; I, implementation difficulty or barrier—anticipated or actual; Int, intentions; K, knowledge; P, practices or techniques used; PF, psychological flexibility; S, skills/competence; Sat, satisfaction/acceptability; T, therapeutic interaction/rapport/working alliance.

TABLE 5.

Intensive training

Author Nathan &Gorman classification Sample size Training topic Amount of training Comparison groups Follow-up Role of consultation Outcome measure domains
Kolko et al. (2012) 1 N= 128 practitioners;
N = 34 supervisors
Alternative for Families (AF-CBT) 32 hr 2; learning community; training as usual 18 months 10 90-min biweekly group consultation A, O, P, K, Sat
German et al. (2017) 2 N = 214 clinicians CBT 22 hr 2; In-person training, expert-led consultation; Web-based training; peer- led consultation 6 months Weekly 2-hr meetings for 6-months with experts; then with peers only K, S
Stirman et al. (2017) 2 N = 85 clinicians Cognitive Therapy 22 hr 2; individual consultation; group consultation 2 years (from baseline) Individual: 1-hr individual; 1-hr group; Group: 2 hr group S
Webster- Stratton et al. (2014) 2 N = 56 community mental health therapists Incredible Years for Conduct Problems 3 days 2; workshop only; workshop + consultation None Weekly consultation calls (duration not reported); Written feedback on video recordings F
Beveridge et al. (2015) 3 N = 143 therapists PCIT 40 hr + two 1-day advanced trainings held 2 months after initial training None None 1-hr biweekly phone consultation group until PCIT completed with 2 clients Sat
Creed et al. (2013) 3 N = 25 school- based therapists Cognitive Therapy 22 hr None 6 months Weekly 2-hr meetings for 6 months K, S, Sat
Creed et al. (2016) 3 N = 321 community mental health therapists CBT 22 hr None 6 months Weekly 2-hr meeting for 6 months S
Herschell et al. (2014) 3 N = 64 mental health therapists DBT 2-day clinical + 1-day admin overview; two 5-day workshops (6-months apart); 2-day follow-up training None 6-, 14-, and 22-months post-training Weekly phone consultation for 14 months A, C, P
Jackson et al. (2017) 3 N = 32 clinicians PCIT 40 hr; advanced 16-hr training 6 months later None 2 years (from baseline) Up to 24 1-hr consultation calls over 1 year I, K, S
Karlin et al. (2012) 3 N = 221 mental health therapists;
N = 356 veteran patients
CBT for depression 3 days None 6-month postconsult.; 3–12 months post-training 6 months of 90-min weekly group calls; submission of audio recordings A, C, Cl, Int. P, S, Sat
Karlin et al. (2013) 3 N = 102 clinicians; N = 182 veteran patients CBT for insomnia 3 days None None 4 mo. of weekly 90-min group calls; submission of audio recordings Cl, S
Lopez and Basco, (2011) 3 N =7 therapists CBT for major depressive disorder 36 hr None None 5 months of 1-hr group phone supervision 2x/week; submission of audiotaped sessions A, Cl, S
Manber et al. (2013) 3 N = 207 therapists CBT for insomnia 3 days None 6 months post-consultation 4 months of weekly 90-min group consultation calls; submission of audiotaped sessions A, C
McManus et al. (2010) 3 N = 278 trainees CBT 36 days over 12-week term; 5 hr workshops each day None None 90 min. supervision at each workshop; submission of audiotaped sessions S
Navarro-Haro et al. (2019) 3 N = 412 participants DBT for borderline personality disorder 2 sets of 5-day workshops (6-months apart) None None Contact with trainers, but no formal consultation A, C, I, P
Ruzek et al. (2016) 3 N = 943 licensed mental health clinicians Prolonged exposure for PTSD 4 days None Post-consultation 6–9 months of 60-min weekly individual and group consultation; audiotape review A, C, Int
Ruzek et al. (2017) 3 N= 1,034 clinicians Prolonged exposure for PTSD 4 days None Post-consultation; 6 months post-consultation Weekly individual (30- min) and group (60-min) consultation for at least two cases; audiotape review A, C, Int, P
Shah, Scogin, Presnell, Morthland, and Kaufman (2013) 3 N = 5 therapists;
N =88 patients
CBT for rural adults 24 hr None Post-treatment Weekly group supervision and feedback on mock sessions (duration not reported) Cl, F, S
Smith et al. (2017) 3 Study 1:N= 234 regional learners; Study 2: N = 24 blended pilot learners; Study 3: N= 40 blended pilot learners Study 1: Cognitive Processing Therapy; Study 2: CBT for depression; Study 3: Prolonged exposure for PTSD Study 1: 3 days
Study 2: 7 hr online + 7 weekly 1–2hr. webinars
Study 3: web-based course + 5 weekly 2-hr webinars
Each study compared alternative training models to traditional VA training model Study 1: none; Study 2:4 months postconsultation; Study 3:6 months postconsultation All studies: Weekly consultation with tape review (call duration not reported)
Blended learning: 4 months Traditional learning: 6 months
Study 1: Cl, Sat
Study 2: S, Sat
Study 3: Cl, K, S
Stephan, Connors, Arora, and Brey (2013) 3 Not reported Core elements of EBIs for anxiety, depression, disruptive behavior disorders, substance abuse 4 days over 13 months None Post-training (13 months) Monthly technical assistance (duration not reported); Learning Collaborative K, P
Walser et al. (2013) 3 N = 391 therapists;
N = 745 patients
ACT for depression 3 days None Post-consultation (3–12 months) 6-months of 90 min weekly consultation; feedback on audiotaped sessions A, C, Cl, Int, K, S, T
Williams, Martinez, Dafters, Ronald, and Garland (2011) 3 N = 267 therapists CBT self-help workbook 38.5 hr None Post-intervention, 3-month follow-up Weekly supervision (5 hr total) K, P, S, Sat

Abbreviations: A, attitudes; ACT, acceptance and commitment therapy; C, confidence (self-efficacy); CBT, cognitive behavioral therapy; Cl, clinical outcome; DBT, dialectical behavior therapy; F, treatment fidelity or adherence; I, implementation difficulty or barrier—anticipated or actual; Int, intentions; K, knowledge; P, practices or techniques used; PCIT, parent–child interaction therapy; PF, psychological flexibility; PTSD, posttraumatic stress disorder; S, skills/competence; Sat, satisfaction/acceptability; T, therapeutic interaction/rapport/working alliance, VA, Veterans Affairs.

T A B L E 6.

Summary of results

Workshop type Conclusions Limitations
Workshop only
  • Demonstrated improvements in declarative knowledge and attitudes compared to baseline

  • Limited evidence for change in therapist behaviors

  • Primarily non-randomized, naturalistic designs

  • Mostly self-report and non-psychometrically evaluated measures

Workshop and consultation
  • Demonstrated improvements in therapist behaviors, including competence and self-reported intervention use, and in client symptoms compared to baseline

  • More participation in consultation calls led to better outcomes (e.g., adherence, skill, intervention use)

  • Following consultation, knowledge and attitudes did not improve above and beyond workshop only conditions

  • Few studies had comparison conditions

  • Many self-report and non-psychometrically evaluated measures

  • Limited data on sustainability

  • Variability in length and frequency of consultation calls

Online training
  • Demonstrated improvements in knowledge and skill relative to baseline; some demonstrated increases in intervention use

  • Generally comparable outcomes to in-person training

  • Inclusion of consultation led to better competence

  • Self-report measures of intervention use

  • Relatively short follow-up periods

  • Some studies showed lack of maintenance of gains

Train-the-trainer
  • Demonstrated improvements in competence, reaching adequate/proficient competence after training

  • Expert-led trainings resulted in higher proficiency than train-the- trainer condition

  • Variability in qualifications to be a trainer

  • Not all studies had comparison conditions

Intensive training
  • Demonstrated improvements in therapist knowledge, intervention use, and observer-rated competence

  • Few studies had comparison conditions

  • Variability in configuration, timing, and components of training

  • Resource and cost-intensive

3.2 |. Workshop only

3.2.1 |. Description of studies

Of the workshop-only studies (n = 9; Table 1), one was Type 2 and eight were Type 3. Most studies used within-subject designs and only included pre- and post-training assessments without follow-up. Sample sizes varied widely, ranging from 30 to 268 participants working primarily in community mental health and public sector settings. Workshop length ranged from 90 minutes to 8 hours. Outcome measures were all self-report, and client outcomes were not examined. Knowledge was the most commonly assessed outcome (in six of nine studies). Two studies used psychometrically supported assessments of knowledge (Lim, Nakamura, Higa-McMillan, Shimabukuro, & Slavin, 2012; Scott, Klech, Lewis, & Simons, 2016); the remainder used knowledge measures developed for the study. Attitudes were the second most frequently examined outcome (five of nine studies); four studies used attitude measures with acceptable psychometric properties (i.e., the Evidence-Based Practice Attitudes Scale (EBPAS; Aarons, 2004) or the Therapist Beliefs about Exposure Scale (TBES; Deacon, Lickel, Farrell, Kemp, & Hipol, 2013).

3.2.2 |. Summary of findings

Consistent with prior reviews, studies of in-person workshops without additional consultation or follow-up indicated improvement in knowledge and attitudes about EBIs after training, although results were variable. Scott et al. (2016) demonstrated that even though training led to improvements in knowledge, overall post-training knowledge scores remained low. Lim et al. (2012) found that providers over-generalized the label for EBIs, applying it to non-EBI strategies. Farrell, Kemp, Blakey, Meyer, and Deacon (2016) conducted the only study comparing two training conditions and found that therapists who received an enhanced training focused on discussing therapist concerns about exposure therapy had significantly greater reductions in such concerns than those in standard training. Those in enhanced training also self-reported superior quality delivery of exposure therapy based on case vignettes.

Several studies examined pre-training therapist characteristics as predictors of training outcomes (Lim et al., 2012; Scott et al., 2016; Waller, Walsh, & Wright, 2016). Results were mixed regarding whether prior experience with the EBI or more favorable attitudes toward EBIs are associated with post-training changes in knowledge and attitudes (Deacon et al., 2013; Lim et al., 2012; Mirick, McCauley, Bridger, & Berkowitz, 2016; Scott et al., 2016; Waller et al., 2016). Richards et al. (2011) reported that participants with higher psychological flexibility had more knowledge post-training. Therapist skill was examined in two studies (Crisanti, Murray-Krezan, Karlin, Sutherland-Bruaw, & Najavits, 2016; Farrell et al., 2016) and EBI use was only examined in one study (Richards et al., 2011), all using self-report measures. Thus, these findings provide limited information about the extent to which workshops alone changed therapist behaviors.

3.2.3 |. Limitations of studies reviewed

Workshop-only studies were predominantly within-subject, nonrandomized, and did not include follow-ups. Most had a naturalistic design that allowed for examination of training efforts that were already underway but were not designed to examine empirical questions. Most of the measures were self-report and had not been psychometrically evaluated, limiting generalizability. In addition, comparison across studies was difficult given variability in the length of training and training approach. Most workshops were reported to include active learning components (e.g., role plays), citing previous research suggesting their importance, but most studies did not expand on the ways or frequency with which these strategies were implemented.

3.3 |. Workshop and consultation

3.3.1 |. Description of studies

Of the workshop and consultation studies (n = 21), only one was Type 1, an RCT comparing different training approaches (Beidas, Edmunds, et al., 2012). There were three Type 2 studies, none of which had explicit inclusion/exclusion criteria, and some of which were lacking in blind assessments or a sufficient sample size. Finally, there were 17 Type 3 studies with predominantly pre-post designs, though some included follow-ups, quasi-experimental designs, retrospective surveys, and qualitative interviews. Of note, several studies had sub-studies from the same datasets (Beidas, Mychailyszyn, et al., 2012; Leathers, Melka-Kaffer, Spielfogel, & Atkins, 2016; Lloyd et al., 2015), which are not included in this review.

Sample sizes ranged from 7 to 1,107 participants working in community, public sector, or Veterans Affairs (VA) settings. Workshop lengths varied from 4 hours to 4 days. The most frequently assessed outcomes included attitudes (10 studies), intervention use (10 studies), and competence (nine studies). A minority of studies used measures of attitudes with psychometric support (Hamblen, Norris, Gibson, & Lee, 2010; Pemberton et al., 2017); most used attitude measures developed for each study. Intervention use was assessed by therapist self-report in all studies, typically using measures designed for each study. Competence, assessed in nine studies, was the only outcome measure that consistently used observer ratings. Competence was assessed through role plays with standardized clients (Beidas, Edmunds, et al., 2012), observation of actual sessions (Brookman-Frazee, Drahota, & Stadnick, 2012; Henggeler, Chapman, Rowland, Sheidow, & Cunningham, 2013; Lu et al., 2012; Simons et al., 2010), or both (Petry, Alessi, & Ledgerwood, 2012). Studies that included client outcomes (n = 6) typically had limited or no exclusion criteria for clients and were said to be representative of clients in community settings.

Consultation was weekly or biweekly for most studies and typically occurred over the phone or via video chat for up to 1-year post-training. Most consultation was in a group format for 60–90 min. However, some studies offered individual consultation calls that lasted 30–60 min. One study provided on-site weekly group supervision (Lu et al., 2012), and others provided feedback on fidelity based on reviews of recorded sessions (e.g., Brookman-Frazee et al., 2012; Petry et al., 2012). Leather, Spielfogel, et al. (2016) incorporated a project-trained change agent (i.e., someone who demonstrates advanced knowledge and generates interest in the intervention) into the agency setting. The change agent initiated informal discussions about the intervention and offered formal consultation, which was rarely used (Table 2).

TABLE 2.

Workshop and consultation

Author Nathan & Gorman classification Sample size Training topic Amount of training Comparison groups Follow-up Role of consultation Outcome measure domains
Beidas, Edmunds, et al. (2012) 1 N= 115 community clinicians CBT for child anxiety 6 hr 3: Routine, Computerized, & Augmented training 3 months Weekly for 3 months F, K, S, Sat.
Henggeler et al. (2013) 2 N = 161 therapists Contingency Management for substance abuse 1 day 3: Workshop/Resources (WS+); WS+ computer-assisted training (CAT); WS+/ CAT+ supervisory support 1 year Biweekly review of barriers to implementation; duration not reported F, KP
Leathers, Spielfogel, et al. (2016) 2 N = 57 providers Keeping Foster Parents Trained and Supported 16 hr over 4 sessions 2: TAU versus. Enhanced Training (contact with change agent) 5 time points over 14-month period TAU: As needed; Enhanced: change agent offered 6 mo. of informal and formal consultation P
Luoma and Vilardaga (2013) 2 N =22 mental health
professionals
Acceptance and Commitment Therapy 2 days 2; training alone; training + 6 sessions phone consultation 3 months 6 30-min phone consultations K, PF, Sat.
Accurso, Astrachan- Fletcher, O’Brien, McClanahan, and Le Grange (2018) 3 N =7 therapists; N = 11 youth participants Family-based treatment for anorexia nervosa (+4 DBT skills sessions) 2 days None None Weekly 1-hr group supervision (number of weeks not reported) A, F, Cl, T
Brookman-Frazee et al. (2012) 3 N= 13 therapist/family dyads Individualized Mental Health Intervention for Children with ASD 6 hr None None 1 hr twice per month for 5 months A, Cl, F, K, P, S, Sat.
Chard, Ricksecker, Healy, Karlin, and Resick (2012) 3 Varies depending on measure (range from N = 237 to ii = 1,107); N = 374 veteran patients Cognitive Processing Therapy 3 days None 3 years 6 months of weekly phone consultation with expert trainers A, C, Cl, I, Sat
Chu et al. (2015) 3 N = 23 mental health clinicians CBT for child anxiety and depression 6 hr None 3–5 years Weekly supervision (duration not reported) A, P
Dorsey et al. (2016) 3 N = 180 clinicians CBT for anxiety, PTSD, depression, and parent management training for behavior problems (“CBT+”) 3 days; optional annual 1-day booster training None Post-assessment was 6-months post-consultation; subsample completed measures 3-months post-consultation 6 weeks biweekly phone consultation; access to CBT+ listserv P. S
Gellis et al. (2014) 3 N = 16 practitioners Problem-solving therapy in community- based settings for depression 1 day None None Weekly 30 min phone-based consultation for 6 months; supervisors reviewed audiotaped sessions A, I (qual)
Hamblen et al. (2010) 3 N = 111 psychologists, LPCs, and social workers CBT for postdisaster distress 2 days None None Biweekly consultation (duration not reported) A, F, K
Leffler et al. (2013) 3 N = 28 professional clinicians and psychology trainees Multi-Family Psychoeducation Psychotherapy for mood disorders 4, 5, or 7 hr None None Received feedback while facilitating group; 30 min consultation after group C
Lewis and Simons (2011) 3 N = 24 therapists from community agencies CBT for depression 3 days None 8 months Every 3–5 weeks over 8 months A, I, P
Lu et al. (2012) 3 N =25 clinicians CBT for PTSD 2 days; 1-day booster 1 year post-training None None Weekly group supervision; submission of audiotaped sessions F, S, Sat
Lyon, Charlesworth- Attie, Stoep, and McCauley (2011) 3 N = 18 therapists Managing and Adapting Practice system for depression and anxiety 3 half days None None Biweekly, in-person, 90-min group consultation for academic year A, K, P
Lyon et al. (2015) 3 N = 11 therapists and supervisors CBT for children and families 3 days None 3 months 6 months of expert- provided biweekly 1-hr phone consultation A, P, S
Ngo et al. (2011) 3 N = 35 therapists CBT for depression in disaster-impacted areas 1–2 days None None 1-hr weekly open group conference call + 1-hr weekly individual phone consultation with review of audiotaped sessions (duration not reported) K, S, Sat.
Petry et al. (2012) 3 N = 15 clinicians; N = 78 patients Contingency management for substance abuse 2 half-day trainings None Post-supervision Weekly meetings and feedback on supervisors’ ratings of audiotaped sessions until competency reached Cl, F, S, K, Sat
Reese et al. (2016) 3 N = 161 clinicians CBT for OCD 3 days None None 3 30-min consultation calls I, P. S
Simons et al. (2010) 3 N = 12 community mental health therapists; N = 116 clients CBT for depression 2 days None 1 year 16 1-hr group phone consultation for 1 year Cl, s
van den Berg et al. (2016) 3 N = 16 therapists; N = 19 patients PE and EMDR for PTSD 4 days(2 theoretical, 2 technical) for each intervention None 2 years Monthly 4-hr group supervision for trial duration (−19 months) A, Cl

Abbreviations: A, attitudes; C, confidence (self-efficacy); CBT, cognitive behavioral therapy; Cl, clinical outcome; DBT, dialectical behavior therapy; E, eye movement desensitization and reprocessing; F, treatment fidelity or adherence; I, implementation difficulty or barrier—anticipated or actual; Int, intentions; K, knowledge; P, practices or techniques used; PE, prolonged exposure; PF, psychological flexibility; S, skills/competence; Sat, satisfaction/acceptability; T, therapeutic interaction/rapport/working alliance; TAU, training as usual.

3.3.2 |. Summary of findings

Findings indicated that workshops combined with consultation are a successful approach to increase self-reported intervention use, effect change in competence, and improve client outcomes. Results from these studies indicated improved training outcomes among participants who received consultation. For example, Beidas, Edmunds, et al. (2012) found that although all training modalities (routine, computer, and active learning) resulted in limited improvements in therapist adherence and skill, the more consultation calls attended, the higher participants performed on measures of adherence and skill at a three month follow-up. Another study included a comparison of therapists who did and did not receive consultation after an acceptance and commitment therapy workshop on psychological flexibility and burnout (Luoma & Vilardaga, 2013). Those who received consultation had higher psychological flexibility at three month follow-up than those who did not receive consultation. Reese et al. (2016) also reported that among therapists who attended a workshop on OCD treatment, using phone consultation and peer consultation after training was associated with more self-reported use of the intervention. Several studies found that intervention use decreased in the absence of consultation (i.e., when consultation ended or if consultation attendance was poor; Lyon, Dorsey, Pullmann, Silbaugh-Cowdin, & Berliner, 2015; Ngo, Centanni, Wong, Wennerstrom, & Miranda, 2011; Petry et al., 2012), although the presence of a change agent in the organization increased participants’ use of the intervention (Leathers, Spielfogel, Blakey, Christian, & Atkins, 2016) in the absence of formal consultation.

Only one study found a lack of improved training outcomes with ongoing supervision (Henggeler et al., 2013). This study included a comparison of three conditions: (a) workshop only, (b) workshop and access to computer-assisted training (CAT), and (c) workshop, access to CAT, and supervisory support and coaching, and found no differences across conditions. All conditions demonstrated comparable improvements in intervention use, knowledge, and adherence to contingency management for substance use. The authors suggest that the lack of condition differences may have been due to strong organizational support, high practitioner turnover (27% turnover during the study), or the fact that the support was primarily for supervisors (i.e., not direct consultation with therapists implementing the intervention).

Several studies examined outcomes with gold-standard measurements of fidelity (i.e., direct observation of videotaped sessions). Simons et al. (2010) trained therapists in cognitive behavioral therapy (CBT) for depression and found that therapists demonstrated improved EBI competence 6 and 12 months after training relative to baseline. Lu et al. (2012) used videotaped sessions to provide feedback on delivery of CBT for PTSD to therapists and found that 91% of therapists achieved competency with their first case. Brookman-Frazee et al. (2012) developed a training protocol following recommendations made by Herschell et al. (2010). Their training in an intervention for children with autism was interactive and included opportunities for practicing treatment planning during the training. It also included self-study, ongoing consultation, and feedback on delivery with an actual client. Similar to other studies, Brookman-Frazee et al. (2012) found that observer-rated therapist fidelity during consultation was high and that parent-reported child problem behaviors significantly decreased from pretreatment to follow-up. All three of these studies also examined client outcomes and found improvements in client symptoms after trained therapists delivered the EBI.

3.3.3 |. Limitations of studies reviewed

Despite generally positive findings, there was variability in the type and quality of outcomes measured, the type of training provided, and the study design. Several studies were limited by the lack of emphasis on sustainment and follow-up greater than one year. Many studies included only pre- and post-assessments, or had brief (e.g., 3 month) follow-ups. Studies with longer follow-ups had mixed results (e.g., Leathers, Spielfogel, et al., 2016; van den Berg et al., 2016). Sustainment of gains, especially after the discontinuation of consultation calls, merits further examination.

Measurement issues persist. Many studies used measures that have not been psychometrically evaluated. A notable concern is the number of studies that assessed fidelity and intervention use with only therapist self-reports. Several studies did use direct observation of sessions to assess fidelity, but many relied on therapist-report (e.g., Henggeler et al., 2013; Reese et al., 2016) or administrative claims data (Hamblen et al., 2010). Another used behavioral role plays as an analogue fidelity measure (Beidas, Edmunds, et al., 2012), which is preferred over self-report, but less preferred than observation of in-session behavior. Given limitations in self-reported fidelity measures (Hurlburt, Garland, Nguyen, & Brookman-Frazee, 2010), conclusions from studies that did not use direct observation of fidelity of intervention delivery should be made with caution.

Another concern is the variability in training, including the duration/dose of training and the workshop training content. The length of training ranged from 4 hours to 4 days. Only one study (Leffler, Jackson, West, McCarty, & Atkins, 2013) compared various workshop durations, so it is not known how much training time is sufficient to successfully transfer knowledge or to change practice. The decision for the workshop duration was typically not discussed nor empirically determined. A similar limitation is true for the dosage of consultation calls, where the length and frequency were rarely justified. No studies examined the ideal dosing or delivery of consultation, though this is under study (Monson et al., 2018; Stirman et al., 2013).

3.4 |. Online training with or without consultation

3.4.1 |. Description of studies

Of the 20 unique studies of online training (Table 3), six were Type 1, four were Type 2, and 10 were Type 3. Type 1 and Type 2 studies used between-subjects designs and compared two or more different training approaches. Most Type 3 studies did not include a comparison condition, with the exception of Mallonee, Phillips, Holloway, and Riggs (2018), which compared in-person training to online training. Three studies (Harned et al., 2014; Jones et al., 2015; Ruzek et al., 2014) had sub-studies (Edwards et al., 2017; Harned, Dimeff, Woodcock, & Contreras, 2013; Nowrouzi, Manassis, Jones, Bobinski, & Mushquash, 2015), which were excluded. Sample sizes ranged from 8 to 706 participants working primarily in community mental health and VA settings. Length of training also varied (2–20 hours). The most commonly assessed outcomes were knowledge (14 studies), followed by skill, use of the intervention, and satisfaction with training (11 studies each). Measures of knowledge typically lacked psychometrics. Measures of skill were mostly based on role plays with standardized clients (e.g., Dimeff et al., 2015) or observation of actual therapy sessions (e.g., Rakovshik, McManus, Vazquez-Montes, Muse, & Ougrin, 2016). Intervention use was generally measured by self-report, except for one study that used client report of therapist behavior (Gryglewicz, Chen, Romero, Karver, & Witmeier, 2016). Satisfaction with training was measured using a wide array of measures, some of which have psychometric support (e.g., Kobak, Craske, Rose, & Wolitsky-Taylor, 2013; Kobak, Lipsitz, Lipsitz, Markowitz, & Bleiberg, 2017; Kobak, Wolitzky-Taylor, Wolitzky-Taylor, Craske, & Rose, 2017), but many of which were designed for each study. Two studies examined client outcomes and did not have strict client inclusion or exclusion criteria.

TABLE 3.

Online training

Author Nathan &Gorman classification Sample size Training topic Amount of training Comparison groups Follow-up Role of consultation Outcome measure domains
Cooper et al. (2017) 1 N = 156 therapists CBT for eating disorders ~8–9 hr online training 2: independent training; support with nonspecialist worker 6 months Support Condition Only: 12 30-min phone calls over 20 weeks K, S
Dimeff et al. (2011) 1 N = 132 participants Distress Tolerance DBT module 2.5 hr (timed with proctors present) 3; self-study; e-learning course; placebo attention control e-learning course 2-, 7-, 11-, 15-week None C, K, P, Sat
Dimeff et al. (2015) 1 N = 172 license mental health professionals DBT Core Strategies Online: ~12 hr
In-Person: 12 hr
2; online; in-person 3 months None C, K, P, S, Sat
Ehrenreich-May et al. (2016) 1 N = 140 community clinicians CBT for anxiety/panic disorders 7 hr online training 3; text-alone; text + online training; text + online + learning community 3 months 8 weekly calls for learning community C, I, K, P, S, Sat
Harned et al. (2014) 1 N = 181 mental health providers & students in mental health field Exposure therapy for anxiety disorders ~10 hr online training 3; online training, online + ME; online + ME + online learning community 3 months Integrated into last 3 weeks of learning community A, K, P, S, Sat
Ruzek et al. (2014) 1 N = 168 clinicians CBT for PTSD ~4 hr online training 3; web-based training; web-based training + consultation; no-training control None 6 weekly 45–60 min group calls C, K, P, S
Bennett-Levy et al. (2012) 2 N = 49 participants CBT 12-weeks 2; online only; online + telephone/Skype support 4 weeks Biweekly 15-min support sessions over 12 weeks C, K, P, S
Harned et al. (2011) 2 N = 46 mental health providers Exposure therapy for anxiety disorders ~2 hr online training 3; online training; online training + MI; placebo control online training None None A, C, K, P, Sat
Rakovshik et al. (2016) 2 N = 61 clinicians CBT for anxiety 20-hr online CBT training program 3; internet-based training with consultation worksheet; internet + supervision; delayed control None 3 30-min individual supervision sessions once per month S
Stein et al. (2015) 2 N = 36 clinicians; N = 136 patients Interpersonal and social rhythm therapy for bipolar disorder Online: 12-hr asynchronous online program + learning collaborative; TAU: 12-hr in-person training 2; Online training; TAU (in- person training) Monthly follow-up for 12 months posttraining Online: Monthly 60 min group phone supervision for 3–6 months after training; TAU: onsite supervision as usual; able to contact experts P
Fairburn, Allen, Bailey- Straebler, O’Connor, and Cooper (2017) 3 N = 102 therapists CBT for eating disorders 9 hr minimum online training None None Up to 12 30-min calls S
Gryglewicz et al. (2016) 3 N = 178 participants QPRT suicide risk assessment and management training 8–12 hr (within 4 weeks) None None None A, C, K, Sat
Jones et al. (2015) 3 N = 78 therapists; N =71 youth CBT for anxious youth 20-session, weekly group supervision training model with didactic component None None 20-session weekly group seminar included didactics and supervision Cl, K
Kobak, Lipsitz, et al. (2017) 3 N =26 clinicians Interpersonal therapy for depression 3–4 hr web-based tutorial None 1–3 months after applied training 45–60 min live remote training (role play) via videoconference with feedback; web-based training portal allowed option to request case consultation K, Sat
Kobak, Wolitzky- Taylor, et al. (2017) 3 N = 10 community clinicians; N= 33 patients CBT for anxiety disorders Online training time not reported None Post-live applied training 4 60-min remote live training sessions with feedback Cl, K, S, Sat
Kobak et al. (2013) 3 N = 39 social workers, psychologists, and graduate students CBT for anxiety disorders ~5.5 hr online training; 3 hr live feedback None None None K, S, Sat
Mallonee et al. (2018) 3 N = 706 mental health professionals (pre) N = 780 (post) CBT for PTSD, depression, insomnia, pain, suicidality Online training time not reported; 2 days in person 2; in-person; online ”3D” training None None K, Sat.
Martin, Gladstone, Diehl, and Beardslee (2016) 3 N = 58 clinicians Family Talk prevention intervention for depression 4-hr web- based + 3.5 hr face-to-face None 4-months None P, Sat
Persons et al. (2016) 3 N =26 clinicians Progress Monitoring 60-min orientation + 4 1.5 hr online classes None 1 year Listserv P
Puspitasari, Kanter, Murphy, Crowe, and Koerner (2013) 3 Study 1: N = 8 participants; Study 2: N =9 participants Behavioral Activation Not specified (“self-paced”) None Study 2 only: 6-week follow-up Studies 1 + 2:3 live 90-min online training sessions Study 1: P, Sat; Study 2: C, P, S, Sat

Abbreviations: A, attitudes; C, confidence (self-efficacy); CBT, cognitive behavioral therapy; Cl, clinical outcome; DBT, dialectical behavior therapy; F, treatment fidelity or adherence; I, implementation difficulty or barrier—anticipated or actual; Int, intentions; K, knowledge; MI, motivational interviewing; P, practices or techniques used; PF, psychological flexibility; QPRT, Question, Persuade, Refer, Treat; S, skills/competence; Sat, satisfaction/acceptability; T, therapeutic interaction/rapport/working alliance; TAU, training as usual.

3.4.2 |. Summary of findings

Similar to findings regarding in-person workshops alone, there was clear evidence that online trainings can improve therapist knowledge and skill in the short-term. Dimeff, Woodcock, Harned, and Beadnell (2011) compared training in the dialectical behavior therapy (DBT) distress tolerance module across conditions: online training, self-study of treatment manual, and placebo attention control e-learning. The online training resulted in the greatest improvements in therapist-reported use of the intervention, but the manual and online training resulted in similar increases in knowledge and self-efficacy. Other less rigorous studies also demonstrated an increase in CBT knowledge, skill, and use of the intervention after online training (Kobak et al., 2013; Kobak, Lipsitz, et al., 2017; Kobak, Wolitzky-Taylor, et al., 2017).

Three studies directly compared in-person and online trainings. Stein et al. (2015) found comparable improvements in use of interpersonal and social rhythm therapy for bipolar disorder for an asynchronous online and an in-person training. Similarly, Mallonee et al. (2018) compared in-person training to live “3-D” (i.e., involving interactive virtual worlds) synchronous training for a variety of EBIs. They found that participants in both conditions had significant and comparable gains in knowledge and readiness to implement EBIs, suggesting that online and in-person training may be equally effective. However, other findings are in contrast: Dimeff et al. (2015) compared online training, in-person training, and self-study of a DBT manual and found that knowledge was highest for participants who received online training, although there were no differences for observer-rated proficiency by condition. This study also found that instructor-led training was better than online training at improving motivation to use a more complex strategy (chain analysis) than a simpler skill (validation). This suggests that there may need to be a match between training and treatment strategies, such that more complex treatment strategies may benefit from being taught in person. Consistent with this, Persons, Koerner, Eidelman, Thomas, and Liu (2016) reported that a more straightforward EBI (i.e., measurement-based care; using quantitative measures to monitor symptom change) improved significantly from online training alone.

Two studies examined the role of consultation following online training. Ruzek et al. (2014) conducted an RCT of web-based training approaches for PTSD comparing webbased training and consultation, web-based training alone, and a no-training control condition. They found that participants who received consultation demonstrated greater skill on a standardized client assessment for two of the three skills relative to the other conditions. There were no condition differences on the third skill or on self-reported implementation of the strategies taught in the training, which improved for all conditions. Rakovshik et al. (2016) conducted a similar RCT in which they compared internet-based training with a consultation worksheet, internet-based training with CBT supervision, and a delayed-training control condition. The consultation worksheet asked questions about CBT approaches for the case and whether the therapist had the skills to implement the planned interventions. Those in the consultation worksheet condition turned in these self-report worksheets, but did not discuss them with a supervisor, whereas those in the supervision condition received supervision based on this worksheet. Similar to Ruzek et al. (2014), participants who received supervision had better observer-rated CBT competence than participants in both other conditions. Consistent with in-person training, online training alone does not seem sufficient to improve CBT competence in practice, but consultation improves outcomes.

Other studies examined additional enhancements to webbased trainings. Cooper et al. (2017) conducted an RCT of web-based training for CBT for eating disorders. In one condition, participants completed the training independently and in the other condition they received support by a nonspecialist worker who encouraged the participants to complete all aspects of the online program. There were no condition differences on competence in the intervention following training. Similarly, Bennett-Levy, Hawkins, Perry, Cromarty, and Mills (2012) provided brief (15-minute) post-training support sessions for a subset of therapists receiving training and also did not find condition differences in confidence, knowledge, skills, or self-reported use of the intervention. However, those who received the support sessions were significantly more likely to complete all elements of the training program. Another study trained therapists in CBT for anxiety and panic (Ehrenreich-May et al., 2016). Participants were randomized to manual-alone, online training and manual, or online training, manual, and a learning community that incorporated weekly conference calls and online discussions via Twitter. Participants who were engaged in the learning community had the highest clinical proficiency ratings based on role play performance, but this effect was not maintained at two and three month follow-ups. Harned et al. (2014) conducted an RCT that incorporated a web-based learning community with online training and motivational enhancement. Other conditions included online training alone and online training plus a motivational enhancement. Results indicated that all conditions had increased self-efficacy and self-reported use of exposure, but the learning community condition had the highest knowledge, better attitudes, and higher clinical proficiency. The motivational enhancement component did not improve attitudes relative to the online training alone condition in this study, which is in contrast to a previous similar study conducted by Harned, Dimeff, Woodcock, and Skutch (2011), that provided the motivational enhancement in person rather than online. Taken together, studies examining enhancements to web-based trainings found that brief support after training improved compliance to training requirements but did not have a substantive effect on key training outcomes. Online learning communities may improve outcomes, but sustainability of these gains merits further examination.

3.4.3 |. Limitations of studies reviewed

Findings indicated that online training can be an effective method for training, but there were variants of online training and questions about the ideal approach to online training. Online training had more Type 1 and Type 2 studies than any other category. Many Type 3 studies were limited by only using within-subject designs. Only three studies made direct comparisons between online training and in-person training (Dimeff et al., 2011; Mallonee et al., 2018; Stein et al., 2015). All studies examining online training demonstrated some positive effects. However, additional work is needed to determine strategies to continue improving online trainings, especially given that satisfaction for in-person trainings may be rated more highly (e.g., Mallonee et al., 2018).

Regarding measurement, several studies used standardized client role plays to measure clinical skill (e.g., Rakovshik et al., 2013; Ruzek et al., 2014), and one study used observation of therapy sessions (Rakovshik et al., 2016). Kobak and colleagues (2013; Kobak, Lipsitz, et al., 2017; Kobak, Wolitzky-Taylor, et al., 2017) used standardized clients as a tool for training in addition to using it as a method for measuring skill. Although measures of clinical skill were relatively strong in several studies, measures of clinical use were generally limited by reliance on therapist self-report.

Another limitation was short follow-ups. The longest follow-up was one year, but the majority were three months, and several studies had no follow-up. Some studies that did include follow-ups found that some gains were not maintained (e.g., Ehrenreich-May et al., 2016; Persons et al., 2016). In addition, the length of training varied widely, making direct comparison of studies difficult and leaving unanswered questions about optimal length of training.

To address these limitations, future studies should compare online training with consultation to in-person training with consultation, include longer follow-ups, and compare synchronous (trainers and trainees move through training at the same time on a set schedule) to asynchronous (self-paced) training. In addition, online trainings often incorporated interactive components, such as completing virtual role plays with other attendees as an adjunct to training. However, there has not been empirical study of what components and teaching strategies for online training are most effective and likely to enhance the probability of therapist behavior change.

3.5 |. Train-the-trainer

3.5.1 |. Description of studies

There were four TT studies involving training supervisors or select therapists who in turn trained other staff (Table 4). These studies involved expert-led group training sessions for trainees to be eligible to train other participants. The first study (Martino et al., 2010) was Type 1 and randomized 12 outpatient programs to one of three conditions for receiving training in motivational interviewing (MI). The other studies were Type 3. Nakamura et al. (2014) used a within-subject multiple baseline design and included four supervisor–therapist dyads. Southam-Gerow et al. (2014) trained therapists in an evidence-informed model of care, Managing and Adapting Practice (MAP). This study did not randomize therapists to conditions but described outcomes for therapists who were trained by expert trainers, as well as for those trained via a TT model. The final study (Wade et al., 2014) examined a TT approach for a post-disaster mental health training program.

TABLE 4.

Train-the-trainer

Author Nathan & Gorman classification Sample size Training topic Amount of training Comparison groups Follow-up Role of consultation Outcome measure domains
Martino et al. (2010) 1 N = 92 therapists Motivational interviewing for substance use problems Self-study: 20 hr; Expert-led: 15-hr; Train-the-trainer: 15 hr + additional 15 hr for trainers 3; Self-study; Expert-led; Train-the-trainer Postsupervision and 12-week follow-up Expert: 3 monthly individual face- to-face sessions to review audiotaped sessions; Train- the-trainer: Expert calls + monthly consultation calls for trainers (duration not reported) F, S, Sat
Nakamura et al. (2014) 3 N = 4 supervisors
N = 4 therapists
Modular Approach to Therapy for Children with Anxiety, Depression, Trauma, or Conduct Problems Supen’isors: three 2-day workshops None None None S
Southam- Gerow et al. (2014) 3 N = 504 for national training pathway;
N = 283 for agency supervisor pathway (train-the-trainer);
N= 1,172 youth clients
Managing and Adapting Practices (MAP) 40 hr; Supen’isor training: additional 16 hr of training None None 6 months of biweekly 1-hr phone consultation and submission of materials for 2 cases; Supen’isor training: additional 6 months of 1-hr consultation calls Cl, F, P
Wade et al. (2014) 3 N = 40 trainers;
N = 684 practitioners
Skills for Psychological Recovery Trainers: 2-days + competency assessment; Practitioners: 1 day;
Both groups: online booster modules at 3- and 6-months
None 3- and 6-month post-training Trainers: 1-hr teleconference at 3- and 6-months to receive advice, support, and feedback; feedback on actual training sessions for a subset of trainers C, I, P; Trainers Only: S

Abbreviations: A, attitudes; C, confidence (self-efficacy); Cl, clinical outcome; F, treatment fidelity or adherence; I, implementation difficulty or barrier—anticipated or actual; Int, intentions; K, knowledge; P, practices or techniques used; PF, psychological flexibility; S, skills/competence; Sat, satisfaction/acceptability; T, therapeutic interaction/rapport/working alliance.

Sample sizes ranged from four supervisor–therapist dyads to 684 practitioners working in community mental health. The length of training ranged from 1 day to 56 hours. Three out of four studies examined therapist and/or trainer competence. Measures of competence included observer ratings using standardized scales (Martino et al., 2010; Nakamura et al., 2014). One study had practitioners use a (non-standardized) measure to assess trainers’ competence (Wade et al., 2014). The other most frequent outcomes were treatment fidelity and EBI use, measured in two studies. Fidelity was measured using observer ratings of submitted session recordings (Martino et al., 2010; Southam-Gerow et al., 2014). Finally, EBI use was measured via therapist-report (Wade et al., 2014) and claims data (Southam-Gerow et al., 2014). Southam-Gerow et al. (2014) measured client outcomes and did not have client inclusion or exclusion criteria.

3.5.2 |. Summary of findings

Martino et al. (2010) conducted the most rigorous of the TT studies and compared self-study, expert-led training, and TT conditions. They found improvements in therapists’ MI performance for the expert and TT conditions, such that a greater percentage performed MI adequately compared to those in the self-study condition. However, contrary to expectations, participants in the expert-led condition achieved higher standards of MI performance than those in the TT condition at 12-week follow-up. The authors suggest that this may be because trainers for the TT condition had no prior MI training and were rated as covering MI material less extensively and skillfully than experts during the workshops they conducted. Southam-Gerow et al. (2014) similarly found that participants who received training from expert trainers had higher proficiency ratings than those trained by trainers in the TT condition. However, overall ratings of therapists’ portfolios were in the proficient range, and time to successful completion of training was comparable for both types of training. These findings highlight the promise of the TT model to achieve positive therapist training outcomes, though not at the level of expert-led trainings.

The other two TT studies also suggest positive results, although methodological limitations preclude firm conclusions. Nakamura et al. (2014) completed observational coding of four therapist–supervisor dyads and found that supervisors improved in their teaching (i.e., style; content) and therapists improved in their EBI performance. Similarly, Wade et al. (2014) found that participants met minimum trainer-rated competency standards to conduct trainings and were rated by other therapists as having high competence in delivering training. Furthermore, therapists who received training from those participants had increases in confidence in using the intervention.

3.5.3 |. Limitations of studies reviewed

Herschell et al.’s (2010) prior review noted that, although potentially cost-efficient, there were limited data to support the use of the TT method. The four studies conducted since the 2010 review provide additional evidence for its support, but gaps remain. Only one study was Type 1 (Martino et al., 2010); the others did not randomize participants, limiting the ability to compare TT to other training approaches. Further, the studies conducted by Nakamura et al. (2014) and Martino et al. (2010) were conducted within a small number of agencies, which limits generalizability.

Other limitations are similar to those noted for other training approaches. For example, follow-up length was absent (Nakamura et al., 2014) or limited (longest was 6 months; Wade et al., 2014); thus, sustainability of this approach is unknown. Client outcomes were only measured in one study (Southam-Gerow et al., 2014) and only represented a subset of cases. Client outcomes may be particularly important in a TT model, where the knowledge passed on to therapists may not be as comprehensive as the training provided to supervisors, which could in turn affect client outcomes. The specific amount and content of consultation provided to supervisors and trainees warrants further examination, as does fidelity to training and consultation by TT trainers.

3.6 |. Intensive training

3.6.1 |. Description of studies

Twenty-two studies evaluated intensive training, which was defined as including at least 20 hours of training, as well as two or more additional training components (Table 5). Among studies of intensive training, only one was Type 1 (Kolko et al., 2012). Three were Type 2, of which two were part of the Beck Initiative. The Beck Initiative studies (German et al., 2017; Stirman et al., 2017) involved condition comparisons with sufficient sample sizes for planned analyses, but neither included randomization to a condition. Webster-Stratton, Reid, and Marsenich’s (2014) randomized trial was considered Type 2 due to lack of clear inclusion and exclusion criteria. The remaining 18 studies were Type 3. All except one of the Type 3 studies (Smith et al., 2017) did not have comparison conditions. Most were uncontrolled within-subjects designs, but many included follow-ups of 2 years after initial training.

Sample sizes ranged from 5 to 1,034 therapists practicing in community mental health, school-based, and VA settings. Intensive training models ranged from the minimum of 20 training hours for inclusion in this category to up to 40 training hours prior to consultation or booster training. In terms of outcome measures, therapist skill was most frequently measured (16 of 22 studies). The Beck Initiative studies all included measures of therapist skill based on observations of therapy session recordings, as did a subset of VA studies (Smith et al., 2017 (study 2); Karlin et al., 2012; Karlin, Trockel, Taylor, Gimeno, & Manber, 2013; Walser, Karlin, Trockel, Mazina, & Barr Taylor, 2013). Several other studies also used observer-rated measures of therapist skill. Confidence using the intervention was the second most frequently measured outcome (13 studies), followed by attitudes (9 studies). In most studies, confidence was measured with questions developed for the study. Attitudes were measured with psychometrically supported measures (the EBPAS) in three studies (Karlin et al., 2012; Kolko et al., 2012; Navarro-Haro et al., 2019) but were measured using study-specific measures for the others. Studies that measured client outcomes (n = 6) had no or limited client exclusion criteria.

Besides consultation, additional components of intensive trainings included homework between training sessions, therapy session tape review with feedback, role play feedback, advanced booster trainings, and a learning collaborative. Studies were part of several different initiatives, including statewide efforts to implement parent–child interaction therapy (PCIT; Beveridge et al., 2015; Jackson, Herschell, Schaffner, Turiano, & McNeil, 2017), examination of adoption of DBT in community clinics (Herschell, Lindhiem, Kogan, Celedonia, & Stein, 2014; Navarro-Haro et al., 2019), VA initiatives for training in EBIs (Karlin et al., 2012, 2013; Manber et al., 2013; Ruzek et al., 2017, 2016; Smith et al., 2017; Walser et al., 2013), and the Beck Initiative for training in Cognitive Therapy (as described in Creed et al., 2013; Creed, Stirman, & Evans, 2014; Creed, Wolk, Feinberg, Evans, & Beck, 2016; German et al., 2017; Stirman et al., 2017).

3.6.2 |. Summary of findings

Studies of intensive training approaches demonstrated evidence for improvements in therapist knowledge, intervention use, and observer-rated competence. All except one study (Navarro-Haro et al., 2019) included formal consultation and many included expert feedback on recorded therapy sessions. Webster-Stratton et al. (2014) compared training conditions with and without consultation. Those who received ongoing consultation and feedback on recorded sessions demonstrated higher fidelity in several domains, suggesting that consultation enhanced adherence. However, fidelity in some areas, such as conducting role plays, was still limited even in the presence of phone consultation. This suggests that phone consultation was insufficient for addressing some of the challenging skills taught as part of the intervention.

Consultation appears to be a key element of intensive trainings. VA training initiatives led to improvements in competency after consultation based on observer-rated sessions (e.g., Karlin et al., 2012, 2013; Walser et al., 2013). Similarly, the Beck Initiative included consultation calls and examined outcomes across multiple settings (i.e., schools; Creed et al., 2013; community settings; Creed et al., 2016). These studies found that there was a significant increase in therapists’ competency in delivering cognitive therapy at 6-month follow-up. Furthermore, over 70% of therapists achieved the competency threshold by this final assessment. Stirman et al. (2017) compared two consultation call formats for participants in the Beck Initiative training. One condition received individual feedback from expert consultants who reviewed full audio-recorded therapy sessions. The other condition received group consultation in which short segments of tape were reviewed in group consultation with the expert consultant. At the end of consultation, the majority of therapists in both groups achieved the required level of competence in cognitive therapy. At 2-year follow-up, participants in the individual feedback condition who had high competence scores at post-consultation had a decrease in competence, but those in the group feedback condition had increases in competence. This provides support for group consultation, which reduces the amount of expert consultation time required. However, nearly half of therapists were ineligible for recertification at the 2-year follow-up, largely due to turnover. Herschell et al. (2014) also reported low rates of DBT training completion (55%) due to turnover.

Intensive trainings tend to be cost- and resource-intensive because of the amount of expert time required. For example, Kolko et al. (2012) randomized participants to either a learning community model (i.e., workshops and consultation) or training as usual (TAU) per agency standards. The former condition required more resources (i.e., expert time), with therapists attending an average of 22.8 hours (SD = 9.1) of workshop training and 9.8 hours (SD = 5.2) of consultation. Therapists in the active condition demonstrated greater improvements in knowledge and self-reported EBI use 6 months after training compared to therapists receiving TAU. These gains were generally maintained at 12- and 18-month follow-ups. German et al. (2017) examined a modification to the Beck Initiative training model that aimed to reduce the need for expert time and resources. They compared in-person workshops with expert-led consultation to web-based training followed by peer-led consultation. Both conditions demonstrated comparable knowledge gains and improvements in therapy competence. However, those who received the web-based training with peer-led consultation were significantly less likely to complete the training than those who received in-person training. Despite this limitation, the model is a promising approach given that web-based training required only 8% of the resources of the in-person training. Furthermore, there are strategies that may improve engagement and completion of a web-based training (e.g., training with a cohort of trainees, pay increases for EBI certification).

3.6.3 |. Limitations of studies reviewed

As in other categories, most intensive studies were within-subject designs without comparison groups. However, these studies are notable for the number of studies with follow-ups, with many having at least a 6-month or 1-year follow-up. This extended contact involved booster training, consultation, and/or tape review. That said, studies of sustainability would benefit from longer follow-ups, as a minimum of 2 years typically represents the duration needed to determine EBI sustainability (Jackson et al., 2017; Stirman et al., 2017).

Consistent with other training methods, limitations related to outcome measurement persist. Studies relied exclusively on self-reports (e.g., Herschell et al., 2014) and used measures that lacked adequate psychometrics (e.g., Jackson et al., 2017). Several studies addressed previous limitations by using observational measures of therapist competence (e.g., the Cognitive Therapy Ratings Scale; CTRS: Young & Beck, 1980), but the lack of their widespread use may be due to these measures being time-intensive and having limited feasibility in low-resourced settings. In addition, there may be some conflation in the literature due to variability in quality of outcome measurement. Specifically, intensive training models have the most rigorous measurement, whereas less intense models tend to rely on self-report, which may inflate outcomes. Thus, there may be even greater differences between intensive and non-intensive training approaches.

The variability in the timing and length of training merits attention. Evaluated approaches were consistent in being intensive, but the configurations for the timing of the multiday workshops varied. For example, Kolko et al. (2012) had four full-day workshops once a week for a month, whereas Navarro-Haro et al. (2019) had two separate blocks of 5-day workshops approximately 6 months apart. Other intensive trainings (e.g., Jackson et al., 2017) had an initial 40 hours of training over several days and an advanced booster training 6 months later. The Beck Initiative studies (Creed et al., 2013, 2016; German et al., 2017; Stirman et al., 2017) did not space out the trainings but included substantial consultation and feedback that was spread out over time. Given that massed and spaced trainings may differ, there is need for research on how spacing of training may enhance training outcomes and/or improve consolidation of learning.

4 |. DISCUSSION

This review of therapist training studies confirms and extends findings from previous reviews (Beidas & Kendall, 2010; Herschell et al., 2010; Rakovshik & McManus, 2010). Therapist training in EBIs has been a growing area of research since 2010, with 76 studies added to the literature. Most studies examined intensive training, in-person workshops with consultation, or online workshops. Only four studies examined TT methods since the last review, and nine studies examined in-person workshops only. Consistent with conclusions from prior reviews, therapist knowledge and attitudes toward EBIs improve after attending a workshop; yet workshops alone are unlikely to increase EBI use. There remains limited evidence that those seeking to change therapist behavior should rely solely on manuals and workshops without additional consultation or other post-training activities. An extension to prior work is that this appears to be true regardless of whether training is conducted in person or online. Importantly, this review suggests that training is more likely to translate into increased EBI use and improved fidelity if therapists are supported by consultation after training, although major questions about optimal dosage and content remain.

This review notably expands upon previous work in the areas of online training, intensive training, and TT approaches. Recent studies suggest that more straightforward EBIs (i.e., those involving one component, such as measurement-based care) can be taught with comparable outcomes using online or in-person trainings, but more complex EBIs (i.e., involving several components or more nuanced application, such as CBT; Wolk et al., 2019) might benefit from being taught in person. Additional work is needed to identify which EBIs can be taught effectively online. Intensive training appears to have the most promise for increasing competence and use of EBIs, including more complex ones. However, limitations to intensive trainings include the amount of resources needed, uncertainty about sustainability, and investment in staff who may turnover (Beidas et al., 2016; Okamura et al., 2018). The TT model demonstrates the most promise with respect to sustainability, but raises important questions about the benefits of sustaining “watered down” EBIs (Demchak & Browder, 1990) if high fidelity is not achieved, as client outcomes were rarely examined in these studies. Without concomitant studies examining clinical outcomes, the merit of TT (and other intensive training) models will remain unclear.

Studies included in this review also included assessment of therapist variables (e.g., characteristics of the trainee therapists). There were mixed findings, such as whether having previous experience with the intervention leads to more or less change in knowledge and attitudes after training. However, among therapists who attended workshops followed by consultation, more positive attitudes toward EBIs generally predicted better training outcomes (e.g., adherence, skill, use of the intervention). A recent meta-analysis concluded that therapist experience has only a modest impact on client outcomes (Walsh, Roddy, Scott, Lewis, & Jensen-Doss, 2018), but there are other potential therapist variables for evaluation. For instance, there is evidence that therapist clinical practices (i.e., use of cognitive, behavioral, family, and psychodynamic techniques) are related to therapist background characteristics (e.g., type of graduate training; Becker-Haimes, Lushin, Creed, & Beidas, 2019). Some studies have begun to examine therapist constructs such as psychological flexibility (e.g., Richards et al., 2011), which may affect openness to learning and therapists’ ability to adopt new methods. Assessing a broad range of therapist constructs might allow for better tailoring of training methods to learner needs. As measurement across training studies becomes more standardized and meta-analysis becomes feasible, therapist variables should be examined as moderators.

Few studies delineated how formal learning theory and behavioral change principles guided the design of the training packages, despite the fact that there are established principles for facilitating both knowledge acquisition and individual behavior change. Application of established behavior change principles (e.g., those that are used in EBIs) should be considered in efforts to change therapist behaviors through training (e.g., Michie, Stralen, & West, 2011). Such principles could also guide the selection of variables most important to measure in the context of training efforts. For example, consistent with learning theory, evaluating procedural knowledge as a training outcome is likely a critical variable proximal to therapist behavior change (Bennett-Levy, McManus, Westling, & Fennell, 2009), yet was only studied relatively infrequently in the reviewed literature. Furthermore, few studies in this review were guided by implementation science frameworks, which highlight the multiple contextual factors that may affect therapist EBI use (e.g., Damschroder et al., 2009). Some studies within the present review identified ways in which organizational factors, such as having a champion within the organization, related to implementation and training outcomes (Dorsey, Berliner, Lyon, Pullmann, & Murray, 2016; Ngo et al., 2011). However, because most implementation science frameworks such as the CFIR are not causal and do not delineate mechanisms of action, work within the broader field of implementation science is needed to identify mechanisms through learning theories and theories of behavior change (Lewis et al., 2015). Future training studies should seek to identify the mechanisms through which contextual factors, such as organizational support or leadership, interact with training packages to affect therapist EBI adoption.

Some of the limitations and future directions noted in the previous reviews were addressed in recent studies. One earlier limitation was that few studies measured client outcomes, even though improving client outcomes is the ultimate goal. Several studies in this review specifically examined client outcomes using standardized measures with known psychometric properties. Relative to the 2010 reviews, there were also improvements in therapist competence measurement, as many studies used direct observation of therapy sessions, which were assessed using standardized measures (e.g., the CTRS). The use of observation rather than therapist self-report of competence is to be applauded and encouraged for its methodological advantages, especially given evidence that therapist self-report of skill tends to be inflated (e.g., Nakamura et al., 2014). Another previously noted limitation that has been addressed is the direct comparison of different training methods (e.g., online vs. in-person training; Dimeff et al., 2015; Mallonee et al., 2018; Stein et al., 2015), many of which were shown to be comparably effective.

Despite advances, several limitations persist. First, there are issues related to measurement. Many studies used measures developed for their specific study, preventing direct comparison across studies and reducing confidence in the interpretations. Of note, measurement development in implementation science is in its infancy (Lewis et al., 2015) and most implementation studies rely on measures with limited psychometric properties out of necessity. However, sound measurement is essential; the development of consistent measures to assess attitudes, knowledge, skill, and related constructs across multiple EBIs clearly merits future research. Second, many studies relied solely on therapists’ own reports of outcomes, which is especially troublesome for constructs known to be unreliable when assessed with self-report (e.g., intervention use; Hurlburt et al., 2010). To their credit, several studies addressed this limitation by using role plays and standardized clients. Future work should continue to evaluate therapist behavior with as-close-to-real behavior as possible. Third, sample sizes varied widely, ranging from four therapists to thousands. Many studies were underpowered, and thus, may have found fewer condition differences than would have been found with larger samples. This limitation was addressed in part in the present review by giving greater weight to Type 1 studies, which by definition conducted analyses that were sufficiently powered. A fourth limitation is that there exists a lack of clarity regarding what constitutes a workshop versus a course or ongoing consultation. It can be difficult to evaluate the outcomes of “training” when there is such wide variability in how the training condition is defined, as well as in the length of training times.

A final and persistent limitation in the training literature is the lack of clarity regarding the expected outcome of training. Training is effective at impacting therapists’ declarative knowledge about EBIs, but knowledge gain is not sufficient when the goal is to favorably impact client outcomes via therapist fidelity to EBIs. Even the question of fidelity to EBI merits consideration, and the degree of fidelity versus adaptation (Stirman et al., 2015) required to “successfully” deliver an intervention in community settings needs to be studied. Chu et al. (2015) found that several years after receiving training as part of an effectiveness trial, therapists reported using only part of the original intervention (i.e., low fidelity). Is it better for therapists to use some elements of the intervention even if not the full program? Is it better to use an intervention with high fidelity and low frequency or with low fidelity and high frequency? These questions are often ignored when training therapists, but they are considerations to weigh when designing a training and may vary depending on the specific EBI. Some trainers have created a process for certification of therapists following training to set benchmarks for fidelity (e.g., Creed et al., 2014), but research is needed to examine whether certification is an ideal measure of long-term therapist competency.

4.1 |. Recommendations and future directions

Findings from the current review suggest the following questions for future research:

  1. How should training outcomes be measured? Efforts are needed to improve measures of fidelity, skill, and use of the intervention. More studies need to assess training outcomes using observations and coding of actual sessions. In addition, accurate and, if possible, briefer measures of therapist fidelity and skill should be developed (Beidas et al., 2016; Herschell et al., 2014).

  2. What is the long-term sustainability of different training approaches? For example, comparisons may include online versus TT, 1 day versus 5 days, and variations to the length/frequency/modality of consultation calls.

  3. What specific strategies should be used during training? Many studies lack a detailed description of the training processes and content, but training strategies may have a meaningful impact on participant engagement, motivation, and learning. For example, studies mention using “experiential” training, but do not clarify what that means or how such a strategy differs from what others refer to as active learning strategies. In addition, there are no empirical guidelines for selecting the dose (length and frequency) of training and consultation, and this merits further examination as well.

  4. What training approaches are the most cost-effective? Such work requires taking into account expert time, measuring fidelity and client outcomes, staff turnover, and sustainability. Although it is likely that how well a training “works” may be highly correlated with how much it costs, research is needed to guide future training activities.

5 |. CONCLUSIONS

Providing optimal treatment to those seeking mental health treatments is likely a universal goal. Improving EBI access is an important step toward attaining that goal. Although the literature offers modest agreements about effective EBIs, questions remain about how best to train therapists in EBI use. Consistent with prior reviews, this updated synthesis suggests that manuals and brief workshops in isolation are likely insufficient to lead to therapist behavior change. Training is more effective at improving competence and intervention use when it is followed by consultation. More research is needed on identifying the critical components of “intensive” trainings, which showed improvements in therapist competence, but are cost- and resource-intensive. Approaches that may help to expand the reach and cost-effectiveness of therapist training include online training and TT models. Future work should focus on improving training outcome measurement, addressing issues related to sustainability, identifying key elements of training, and improving cost-effectiveness.

Public Health Significance.

Therapist training plays a critical role in the implementation of evidence-based interventions. This review provides an update of the past decade of literature related to therapist training in evidencebased interventions and examines five approaches to therapist training: workshops alone, workshops followed by consultation, online training, train-thetrainer, and intensive training. Results suggest that more intensive training models are most likely to facilitate clinician behavior change, although optimal dosage and training content remains unclear. Future work is needed to identify key elements of training to optimize the success of implementation efforts and ensure that treatment-seeking individuals receive effective care

Funding information

National Institute of Mental Health, Grant/Award Number: F31MH112211

Footnotes

1

The term EBI is used rather than evidence-based practice (EBP), as EBP has multiple definitions, and EBI is a more precise term that has begun to be used to describe treatments with empirical support (e.g., Aarons et al., 2016; Becker-Haimes et al., 2018).

2

As defined by Nadeem et al. (2013), consultation (sometimes referred to as “coaching” in education) is more often the appropriate term for the studies included in this review, as it refers to external support provided by experts on the intervention being disseminated and implemented. In contrast, supervision typically refers to internal support provided to therapists by their agency, clinic, or school district.

REFERENCES

  1. Aarons GA (2004). Mental health provider attitudes toward adoption of evidence-based practice: The Evidence-Based Practice Attitude Scale (EBPAS). Mental Health Services Research, 6(2), 61–74. 10.1023/b:mhsr.0000024351.12294.65 [DOI] [PMC free article] [PubMed] [Google Scholar]
  2. Aarons GA, Green AE, Trott E, Willging CE, Torres EM, Ehrhart MG, & Roesch SC (2016). The roles of system and organizational leadership in system-wide evidence-based intervention sustainment: A mixed-method study. Administration and Policy in Mental Health and Mental Health Services Research, 43(6), 991–1008. 10.1007/s10488-016-0751-4 [DOI] [PMC free article] [PubMed] [Google Scholar]
  3. Accurso EC, Astrachan-Fletcher E, O’Brien S, McClanahan SF, & Le Grange D (2018). Adaptation and implementation of family-based treatment enhanced with dialectical behavior therapy skills for anorexia nervosa in community-based specialist clinics. Eating Disorders, 26(2), 149–163. 10.1080/10640266.2017.1330319 [DOI] [PMC free article] [PubMed] [Google Scholar]
  4. Becker EM, Smith AM, & Jensen-Doss A (2013). Who’s using treatment manuals? A national survey of practicing therapists. Behaviour Research and Therapy, 51(10), 706–710. 10.1016/j.brat.2013.07.008 [DOI] [PubMed] [Google Scholar]
  5. Becker-Haimes EM, Lushin V, Creed TA, & Beidas RS (2019). Characterizing the heterogeneity of clinician practice use in community mental health using latent profile analysis. BMC Psychiatry, 19(1), 1–11. 10.1186/s12888-019-2234-0 [DOI] [PMC free article] [PubMed] [Google Scholar]
  6. Becker-Haimes EM, Okamura KH, Baldwin CD, Wahesh E, Schmidt C, & Beidas RS (2018). Understanding the landscape of behavioral health pre-service training to inform evidence-based intervention implementation. Psychiatric Services, 70(1), 68–70. 10.1176/appi.ps.201800220 [DOI] [PubMed] [Google Scholar]
  7. Beidas RS, Edmunds JM, Marcus SC, & Kendall PC (2012). Training and consultation to promote implementation of an empirically supported treatment: A randomized trial. Psychiatric Services, 63(7), 660–665. 10.1176/appi.ps.201100401 [DOI] [PMC free article] [PubMed] [Google Scholar]
  8. Beidas RS, & Kendall PC (2010). Training therapists in evidence-based practice: A critical review of studies from a systems-contextual perspective. Clinical Psychology: Science and Practice, 17(1), 1–30. 10.1111/j.1468-2850.2009.01187.x [DOI] [PMC free article] [PubMed] [Google Scholar]
  9. Beidas RS, Mychailyszyn MP, Edmunds JM, Khanna MS, Downey MM, & Kendall PC (2012). Training school mental health providers to deliver cognitive-behavioral therapy. School Mental Health, 4(4), 197–206. 10.1007/s12310-012-9074-0 [DOI] [PMC free article] [PubMed] [Google Scholar]
  10. Beidas RS, Stewart RE, Adams DR, Fernandez T, Lustbader S, Powell BJ, … Barg FK (2016). A multi-level examination of stakeholder perspectives of implementation of evidence-based practices in a large rrban publicly-funded mental health system. Administration and Policy in Mental Health and Mental Health Services Research, 43(6), 893–908. 10.1007/s10488-015-0705-2 [DOI] [PMC free article] [PubMed] [Google Scholar]
  11. Beidas RS, Williams NJ, Becker-Haimes EM, Aarons GA, Barg FK, Evans AC, … Mandell DS (2019). A repeated cross-sectional study of clinicians’ use of psychotherapy techniques during 5 years of a system-wide effort to implement evidence-based practices in Philadelphia. Implemention Science, 14(1), 67. 10.1186/s13012-019-0912-4 [DOI] [PMC free article] [PubMed] [Google Scholar]
  12. Bennett-Levy J, Hawkins R, Perry H, Cromarty P, & Mills J (2012). Online cognitive behavioural therapy training for therapists: Outcomes, acceptability, and impact of support. Australian Psychologist, 47(3), 174–182. 10.1111/j.1742-9544.2012.00089.x [DOI] [Google Scholar]
  13. Bennett-Levy J, McManus F, Westling BE, & Fennell M (2009). Acquiring and refining CBT skills and competencies: Which training methods are perceived to be most effective? Behavioral and Cognitive Psychotherapy, 37(5), 571–583. 10.1017/s1352465809990270 [DOI] [PubMed] [Google Scholar]
  14. Beveridge RM, Fowles TR, Masse JJ, McGoron L, Smith MA, Parrish BP, … Widdoes N (2015). State-wide dissemination and implementation of parent–child interaction therapy (PCIT): Application of theory. Children and Youth Services Review, 48, 38–48. 10.1016/j.childyouth.2014.11.013 [DOI] [Google Scholar]
  15. Brookman-Frazee LI, Drahota A, & Stadnick N (2012). Training community mental health therapists to deliver a package of evidence-based practice strategies for school-age children with autism spectrum disorders: A pilot study. Journal of Autism & Developmental Disorders, 42(8), 1651–1661. 10.1007/s10803-011-1406-7 [DOI] [PMC free article] [PubMed] [Google Scholar]
  16. Chard KM, Ricksecker EG, Healy ET, Karlin BE, & Resick PA (2012). Dissemination and experience with cognitive processing therapy. Journal of Rehabilitation and Research Development, 49(5), 667–678. 10.1682/JRRD.2011.10.0198 [DOI] [PubMed] [Google Scholar]
  17. Chin EG, Bernecker S, Buchanan EM, Cunningham S, Schumacher JA, & Coffey SF (2018). A novel dissemination effort for prolonged exposure: Practice and dissemination curriculum. Psychological Trauma: Theory, Research, Practice, and Policy, 11(3), 314–318. 10.1037/tra0000362 [DOI] [PubMed] [Google Scholar]
  18. Chu BC, Crocco ST, Arnold CC, Brown R, Southam-Gerow MA, & Weisz JR (2015). Sustained implementation of cognitive-behavioral therapy for youth anxiety and depression: Long-term effects of structured training and consultation on therapist practice in the field. Professional Psychology, Research and Practice, 46(1), 70–79. 10.1037/a0038000 [DOI] [PMC free article] [PubMed] [Google Scholar]
  19. Cooper Z, Bailey-Straebler S, Morgan KE, O’Connor ME, Caddy C, Hamadi L, & Fairburn CG (2017). Using the Internet to train therapists: Randomized comparison of two scalable methods. Journal of Medical Internet Research, 19(10). e355. 10.2196/jmir.8336 [DOI] [PMC free article] [PubMed] [Google Scholar]
  20. Creed T, Jager-Hyman S, Pontoski K, Feinberg B, Rosenberg Z, Evans A, … Beck AT (2013). The Beck initiative: Training school-based mental health staff in cognitive therapy. International Journal of Emotional Education, 5(2), 49–66. [Google Scholar]
  21. Creed T, Stirman S, & Evans A (2014). A model for implementation of cognitive therapy in community mental health: The Beck initiative. The Behavior Therapist, 37(3), 56–64. [Google Scholar]
  22. Creed T, Wolk CB, Feinberg B, Evans AC, & Beck AT (2016). Beyond the label: Relationship between community therapists’ self-report of a cognitive behavioral therapy orientation and observed skills. Administration and Policy in Mental Health and Mental Health Services Research, 43(1), 36–43. 10.1007/s10488-014-0618-5 [DOI] [PubMed] [Google Scholar]
  23. Crisanti AS, Murray-Krezan C, Karlin LS, Sutherland-Bruaw K, & Najavits LM (2016). Evaluation of an evidence-based practice training for peer support workers in behavioral health care. Cogent Psychology, 3(1). 10.1080/23311908.2016.1212453 [DOI] [Google Scholar]
  24. Damschroder LJ, Aron DC, Keith RE, Kirsh SR, Alexander JA, & Lowery JC (2009). Fostering implementation of health services research findings into practice: A consolidated framework for advancing implementation science. Implementation Science, 4(1), 50. 10.1186/1748-5908-4-50 [DOI] [PMC free article] [PubMed] [Google Scholar]
  25. Deacon BJ, Lickel JJ, Farrell NR, Kemp JJ, & Hipol LJ (2013). Therapist perceptions and delivery of interoceptive exposure for panic disorder. Journal of Anxiety Disorders, 27(2), 259–264. 10.1016/j.janxdis.2013.02.004 [DOI] [PubMed] [Google Scholar]
  26. Demchak M, & Browder DM (1990). An evaluation of the pyramid model of staff training in group homes for adults with severe handicaps. Education and Training in Mental Retardation, 25(2), 150–163. [Google Scholar]
  27. Dimeff LA, Harned MS, Woodcock EA, Skutch JM, Koerner K, & Linehan MM (2015). Investigating bang for your training buck: A randomized controlled trial comparing three methods of training clinicians in two core strategies of dialectical behavior therapy. Behavior Therapy, 46(3), 283–295. 10.1016/j.beth.2015.01.001 [DOI] [PubMed] [Google Scholar]
  28. Dimeff LA, Koerner K, Woodcock EA, Beadnell B, Brown MZ, Skutch JM, … Harned MS (2009). Which training method works best? A randomized controlled trial comparing three methods of training clinicians in dialectical behavior therapy skills. Behaviour Research and Therapy, 47(11), 921–930. 10.1016/j.brat.2009.07.011 [DOI] [PubMed] [Google Scholar]
  29. Dimeff LA, Woodcock EA, Harned MS, & Beadnell B (2011). Can dialectical behavior therapy be learned in highly structured learning environments? Results from a randomized controlled dissemination trial. Behavior Therapy, 42(2), 263–275. 10.1016/j.beth.2010.06.004 [DOI] [PubMed] [Google Scholar]
  30. Dorsey S, Berliner L, Lyon AR, Pullmann MD, & Murray LK (2016). A statewide common elements initiative for children’s mental health. The Journal of Behavioral Health Services & Research, 43(2), 246–261. 10.1007/s11414-014-9430-y [DOI] [PMC free article] [PubMed] [Google Scholar]
  31. Eccles MP, & Mittman BS (2006). Welcome to implementation science. Implementation Science, 1(1), 1. 10.1186/1748-5908-1-1 [DOI] [Google Scholar]
  32. Edwards KS, Rosen RC, Smith LD, Garvert DW, Graham BC, Hoyman LC, … Ruzek JI (2017). Group telephone consultation after online self-administered training: Acceptability and feasibility. Training and Education in Professional Psychology, 11(3), 198–206. 10.1037/tep0000153 [DOI] [Google Scholar]
  33. Ehrenreich-May J, Dimeff LA, Woodcock EA, Queen AH, Kelly T, Contreras IS, … Barlow DH (2016). Enhancing online training in an evidence-based treatment for adolescent panic disorder: A randomized controlled trial. Evidence-Based Practice in Child and Adolescent Mental Health, 1(4), 241–258. 10.1080/23794925.2016.1245116 [DOI] [Google Scholar]
  34. Fairburn CG, Allen E, Bailey-Straebler S, O’Connor ME, & Cooper Z (2017). Scaling up psychological treatments: A countrywide test of the online training of therapists. Journal of Medical Internet Research, 19(6), e214. 10.2196/jmir.7864 [DOI] [PMC free article] [PubMed] [Google Scholar]
  35. Farrell NR, Kemp JJ, Blakey SM, Meyer JM, & Deacon BJ (2016). Targeting clinician concerns about exposure therapy: A pilot study comparing standard vs. enhanced training. Behaviour Research and Therapy, 85, 53–59. 10.1016/j.brat.2016.08.011 [DOI] [PubMed] [Google Scholar]
  36. Fixsen DL, Naoom SF, Blase KA, Friedman RM, & Wallace F (2005). Implementation research: A synthesis of the Literature. The National Implementation Research Network: University of South Florida, Louis de la Parte Florida Mental Health Institute. [Google Scholar]
  37. Garland AF, Bickman L, & Chorpita BF (2010). Change what? Identifying quality improvement targets by investigating usual mental health care. Administration and Policy in Mental Health and Mental Health Services Research, 37(1), 15–26. 10.1007/s10488-010-0279-y [DOI] [PMC free article] [PubMed] [Google Scholar]
  38. Gellis ZD, Solomon P, Cullen SW, Lukens J, Gleba J, Zalucki D, & Huz S (2014). Dissemination of evidence-based depression care for community-dwelling older adults. Best Practices in Mental Health: An International Journal, 10(1), 1–15. [Google Scholar]
  39. German RE, Adler A, Frankel SA, Stirman SW, Pinedo P, Evans AC, … Creed TA (2017). Testing a web-based, trainedpeer model to build capacity for evidence-based practices in community mental health systems. Psychiatric Services, 69(3), 286–292. 10.1176/appi.ps.201700029 [DOI] [PubMed] [Google Scholar]
  40. Greenhalgh T, Robert G, Bate P, Kyriakidou O, Macfarlane F, & Peacock R (2004). How to spread good ideas – A systematic review of the literature on diffusion, dissemination and sustainability of innovations in health service delivery and organisation. Report for the National Coordinating Center for NHS Service Delivery and Organization, 1–426. [Google Scholar]
  41. Gryglewicz K, Chen JI, Romero GD, Karver MS, & Witmeier M (2016). Online suicide risk assessment and management training. Crisis: The Jounral of Crisis Intervention and Suicide. Prevention, 38(3), 186–194. 10.1027/0227-5910/a000421 [DOI] [PubMed] [Google Scholar]
  42. Gyani A, Shafran R, Myles P, & Rose S (2014). The gap between science and practice: How therapists make their clinical decisions. Behaior Therapy, 45(2), 199–211. 10.1016/j.beth.2013.10.004 [DOI] [PubMed] [Google Scholar]
  43. Hamblen JL, Norris FH, Gibson L, & Lee L (2010). Training community therapists to deliver cognitive behavioral therapy in the aftermath of disaster. International Journal of Emergency Mental Health, 12(1), 33–40. [PubMed] [Google Scholar]
  44. Harned MS, Dimeff LA, Woodcock EA, & Contreras I (2013). Predicting adoption of exposure therapy in a randomized controlled dissemination trial. Journal of Anxiety Disorders, 27(8), 754–762. 10.1016/j.janxdis.2013.02.006 [DOI] [PMC free article] [PubMed] [Google Scholar]
  45. Harned MS, Dimeff LA, Woodcock EA, Kelly T, Zavertnik J, Contreras I, & Danner SM (2014). Exposing clinicians to exposure: A randomized controlled dissemination trial of exposure therapy for anxiety disorders. Behavior Therapy, 45(6), 731. 10.1016/j.beth.2014.04.005 [DOI] [PMC free article] [PubMed] [Google Scholar]
  46. Harned MS, Dimeff LA, Woodcock EA, & Skutch JM (2011). Overcoming barriers to disseminating exposure therapies for anxiety disorders: A pilot randomized controlled trial of training methods. Journal of Anxiety Disorders, 25(2), 155–163. 10.1016/j.janxdis.2010.08.015 [DOI] [PMC free article] [PubMed] [Google Scholar]
  47. Henggeler SW, Chapman JE, Rowland MD, Sheidow AJ, & Cunningham PB (2013). Evaluating training methods for transporting contingency management to therapists. Journal of Substance Abuse Treatment, 45(5), 466–474. 10.1016/j.jsat.2013.06.008 [DOI] [PMC free article] [PubMed] [Google Scholar]
  48. Herschell AD, Kolko DJ, Baumann BL, & Davis AC (2010). The role of therapist training in the implementation of psychosocial treatments: A review and critique with recommendations. Clinical Psychology Review, 30(4), 448–466. 10.1016/j.cpr.2010.02.005 [DOI] [PMC free article] [PubMed] [Google Scholar]
  49. Herschell AD, Lindhiem OJ, Kogan JN, Celedonia KL, & Stein BD (2014). Evaluation of an implementation initiative for embedding dialectical behavior therapy in community settings. Evaluation and Program Planning, 43, 55–63. 10.1016/j.evalprogplan.2013.10.007 [DOI] [PMC free article] [PubMed] [Google Scholar]
  50. Huppert JD, Bufka LF, Barlow DH, Gorman JM, Shear MK, & Woods SW (2001). Therapist, therapist variables, and cognitive-behavioral therapy outcome in multicenter trial for panic disorder. Journal of Consulting and Clinical Psychology, 69, 747–755. 10.1037/0022-006X.69.5.747 [DOI] [PubMed] [Google Scholar]
  51. Hurlburt MS, Garland AF, Nguyen K, & Brookman-Frazee L (2010). Child and family therapy process: Concordance of therapist and observational perspectives. Administration and Policy in Mental Health and Mental Health Services Research, 37(3), 230–244. 10.1007/s10488-009-0251-x [DOI] [PMC free article] [PubMed] [Google Scholar]
  52. Jackson CB, Herschell AD, Schaffner KF, Turiano NA, & McNeil CB (2017). Training community-based clinicians in parent-child interaction therapy: The interaction between expert consultation and caseload. Professional Psychology: Research and Practice, 48(6), 481. 10.1037/pro0000149 [DOI] [PMC free article] [PubMed] [Google Scholar]
  53. Jones E, Manassis K, Arnold P, Ickowicz A, Mendlowitz S, Nowrouzi B, … Schmidt F (2015). Translating cognitive behavioral therapy for anxious youth to rural-community settings via tele-psychiatry. Community Mental Health Journal, 51(7), 852–856. 10.1007/s10597-015-9882-4 [DOI] [PubMed] [Google Scholar]
  54. Karlin BE, Brown GK, Trockel M, Cunning D, Zeiss AM, & Taylor CB (2012). National dissemination of cognitive behavioral therapy for depression in the Department of Veterans Affairs health care system: Therapist and patient-level outcomes. Journal of Consulting and Clinical Psychology, 80(5), 707–718. 10.1037/a0029328 [DOI] [PubMed] [Google Scholar]
  55. Karlin BE, Trockel M, Taylor CB, Gimeno J, & Manber R (2013). National dissemination of cognitive behavioral therapy for insomnia in veterans: Therapist- and patient-level outcomes. Journal of Consulting and Clinical Psychology, 81(5), 912–917. 10.1037/a0032554 [DOI] [PubMed] [Google Scholar]
  56. Kilbourne AM, Beck K, Spaeth-Rublee B, Ramanuj P, O’Brien RW, Tomoyasu N, & Pincus HA (2018). Measuring and improving the quality of mental health care: A global perspective. World Psychiatry, 17(1), 30–38. 10.1002/wps.20482 [DOI] [PMC free article] [PubMed] [Google Scholar]
  57. Kobak KA, Craske MG, Rose RD, & Wolitsky-Taylor K (2013). Web-based therapist training on cognitive behavior therapy for anxiety disorders: A pilot study. Psychotherapy, 50(2), 235. 10.1037/a0030568 [DOI] [PMC free article] [PubMed] [Google Scholar]
  58. Kobak KA, Lipsitz JD, Markowitz JC, & Bleiberg KL (2017). Web-based therapist training in interpersonal psychotherapy for depression: Pilot study. Journal of Medical Internet Research, 19(7), e257. 10.2196/jmir.7966 [DOI] [PMC free article] [PubMed] [Google Scholar]
  59. Kobak KA, Wolitzky-Taylor K, Craske MG, & Rose RD (2017). Therapist training on cognitive behavior therapy for anxiety disorders using internet-based technologies. Cognitive Therapy and Research, 41(2), 252–265. 10.1007/s10608-016-9819-4 [DOI] [PMC free article] [PubMed] [Google Scholar]
  60. Kolko DJ, Baumann BL, Herschell AD, Hart JA, Holden EA, & Wisniewski SR (2012). Implementation of AF-CBT by community practitioners serving child welfare and mental health: A randomized trial. Child Maltreatment, 17(1), 32–46. 10.1177/1077559511427346 [DOI] [PubMed] [Google Scholar]
  61. Leathers SJ, Melka-Kaffer C, Spielfogel JE, & Atkins MS (2016). Use of evidence-based interventions in child welfare: Do attitudes matter? Children and Youth Services Review, 70, 375–382. 10.1016/j.childyouth.2016.10.022 [DOI] [PMC free article] [PubMed] [Google Scholar]
  62. Leathers SJ, Spielfogel JE, Blakey J, Christian E, & Atkins MS (2016). The effect of a change agent on use of evidence-based mental health practices. Administration and Policy in Mental Health and Mental Health Services Research, 43(5), 768–782. 10.1007/s10488-015-0694-1 [DOI] [PMC free article] [PubMed] [Google Scholar]
  63. Leffler JM, Jackson Y, West AE, McCarty CA, & Atkins MS (2013). Training in evidence-based practice across the professional continuum. Professional Psychology: Research and Practice, 44(1), 20–28. 10.1037/a0029241 [DOI] [Google Scholar]
  64. Lewis CC, Fischer S, Weiner BJ, Stanick C, Kim M, & Martinez RG (2015). Outcomes for implementation science: An enhanced systematic review of instruments using evidence-based rating criteria. Implementation Science, 10(1), 155. 10.1186/s13012-015-0342-x [DOI] [PMC free article] [PubMed] [Google Scholar]
  65. Lewis CC, & Simons AD (2011). A pilot study disseminating cognitive behavioral therapy for depression: Therapist factors and perceptions of barriers to implementation. Administration and Policy in Mental Health and Mental Health Services Research, 38(4), 324–334. 10.1007/s10488-011-0348-x [DOI] [PubMed] [Google Scholar]
  66. Lim A, Nakamura BJ, Higa-McMillan CK, Shimabukuro S, & Slavin L (2012). Effects of workshop trainings on evidence-based practice knowledge and attitudes among youth community mental health providers. Behaviour Research and Therapy, 50(6), 397. 10.1016/j.brat.2012.03.008 [DOI] [PubMed] [Google Scholar]
  67. Lloyd D, Couineau A-L, Hawkins K, Kartal D, Nixon RD, Perry D, & Forbes D (2015). Preliminary outcomes of implementing cognitive processing therapy for posttraumatic stress disorder across a national veterans’ treatment service. Journal of Clinical Psychiatry, 76(11), e1405–e1409. 10.4088/JCP.14m09139 [DOI] [PubMed] [Google Scholar]
  68. Lopez MA, & Basco MR (2011). Feasibility of dissemination of cognitive behavioral therapy to Texas community mental health centers. Journal of Behavioral Health Services and Research, 38(1), 91–104. 10.1007/s11414-009-9209-8 [DOI] [PubMed] [Google Scholar]
  69. Lu W, Yanos PT, Gottlieb JD, Duva SM, Silverstein SM, Xie H, … Mueser KT (2012). Using fidelity assessments to train frontline clinicians in the delivery of cognitive-behavioral therapy for PTSD in persons with serious mental illness. Psychiatric Services, 63(8), 785–792. 10.1176/appi.ps.201000458 [DOI] [PMC free article] [PubMed] [Google Scholar]
  70. Luoma JB, & Vilardaga JP (2013). Improving therapist psychological flexibility while training acceptance and commitment therapy: A pilot study. Cognitive Behaviour Therapy, 42(1), 1–8. 10.1080/16506073.2012.701662 [DOI] [PubMed] [Google Scholar]
  71. Lyon AR, Charlesworth-Attie S, Stoep AV, & McCauley E (2011). Modular psychotherapy for youth with internalizing problems: Implementation with therapists in school-based health centers. School Psychology Review, 40(4), 569. [Google Scholar]
  72. Lyon AR, Dorsey S, Pullmann M, Silbaugh-Cowdin J, & Berliner L (2015). Clinician use of standardized assessments following a common elements psychotherapy training and consultation program. Administration and Policy in Mental Health and Mental Health Services Research, 42(1), 47–60. 10.1007/s10488-014-0543-7 [DOI] [PMC free article] [PubMed] [Google Scholar]
  73. Mallonee S, Phillips J, Holloway K, & Riggs D (2018). Training providers in the use of evidence-based treatments: A comparison of in-person and online delivery modes. Psychology Learning & Teaching, 17(1), 61–72. 10.1177/1475725717744678 [DOI] [Google Scholar]
  74. Manber R, Trockel M, Batdorf W, Siebern AT, Taylor CB, Gimeno J, & Karlin BE (2013). Lessons learned from the national dissemination of cognitive behavioral therapy for insomnia in the Veterans Health Administration: Impact of training on therapists’ self-efficacy and attitudes. Sleep Medicine Clinics, 8(3), 399–405. 10.1016/j.jsmc.2013.05.003 [DOI] [Google Scholar]
  75. Martin J, Gladstone T, Diehl A, & Beardslee W (2016). Development and evaluation of a web-based clinician training program for a family-focused depression preventive intervention. Journal of Technology in Human Services, 34(3), 267–284. 10.1080/15228835.2016.1218311 [DOI] [Google Scholar]
  76. Martino S, Ball SA, Nich C, Canning-Ball M, Rounsaville BJ, & Carroll KM (2010). Teaching community program clinicians motivational interviewing using expert and train-the-trainer strategies. Addiction, 106(2), 428–441. 10.1111/j.1360-0443.2010.03135.x [DOI] [PMC free article] [PubMed] [Google Scholar]
  77. McHugh RK, & Barlow DH (2010). The dissemination and implementation of evidence-based psychological treatments: A review of current efforts. American Psychologist, 65(2), 73–84. 10.1037/a0018121 [DOI] [PubMed] [Google Scholar]
  78. McManus F, Westbrook D, Vazquez-Montes M, Fennell M, & Kennerley H (2010). An evaluation of the effectiveness of diploma-level training in cognitive behaviour therapy. Behaviour Research and Therapy, 48(11), 1123–1132. [DOI] [PubMed] [Google Scholar]
  79. Michie S, van Stralen MM, & West R (2011). The behaviour change wheel: A new method for characterising and designing behaviour change interventions. Implementation Science, 6(1), 42. 10.1186/1748-5908-6-42 [DOI] [PMC free article] [PubMed] [Google Scholar]
  80. Mirick R, McCauley J, Bridger J, & Berkowitz L (2016). Continuing education on suicide assessment and crisis intervention: What can we learn about the needs of mental health professionals in community practice? Community Mental Health Journal, 52(5), 501–510. 10.1007/s10597-015-9884-2 [DOI] [PubMed] [Google Scholar]
  81. Monson CM, Shields N, Suvak MK, Lane JEM, Shnaider P, Landy MSH, … Stirman SW (2018). A randomized controlled effectiveness trial of training strategies in cognitive processing therapy for posttraumatic stress disorder: Impact on patient outcomes. Behaviour Research and Therapy, 110, 31–40. 10.1016/j.brat.2018.08.007 [DOI] [PubMed] [Google Scholar]
  82. Nadeem E, Gleacher A, & Beidas RS (2013). Consultation as an implementation strategy for evidence-based practices across multiple contexts: Unpacking the black box. Administration and Policy in Mental Health and Mental Health Services Research, 40(6), 439–450. 10.1007/s10488-013-0502-8 [DOI] [PMC free article] [PubMed] [Google Scholar]
  83. Nakamura BJ, Selbo-Bruns A, Okamura K, Chang J, Slavin L, & Shimabukuro S (2014). Developing a systematic evaluation approach for training programs within a train-the-trainer model for youth cognitive behavior therapy. Behaviour Research and Therapy, 53, 10–19. 10.1016/j.brat.2013.12.001 [DOI] [PubMed] [Google Scholar]
  84. Nathan PE, & Gorman JM (2002). A guide to treatments that work (2nd ed.). New York, NY: Oxford University Press. [Google Scholar]
  85. Nathan PE, & Gorman JM (2007). A guide to treatments that work (3rd ed.). New York, NY: Oxford University Press. [Google Scholar]
  86. Navarro-Haro MV, Harned MS, Korslund KE, DuBose A, Chen T, Ivanoff A, & Linehan MM (2019). Predictors of Adoption and Reach Following Dialectical Behavior Therapy Intensive Training™. Community Mental Health Journal, 55(1), 100–111. [DOI] [PubMed] [Google Scholar]
  87. Ngo VK, Centanni A, Wong E, Wennerstrom A, & Miranda J (2011). Building capacity for cognitive behavioral therapy delivery for depression in disaster-impacted contexts. Ethnicity & Disease, 21(3 Suppl 1), S1– 38–44. [PMC free article] [PubMed] [Google Scholar]
  88. Nowrouzi B, Manassis K, Jones E, Bobinski T, & Mushquash CJ (2015). Translating anxiety-focused CBT for youth in a first nations context in Northwestern Ontario. Journal of the Canadian Academy of Child and Adolescent Psychiatry, 24(1), 33–40. [PMC free article] [PubMed] [Google Scholar]
  89. Okamura KH, Benjamin Wolk CL, Kang-Yi CD, Stewart R, Ruin RM, Weaver S, … Mandell DS (2018). The price per prospective consumer of providing therapist training and consultation in seven evidence-based treatments within a large public behavioral health system: An example cost-analysis metric. Frontiers in Public Health, 5, 356. 10.3389/fpubh.2017.00356 [DOI] [PMC free article] [PubMed] [Google Scholar]
  90. Ouzzani M, Hammady H, Fedorowicz Z, & Elmagarmid A (2016). Rayyan – A web and mobile app for systematic reviews. Systematic Reviews, 5, 210. 10.1186/s13643-016-0384-4 [DOI] [PMC free article] [PubMed] [Google Scholar]
  91. Pemberton JR, Conners-Burrow NA, Sigel BA, Sievers CM, Stokes LD, & Kramer TL (2017). Factors associated with clinician participation in TF-CBT post-workshop training components. Administration and Policy in Mental Health and Mental Health Services Research, 44(4), 524–533. 10.1007/s10488-015-0677-2 [DOI] [PubMed] [Google Scholar]
  92. Persons JB, Koerner K, Eidelman P, Thomas C, & Liu H (2016). Increasing psychotherapists’ adoption and implementation of the evidence-based practice of progress monitoring. Behaviour Research and Therapy, 76, 24–31. 10.1016/j.brat.2015.11.004 [DOI] [PMC free article] [PubMed] [Google Scholar]
  93. Petry NM, Alessi SM, & Ledgerwood DM (2012). Contingency management delivered by community therapists in outpatient settings. Drug and Alcohol Dependence, 122(1–2), 86–92. 10.1016/j.drugalcdep.2011.09.015 [DOI] [PMC free article] [PubMed] [Google Scholar]
  94. Puspitasari A, Kanter JW, Murphy J, Crowe A, & Koerner K (2013). Developing an online, modular, active learning training program for behavioral activation. Psychotherapy Theory Research Practice Training, 50(2), 256–265. 10.1037/a0030058 [DOI] [PubMed] [Google Scholar]
  95. Rakovshik SG, & McManus F (2010). Establishing evidence-based training in cognitive behavioral therapy: A review of current empirical findings and theoretical guidance. Clinical Psychology Review, 30(5), 496–516. 10.1016/j.cpr.2010.03.004 [DOI] [PubMed] [Google Scholar]
  96. Rakovshik SG, McManus F, Vazquez-Montes M, Muse K, & Ougrin D (2016). Is supervision necessary? Examining the effects of internet-based CBT training with and without supervision. Journal of Consulting and Clinical Psychology, 84(3), 191. 10.1037/ccp0000079 [DOI] [PubMed] [Google Scholar]
  97. Rakovshik SG, McManus F, Westbrook D, Kholmogorova AB, Garanian NG, Zvereva NV, & Ougrin D (2013). Randomized trial comparing Internet-based training in cognitive behavioural therapy theory, assessment and formulation to delayed-training control. Behaviour Research and Therapy, 51(6), 231–239. 10.1016/j.brat.2013.01.009 [DOI] [PubMed] [Google Scholar]
  98. Reese HE, Pollard CA, Szymanski J, Berman N, Crowe K, Rosenfield E, & Wilhelm S (2016). The behavior therapy training institute for OCD: A preliminary report. Journal of Obsessive Compulsive and Related Disorders, 8, 79–85. 10.1016/j.jocrd.2015.12.005 [DOI] [Google Scholar]
  99. Richards R, Oliver JE, Morris E, Aherne K, Iervolino AC, & Wingrove J (2011). Acceptance and commitment therapy training for clinicians: An evaluation. The Cognitive Behaviour Therapist, 4(3), 114–121. 10.1017/S1754470X11000043 [DOI] [Google Scholar]
  100. Ringel JS, & Sturm R (2001). National estimates of mental health utilization and expenditures for children in 1998. The Journal of Behavioral Health Services & Research, 28(3), 319–333. 10.1007/BF02287247 [DOI] [PubMed] [Google Scholar]
  101. Ruzek JI, Eftekhari A, Crowley J, Kuhn E, Karlin BE, & Rosen CS (2017). Post-training beliefs, intentions, and use of prolonged exposure therapy by clinicians in the Veterans Health Administration. Administration and Policy in Mental Health and Mental Health Services Research, 44(1), 123–132. 10.1007/s10488-015-0689-y [DOI] [PubMed] [Google Scholar]
  102. Ruzek JI, Eftekhari A, Rosen CS, Crowley JJ, Kuhn E, Foa EB, … Karlin BE (2016). Effects of a comprehensive training program on clinician beliefs about and intention to use prolonged exposure therapy for PTSD. Psychological Trauma: Theory, Research, Practice, and Policy, 8(3), 348–355. 10.1037/tra0000004 [DOI] [PubMed] [Google Scholar]
  103. Ruzek JI, Rosen RC, Garvert DW, Smith LD, Sears KC, Marceau L, … Stoddard AM (2014). Online self-administered training of PTSD treatment providers in cognitive–behavioral intervention skills: Results of a randomized controlled trial. Journal of Traumatic Stress, 27(6), 703–711. 10.1002/jts.21977 [DOI] [PubMed] [Google Scholar]
  104. Schoenwald SK, Sheidow AJ, & Letourneau EJ (2004). Toward effective quality assurance in evidence-based practice: Links between expert consultation, therapist fidelity, and child outcomes. Journal of Clinical Child and Adolescent Psychology, 33(1), 94–104. 10.1207/S15374424JCCP3301_10 [DOI] [PubMed] [Google Scholar]
  105. Scott K, Klech D, Lewis CC, & Simons AD (2016). What did they learn? Effects of a brief cognitive behavioral therapy workshop on community therapists’ knowledge. Community Mental Health Journal, 52(8), 998–1003. 10.1007/s10597-015-9876-2 [DOI] [PMC free article] [PubMed] [Google Scholar]
  106. Shah A, Scogin F, Presnell A, Morthland M, & Kaufman A (2013). Social workers as research psychotherapists in an investigation of cognitive-behavioral therapy among rural older adults. Social Work Research, 37(2), 137–145. 10.1093/swr/svt011 [DOI] [PMC free article] [PubMed] [Google Scholar]
  107. Shiner B, D’Avolio LW, Nguyen TM, Zayed MH, Young-Xu Y, Desai RA, … Watts BV (2013). Measuring use of evidence based psychotherapy for posttraumatic stress disorder. Administration and Policy in Mental Health and Mental Health Services Research, 40(4), 311–318. 10.1007/s10488-012-0421-0 [DOI] [PubMed] [Google Scholar]
  108. Simons AD, Padesky CA, Montemarano J, Lewis CC, Murakami J, Lamb K, … Beck AT (2010). Training and dissemination of cognitive behavior therapy for depression in adults: A preliminary examination of therapist competence and client outcomes. Journal of Consulting and Clinical Psychology, 78(5), 751–756. 10.1037/a0020569 [DOI] [PubMed] [Google Scholar]
  109. Smith TL, Landes SJ, Lester-Williams K, Day KT, Batdorf W, Brown GK, … Weingardt KR (2017). Developing alternative training delivery methods to improve psychotherapy implementation in the US Department of Veterans Affairs. Training and Education in Professional Psychology, 11(4), 266–275. 10.1037/tep0000156 [DOI] [Google Scholar]
  110. Southam-Gerow MA, Daleiden EL, Chorpita BF, Bae C, Mitchell C, Faye M, & Alba M (2014). MAPping Los Angeles County: Taking an evidence-informed model of mental health care to scale. Journal of Clinical Child & Adolescent Psychology, 43(2), 190–200. [DOI] [PubMed] [Google Scholar]
  111. Stein BD, Celedonia KL, Swartz HA, DeRosier ME, Sorbero MJ, Brindley RA, … Frank E (2015). Implementing a web-based intervention to train community clinicians in an evidence-based psychotherapy: A pilot study. Psychiatric Services, 66(9), 988–991. 10.1176/appi.ps.201400318 [DOI] [PMC free article] [PubMed] [Google Scholar]
  112. Stephan SH, Connors EH, Arora P, & Brey L (2013). A learning collaborative approach to training school-based health providers in evidence-based mental health treatment. Children and Youth Services Review, 35(12), 1970–1978. 10.1016/j.childyouth.2013.09.008 [DOI] [Google Scholar]
  113. Stirman SW, Gutner C, Crits-Christoph P, Edmunds J, Evans AC, & Beidas RS (2015). Relationships between clinician-level attributes and fidelity-consistent and fidelity-inconsistent modifications to an evidence-based psychotherapy. Implementation Science, 10(1), 115. 10.1186/s13012-015-0308-z [DOI] [PMC free article] [PubMed] [Google Scholar]
  114. Stirman SW, Kimberly J, Cook N, Calloway A, Castro F, & Charns M (2012). The sustainability of new programs and innovations: A review of the empirical literature and recommendations for future research. Implementation Science, 7(1), 17. 10.1186/1748-5908-7-17 [DOI] [PMC free article] [PubMed] [Google Scholar]
  115. Stirman SW, Pontoski K, Creed T, Xhezo R, Evans AC, Beck AT, & Crits-Christoph P (2017). A non-randomized comparison of strategies for consultation in a community-academic training program to implement an evidence-based psychotherapy. Administration and Policy in Mental Health and Mental Health Services Research, 44(1), 55–66. 10.1007/s10488-015-0700-7 [DOI] [PMC free article] [PubMed] [Google Scholar]
  116. Stirman SW, Shields N, Deloriea J, Landy MS, Belus JM, Maslej MM, & Monson CM (2013). A randomized controlled dismantling trial of post-workshop consultation strategies to increase effectiveness and fidelity to an evidence-based psychotherapy for posttraumatic stress disorder. Implementation Science, 8(1), 82. 10.1186/1748-5908-8-82 [DOI] [PMC free article] [PubMed] [Google Scholar]
  117. van den Berg DP, van der Vleugel BM, de Bont PA, Thijssen G, de Roos C, de Kleine R, … van der Gaag M (2016). Exposing therapists to trauma-focused treatment in psychosis: Effects on credibility, expected burden, and harm expectancies. European Journal of Psychotraumatology, 7, 31712. 10.3402/ejpt.v7.31712 [DOI] [PMC free article] [PubMed] [Google Scholar]
  118. Wade D, Crompton D, Howard A, Stevens N, Metcalf O, Brymer M, … Forbes D (2014). Skills for psychological recovery: Evaluation of a post-disaster mental health training program. Disaster Health, 2(3–4), 138–145. 10.1080/21665044.2015.1085625 [DOI] [PMC free article] [PubMed] [Google Scholar]
  119. Waller G, Walsh KDS, & Wright C (2016). Impact of education on clinicians’ attitudes to exposure therapy for eating disorders. Behaviour Research and Therapy, 76, 76–80. 10.1016/j.brat.2015.11.014 [DOI] [PubMed] [Google Scholar]
  120. Walser RD, Karlin BE, Trockel M, Mazina B, & Barr Taylor C (2013). Training in and implementation of acceptance and commitment therapy for depression in the Veterans Health Administration: Therapist and patient outcomes. Behaviour Research and Therapy, 51(9), 555–563. 10.1016/j.brat.2013.05.009 [DOI] [PubMed] [Google Scholar]
  121. Walsh LM, Roddy MK, Scott K, Lewis CC, & Jensen-Doss A (2018). A meta-analysis of the effect of therapist experience on outcomes for clients with internalizing disorders. Psychotherapy Research, 29(7), 846–859. 10.1080/10503307.2018.1469802 [DOI] [PMC free article] [PubMed] [Google Scholar]
  122. Webster-Stratton CH, Reid MJ, & Marsenich L (2014). Improving therapist fidelity during implementation of evidence-based practices: Incredible years program. Psychiatric Services, 65(6), 789–795. 10.1176/appi.ps.201200177 [DOI] [PubMed] [Google Scholar]
  123. Weissman MM, Verdeli H, Gameroff MJ, Bledsoe SE, Betts K, Mufson L, … Wickramaratne P (2006). National survey of psychotherapy training in psychiatry, psychology, and social work. Archives of General Psychiatry, 63(8), 925–934. 10.1001/archpsyc.63.8.925 [DOI] [PubMed] [Google Scholar]
  124. Williams C, Martinez R, Dafters R, Ronald L, & Garland A (2011). Training the wider workforce in cognitive behavioural self-help: The SPIRIT (Structured Psychosocial InteRventions in Teams) training course. Behavioural and Cognitive Psychotherapy, 39(2), 139–149. 10.1017/S1352465810000445 [DOI] [PubMed] [Google Scholar]
  125. Williams NJ, & Beidas RS (2019). Annual research review: The state of implementation science in child psychology and psychiatry: A review and suggestions to advance the field. Journal of Child Psychology and Psychiatry, 60(4), 430–450. 10.1111/jcpp.12960 [DOI] [PMC free article] [PubMed] [Google Scholar]
  126. Wolk CB, Becker-Haimes EM, Fishman J, Affrunti N, Mandell D, & Creed T (2019). Variability in clinician intentions to implement specific cognitive-behavioral therapy components. BMC Psychiatry, 19(1), 406. 10.1186/s12888-019-2394-y [DOI] [PMC free article] [PubMed] [Google Scholar]
  127. Young J, & Beck AT (1980). Cognitive Therapy Scale: Rating manual (Unpublished manuscript). Philadelphia, PA: Center for Cognitive Therapy, University of Pennsylvania. [Google Scholar]

RESOURCES