Skip to main content
Healthcare logoLink to Healthcare
. 2022 Nov 2;10(11):2204. doi: 10.3390/healthcare10112204

Effectiveness of Educational Interventions to Increase Skills in Evidence-Based Practice among Nurses: The EDITcare Systematic Review

Omar Portela Dos Santos 1,2,*, Pauline Melly 1, Roger Hilfiker 3, Katia Giacomino 3, Elodie Perruchoud 1, Henk Verloo 1,4, Filipa Pereira 1
Editor: César Leal-Costa
PMCID: PMC9691114  PMID: 36360544

Abstract

Background: Using evidence-based practice (EBP) improves the implementation of safe, high-quality healthcare for patients, reduces avoidable costs, and plays a crucial role in bridging knowledge–action gaps and reducing health inequities. EBP combines the best available evidence in the relevant literature with patient preferences and values and healthcare professionals’ (HCPs) expertise. Methods: Systematic searches of ten bibliographic databases, unpublished works, and the Grey Literature Report sought studies published up to 30 September 2022. Results: The 15 studies retained involved 2712 nurses. Three types of effective educational interventions were identified: (1) multifaceted educational strategies incorporating mentoring and tutoring; (2) single educational strategies, often delivered online; and (3) multifaceted educational strategies using the five steps of EBP. Eleven primary outcomes (EBP beliefs, EBP self-efficacy, perceived EBP implementation, EBP competencies, EBP knowledge, EBP skills, EBP attitudes, EBP behaviors, EBP desire, EBP practice, and perceptions of organizational culture and readiness) were assessed using 13 qualitative and quantitative instruments. Conclusions: Ensuring the successful implementation of EBP requires effective educational strategies. Computer-based learning seems the most cost-effective and efficient strategy, when considering caregivers’ characteristics, the clinical field, and educational interventions across the pre-, peri-, and post-implementation processes.

Keywords: evidence-based practice, primary healthcare, beliefs, knowledge, implementation, nurses, interventions, education

1. Introduction

The reports “To Err is Human” (1999) [1] and “Crossing the Quality Chasm” (2001) [2] created a level of global awareness that pushed healthcare institutions to take corrective actions to promote and improve the quality of clinical care and patient safety [3]. According to the Institute of Medicine, quality of care is “the ability of health services to increase the likelihood of achieving desired health outcomes in accordance with current professional knowledge” [3]. At the same time, the concept of patient safety—defined as the prevention of harm [4]—emerged as a major focus of professional vigilance. Healthcare systems today face the twin challenges of implementing EBP [5,6] to reduce healthcare costs [7] and increasing healthcare personnel’s job satisfaction [8]. Indeed, EBP contributes to improving the quality of care and patient safety because effectiveness, safety, responsiveness, efficiency, and equity have become indicators of whether care providers are using patient-centered practices. EBP is now an established framework among healthcare professionals (HCPs) [9,10,11,12]. As an integral part of patient-centered approaches, it involves the use of research steps to critically appraise research evidence and implement that evidence in practice [11,13,14]. In a nursing context, EBP is defined as the integration of information from various sources to aid clinical reasoning, including the best and most recent evidence, clinical expertise, personal experiences, patient preferences, and theories underpinning nursing care [5,6,15]. HCPs are expected to use EBP as a gold-standard approach to their daily practice [12,16,17,18,19,20]. One might expect EBP to be well known and understood by HCPs and, equally, to be prevalent in their education programs and clinical practice [15]. However, the routine use of EBP remains challenging [11] since between 30% and 40% of patients do not receive care based on current knowledge and two-thirds of good practice implementations fail [17,18,19]. Moreover, some HCPs feel that engaging in EBP is not within their scope of practice [20]. Because not all HCPs have undergone training on EBP along their career trajectory, there are questions about which educational interventions are most effective at increasing EBP skills among HCPs in their daily practice [12,21].

Nurses and physiotherapists are highly involved in healthcare services [22,23]. Because of their close relationships with healthcare users during their daily practice, they play important decision-making roles, strengthening communication and collaborative practices between community and specialized HCPs to provide the best available overall healthcare [24]. Although it is generally considered that nurses and physiotherapists—like every other HCP—are accountable for providing the best available evidence-based healthcare [25,26], recent research has concluded that only a tiny percentage of them consistently do so [13,27]. Indeed, multiple barriers to the daily use of EBP have been reported, such as inadequate skills in EBP among clinicians, time constraints, negative attitudes, a lack of personal motivation, and professional resistance to research [28,29,30,31,32]. Additionally, several authors have documented administrative and organizational problems in the workplace, a lack of mentors for EBP, inadequate point-of-care resources, gaps between theory and practice, a lack of any meaningful transition between training courses on EBP and the clinical reality, and absent or insufficient basic education on the subject [7,12,18,33,34]. It is therefore urgent that efficient and effective educational interventions that meet HCPs’ needs be identified and developed according to a well-defined implementation process that ensures that EBP’s added value is felt at the patient’s bedside. The importance of providing educational interventions on EBP has been recognized by the World Health Organization [33]. Over the last two decades, EBP in healthcare has been documented in exploratory and observational studies in different settings. Scurlock-Evans et al. [13] summarized attitudes, barriers, enablers, and EBP interventions among physiotherapists, although without specifying employment settings or assessing educational interventions. Melender et al. [34] summarized the educational interventions used to train nursing students to improve outcomes in implementing EBP. Nevertheless, to the best of our knowledge, no systematic reviews have examined the effectiveness of educational interventions to increase skills in EBP in the daily practice of nurses and physiotherapists. Furthermore, no previous international research has ever compared the outcomes of different educational interventions on EBP skills in our population of interest. While EBP is considered as a core competency needed for HCPs, evidence for how to effectively teach EBP remains suboptimal and there is an important heterogeneity of EBP education interventions. Therefore, there is a need to synthetize and evaluate the outcomes of the different EBP educational interventions proposed by the literature, and to evaluate what is the most effective to overcome the barriers cited, meet the characteristics and needs of the professionals, and make EBP a norm for al HCPs need to be identified. It is an essential process for bridging the gap between evidence and practice, for the nursing and physiotherapist professions to advance as well-informed disciplines, for the homogenization of basic education on EBP, and/or the avoidance of a lack of any meaningful transition between training courses on EBP and the clinical reality [35]. This review aimed to answer the questions: (a) What are the most effective educational interventions to increase skills in EBP among nurses and physiotherapists delivering primary healthcare? (b) What are the effects of these educational interventions on EBP skills among primary-care nurses and physiotherapists? We hypothesized that researchers, clinicians, healthcare system managers, healthcare institutions, healthcare policymakers, stakeholders, universities and their faculty, and healthcare professionals could identify the most effective educational interventions that would enable them to use evidence in daily practice, enable HCPs to understand EBP, promote and improve the quality of clinical care and patient safety, and improve responses to current healthcare challenges. If EBP educational interventions become an integral part of continuing HCPs education curricula, it will improve skills in EBP, facilitate its implementation, and guarantee its holistic benefits. Our second hypothesis was that the identification of these educational interventions would enable the recognition of the factors facilitating and constraining the implementation of EBP and would contribute to establishing effective implementation strategies within different healthcare settings.

2. Materials and Methods

This review was conducted following the Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRISMA) recommendations [36], the Meta-analysis Of Observational Studies in Epidemiology (MOOSE) reporting proposals [37], and the methods outlined in the Cochrane Handbook for Systematic Reviews of Interventions [38]. The research protocol for this systematic review has been published elsewhere [39].

2.1. Study Selection Criteria

2.1.1. Participants and Healthcare Settings

This review considers studies involving registered HCPs delivering primary healthcare, including those with bachelor’s, master’s, and doctoral degrees in nursing (RNs = registered nurses) and physiotherapy (physiotherapists = PTs) and nursing and physiotherapy students in their respective educational institutions. Physical therapists and physiotherapists are considered synonymous. We also include all categories of primary healthcare settings for home-dwelling adult care, such as private offices, community, and health maintenance organizations (HMO), community and private primary healthcare settings, hospital outpatient departments (OPDs), and hospital settings.

2.1.2. Types of Studies

Randomized controlled trials (RCTs), cluster randomized controlled trials (RCTs), non-randomized studies (NRSs), prospective cohort studies, case-control studies, controlled before-and-after studies, interrupted-time-series studies, quasi-experimental studies, and mixed-methods studies were included [40,41,42].

2.1.3. Types of Educational Interventions

All kinds of educational interventions to improve the EBP skills of RNs, and PTs delivering active primary healthcare were examined. These included ex cathedra, interactive, online, or individual educational sessions on the steps and components of EBP, organized journal clubs, seminars, educational meetings, the distribution of educational materials, webinars and other individual-oriented educational activities, case studies, grand rounds, and mentoring within a healthcare organization. Educational interventions were excluded if they targeted the regulatory, economic, or financial aspects of EBP, as per the Cochrane Effective Practice and Organization of Care (EPOC, 2015) taxonomy of interventions [43].

2.1.4. Outcome Measures

Our primary outcome was whether educational interventions increased or decreased the EBP knowledge, attitudes, beliefs, competencies, skills, self-efficacy, behavior, practice and perception of RNs, and PTs active in primary healthcare settings. This was measured using methods such as questionnaires, interviews, chart analyses, and self-reporting by RNs, and PTs, with a focus on dichotomous (yes/no), ordinal, or continuous scales or scores. Secondary outcomes explored the production of systematic reviews and numbers of journal clubs and grand rounds [44,45].

2.2. Search and Process Strategies

2.2.1. Search Strategy and Identification of Relevant Studies

In collaboration with a medical librarian (PM) and using predefined search terms, we conducted a systematic literature search for published articles in the following electronic databases, from inception until 30 September 2022: Medline Ovid SP (from 1946), PubMed (NOT Medline[sb]) (from 1996), Embase.com (from 1947), CINAHL Ebesco (from 1937), the Cochrane Central Register of Controlled Trials Wiley (from 1992), PsycINFO Ovid SP (from 1806), Web of Science Core Collection (from 1900), PEDro (from 1999), the JBI Database of Systematic Reviews and Implementation Reports (from 1998), and the Trip Database (from 1997). We conducted a hand search of the bibliographies of all relevant articles and searched for unpublished studies using Google Scholar, ProQuest Dissertations and Theses dissemination, Mednar, and WorldCat. The search was completed a first time by exploring the grey literature in OpenGrey and the Grey Literature Report from inception until 31 December 2021, and a second time until 30 September 2022. We considered publications in any language.

2.2.2. Study Screening and Data Extraction

Three reviewers (OPDS, HV, RH) independently screened the titles and abstracts identified in our searches to assess which studies met our inclusion criteria. The full texts of the articles retained were read to ensure the assessments were correct. Disagreements were resolved through discussion or, if needed, a consensus was reached after discussion with the co-authors (KG, EP, FP). A flowchart of the article selection process was drawn using the PRISMA statement guidelines [36] (Figure 1). Three reviewers (OPDS, EP, HV) extracted the data independently using a specially designed, standardized data-extraction form. Again, discrepancies were resolved through discussion and consultation with the co-authors if necessary. The information extracted from each study included: (1) study authors, year of publication, and country where the study was conducted; (2) study characteristics (including setting and design, duration of follow-up, and sample size); (3) participants’ basic sociodemographic characteristics; (4) characteristics of interventions (description and frequency of educational interventions, healthcare professionals involved, etc.); (5) characteristics of the usual care group; and (6) outcome measures.

Figure 1.

Figure 1

Search strategy for the EDITcare systematic review [46].

2.3. Assessment of the Risks of Bias in the Studies Retained

Two reviewers (OPDS, HV) independently assessed the risks of bias in all the randomized and non-randomized studies of the interventions included. Disagreements were resolved through discussion and consultation with all the co-authors (EP, KG, RH, FP). Any disagreements in the quality assessments were resolved through discussion.

2.4. Statistical Analyses

Statistical analyses followed the recommendations of the Cochrane Handbook for Systematic Reviews of Interventions [38] and the PRISMA and MOOSE statements [36]. Descriptive statistics were used to describe the studies and participants involved. We computed the sum of the population and the mean age of the participants in the studies retained. We used exact statistics and measures of association to report on the effectiveness of the interventions in the studies. The extreme heterogeneity of the data collected meant that we could not compute group or subgroup meta-analyses.

2.5. Methodological Quality of the Studies Retained

Three different tools were used to assess the studies’ methodological quality. The Newcastle–Ottawa Scale (NOS) for assessing the quality of non-randomized studies in meta-analyses [47] was used to assess the only cohort study retrieved [48]. We used the validated Risk Of Bias In Non-randomized Studies of Interventions (ROBINS-I) tool [49] for the eight studies with a quasi-experimental design [50,51,52,53,54,55,56,57]. Finally, we used the Revised Cochrane Risk of Bias tool for randomized trials (RoB 2.0) [58]. Two researchers (OPDS, HV) independently rated the studies’ quality. Any disagreements about quality assessments were resolved by consensus discussion.

3. Results

3.1. Search Strategy and Results

Our literature search took place at two distinct time points, with a baseline search in 2019 and a supplementary search in December 2021: we retrieved a total of 18,299 references. The second search’s goal was to complete the first one and ensure up-to-date evidence. After removing duplicates, three researchers independently analyzed the titles and abstracts of 12,948 references. In the second phase, papers’ full texts were retrieved from the references and analyzed based on the inclusion and exclusion criteria. Finally, a third search was made in September 2022, during the revision process, with the objective of improving the literature review with information about the very latest research publications. Considering the importance of the topic we decided to conduct living systematic review based on the Cochrane recommendations, defined as “a systematic review that is continually updated, incorporating relevant new evidence as it becomes available” [59], by continual and active monitoring of the evidence, and incorporating relevant new evidence as it becomes available. Two appropriate new references were discovered. All the full texts respecting the inclusion criteria were analyzed and described in a structured reporting form (Figure 1).

3.2. Characteristics of the Retrieved Studies

The review included fifteen studies—nine in the United States of America [8,48,50,51,54,55,56,57,60], two in Taiwan [52,61], and one each in Portugal [62], Finland [63], Spain [53], and India [64]. All were published between 2011 and 2022. One study used a pre–post cohort design [48], four were RCTs [8,65,66], and ten used quasi-experimental designs [50,51,52,53,54,55,56,57,60,61]. Twelve studies were conducted in primary-care settings, one took place in a Portuguese nursing school [62], one in the nursing education institution of an Indian university [64], and one in Jaume University in Spain [53]. The total population included in the review was composed of 2712 RNs and licensed practical nurses (LPNs), bachelor’s degree students, Ph.D. students in Nursing Practice–Family Nurse Practice, nurse educators, emergency nurses, nurse managers, and visiting staff nurses. An original intention had been to look at PTs and primary HCPs but no relevant studies were found involving them.

The 1509 female and 108 male participants were aged between 19 and 65 years old. Seven studies did not report on their sample’s gender [50,54,55,56,61,64,66], and only nine studies provided information about their sample’s age distribution [48,50,51,52,53,60,61,62,64,66]. Table 1 describes the studies retained.

Table 1.

Characteristics of the studies retained (n = 13).

Study Authors Year Country Professionals Involved Study Design Setting Sample
Mean Age (SD)
Men/Women Covariates Included in the Study
(Professional and Sociodemographic)
1 Cardoso, D., Couto, F., Cardoso, A. F., Bobrowicz-Campos, E., Santos, L., Rodrigues, R., Coutinho, V., Pinto, D., Ramis, M-A, Alves Rodrigues, M., and Apostolo, J. [62] 2021 Portugal Students in the Bachelor of Nursing program, 8th semester RCT cluster with two parallel brains qualitative section:
  • -

    Evaluation of 18 bachelor’s degree students’ theses

  • -

    IC n = 9

  • -

    CG n = 9

Portuguese nursing school n = 11,148
  • -

    Mean age = 21.9 (SD = 2.2)

  • -

    Demographic characteristics

  • -

    Effectiveness of an EBP educational program

  • -

    Knowledge and skills in EBP

  • -

    Qualitative evaluation of bachelor’s degree students’ theses

2 Chao, W.-Y., Huang, L.C., Hung, H.-C., Hung, S.-C., Chuang, T.-F., Yeh, L.-Y., and Tseng, H.-C. [60] 2022 Taiwan RNs with more than 3 months’ work experience Quasi-experimental design
  • -

    114 licensed nurses

  • -

    CG n = 54

  • -

    IC n = 54

475-bed regional teaching hospital n = 114
  • -

    Mean age IC = 32.5 (SD = 8.84)

  • -

    Mean age CG = 33.84 (SD = 7.53)

Women = 109
Men = 5
  • -

    Demographic characteristics

  • -

    Knowledge, attitudes towards, and practice of EBP

  • -

    Satisfaction score

3 D’Souza, P., George, A., Nair, S., Noronha, J., and Renjith, V. [63] 2021 India Nurse educators involved in undergraduate and postgraduate nursing education programs RCT Nursing education institution in an Indian university n = 51
  • -

    IG n = 27

  • -

    CG n = 24

  • -

    Mean age IG = 34.8 (SD = 6.2)

  • -

    Mean age CG = 35.1 (SD = 7.6)

Not reported
  • -

    Effectiveness of an EBP training program

  • -

    Knowledge, attitudes, usage, and competency

  • -

    Demographic characteristics

4 Gallagher-Ford, L., Koshy Thomas, B., Connor, L., Sinnott, L. T., and Melnyk, B. M. [50] 2020 USA Individuals who attended the EBP immersion program
  • -

    94.3% nurses

  • -

    5.7% non-nurses

Longitudinal pre-experimental design Academic medical center (36.3%), health systems (25.5%), and community hospital (19.3%) n = 400
  • -

    Mean age = 42.0 (SD = 11.0)

Not reported
  • -

    Effects of a 5-day EBP immersion program

  • -

    EBP attributes, competencies, beliefs, implementation, and knowledge

  • -

    Perceptions of organizational culture and readiness for integration of EBP

  • -

    Demographic characteristics

5 Hart, P., Eaton, L.A., Buckner, M., Morrow, B. N., Barrett, D. T., Fraser, D. D., Hooks, D., and Sharrer, R. L. [51] 2008 USA RNs and licensed practical nurses (LPNs) A quasi-experimental design with a one-group pre–post-intervention survey design Integrated healthcare system in a southeastern state in the United States Pre-survey n = 744
Post-survey n = 314
  • -

    Mean age = 43.6 (SD = 10.3)

Women = 676
Men = 42
  • -

    Knowledge of EBP and research utilization

  • -

    Attitudes and skills in EBP and research utilization

6 Hsieh, P.-L. and Chen, S.-H. [52] 2020 Taiwan School nurses Quasi-experimental, pre–posttest design 193 primary schools in Tayuan and New Taipei City n = 401
  • -

    Mean age = 35.7 (min = 26, max = 58)

Women = 401
Men = 0
  • -

    Effectiveness of the multifaceted EBP training program

  • -

    Knowledge, attitudes, skills, and self-efficacy

7 Koota, E., Kääriäinen, M., Kyngäs, H., Lääperi, M., and Melender, H.-L. [65] 2021 Finland Emergency nurses RCT with a parallel group Four emergency departments at two university hospitals in Finland n = 80 (IG = 40 and CG = 40)
Completed study: n = 64 (IG = 34 and CG = 29)
Not reported
  • -

    Attitudes, knowledge, self-efficacy, skills, and behavior toward EBP

  • -

    Satisfaction with the EBP educational intervention

8 Levin, R. F. Fineout-Overholt E., Melnyk, B. M., Barnes, M., and Vetter, M. J. [8] 2011 USA Nurse managers and visiting staff nurses Two-group randomized controlled pilot trial with a repeated-measures design Community Health Services from the three boroughs of Queens, the Bronx, and Manhattan n = 46
  • -

    IG n = 22

  • -

    CG n = 24

Women = 46
Men = 0
  • -

    Demographic characteristics

  • -

    Nurses’ EBP beliefs and implementation behaviors

  • -

    Group cohesion

  • -

    Nurse job satisfaction

  • -

    Nurse productivity

  • -

    Nurse attrition/turnover rate

  • -

    Manipulation checks on the intervention

  • -

    Cost outcomes

9 Mena-Tudela, D., Gonzalez-Chorda. V.-M., Cervera-Gasch, A., Macia-Soler, M., and Orts-Cortés, M. I. [53] 2018 Spain Second-year bachelor’s degree nursing students Quasi-experimental before and after design University Jaume I n = 83
  • -

    Mean age = 21.6 (SD = 5.6)

Women = 70
Men = 13
  • -

    Knowledge, skills, and attitudes toward EBP

  • -

    Demographic characteristics

10 Moore, L. [54] 2017 USA RNs Quasi-experimental, before and after design RNs with a bedside practice in a healthcare facility n = 77
  • -

    Pre-test n = 197

  • -

    Post-test n = 134

Not reported
  • -

    Effectiveness of an EBP educational intervention

  • -

    EBP attitudes, knowledge, and skills

  • -

    Correlation between educational preparation and years of nursing experience in using EBP, EBP attitudes, EBP knowledge, and EBP skills

  • -

    Correlation between years of experience and using EBP, attitudes, knowledge, and skills

  • -

    Demographic characteristics

11 Moore, L. K. [55] 2018 USA RNs with bedside responsibilities Quasi-experimental pre-test–post-test design Regional hospital (nine counties in western Kentucky and two in southern Indiana) n = 197 Not reported
  • -

    EBP skills

  • -

    Practice, attitudes, knowledge, and skills related to EBP

  • -

    Effect of educational preparation and years of nursing experience

  • -

    RNs’ practice, attitudes, knowledge, and skills related to EBP

  • -

    Demographic characteristics

12 Singleton, J.K. [48] 2017 USA Doctor of Nursing Practice–Family Nurse Practice (DNP–FNP) Pre–post cohort study Primary care, hospital, nursing homes n = 89
Mean age = 47.0 (min = 29, max = 65)
Women = 80
Men = 9
  • -

    Comparison between the five cohorts

  • -

    Demographic characteristics

13 Underhill, M., Roper, K., Siefert, L. S., Boucher, J., and Berry, D. [56] 2015 USA Oncology RNs and advanced practice nurses Pre–post-test design Ambulatory oncology setting, Dana-Farber Cancer Institute n = 113 Not reported
  • -

    Barriers to involvement in EBP

  • -

    Factors affecting participation in EBP projects

  • -

    Feedback on how EBP can be better incorporated into nursing practice

  • -

    Oncology nurses’ beliefs and implementation factors affecting EBP

  • -

    Demographic factors associated with EBP beliefs and implementation

14 Waltz, L.-A., Munoz, L., Miller, R.-A., and Johnson, H.-W [61] 2022 USA RNs employed full-time in an acute care setting at a metropolitan hospital Quasi-experimental study with a pre-test–post-test design 179-bed metropolitan acute care hospital n = 30
Age = 40–49 years (n = 9, 30.0%)
Not reported
  • -

    Demographic information

  • -

    Self-perceptions of competencies in EBP understanding, ability, desire, and frequency

  • -

    Self-perceptions of barriers before and after the intervention

  • -

    Changes in knowledge and ability to use EBP

15 Wan, A. L. P. [57] 2017 USA RNs working in medical/surgical/telemetry units, IMCU, ICU, or another unit (staff nurse, charge nurse, quality manager, nurse manager, nurse educator) Pre-test–post-test quasi-experimental randomized design Nurses in a single-site county hospital in the San Francisco Bay Area n = 19
  • -

    IG n = 9

  • -

    CG n = 10

  • -

    Mean age EG = 40–49 (44.4%)

  • -

    Mean age CG = 30–39 (55.6%)

Women = 9
Men = 9
  • -

    Demographic characteristics

  • -

    Knowledge, beliefs, attitudes, and abilities in implementing EBP

  • -

    Correlations between knowledge, beliefs, attitudes, and abilities in implementing EBP

Note: CG = control group; EBP = evidence-based practice; IG = interventional group; HCPs = healthcare professionals; ICU = intensive care unit; IG = intervention group; RNs = registered nurses.

3.3. Methodological Quality of the Studies Retained

The methodological quality of the cohort study was poor (four stars) [48] (Table 2). The different domains of evaluation used in the quasi-experimental interventional studies also scored as moderate [50,51,52,53,54,55,56,57,60,61] (Table 3). Finally, the RCT studies were scored as “of some concern” regarding their methodological risk of bias [8,66,67,68,69] (Table 4).

Table 2.

Newcastle–Ottawa Scale (NOS) adapted for cohort studies.

Selection Comparability Outcome
Study Representativeness of Exposed Cohort Selection of Non-Exposed Cohort Ascertainment of Exposure Demonstration That Outcome of Interest Was Not Present at Start of the Study Adjusted for the Most Important Risk Adjusted for Other Risk Factors Assessment of Outcome Follow-Up Length Loss-to-Follow-Up Rate Total Quality
Singleton et al., 2017 [48] 1 * 0 * 1 * 1 * 0 * 0 * 0 * 1 * 0 * 4 *

Note: Studies were awarded a maximum of one star for each numbered item in the Selection and Outcome categories. A maximum of two stars could be awarded for Comparability. Studies were evaluated on a scale from 0 to 9 stars and classified as low (<6 stars), moderate (6–7 stars), or high (8–9 stars) quality. X * = X star.

Table 3.

ROBINS-I–The Risk Of Bias In Non-randomized Studies of Interventions assessment tool.

Pre-Intervention At Intervention Post-Intervention
Confounding Selection of Participants Classification of the Intervention Intended Interventions Missing Outcome Data Measurement of the Outcome Selection of the Reported Result
Study Effect of Assignment to Intervention Effect of Starting and Adhering to Intervention
Wan, 2017 [57] Low risk Low risk Low risk Low risk Serious risk Serious risk Serious risk Low risk
Chao et al., 2022 [60] Moderate risk Low risk Low risk Low risk Low risk Low risk Moderate risk Low risk
Gallagher-Ford et al., 2020 [50] Low risk Low risk Moderate risk Moderate risk Serious risk Low risk Serious risk Low risk
Hart et al., 2008 [51] Serious risk Low risk Moderate risk Moderate risk Serious risk Low risk Serious risk Low risk
Hsieh et al., 2020 [52] Serious risk Low risk Moderate risk Moderate risk Low risk Low risk Moderate risk Low risk
Mena-Tudela et al., 2018 [53] Serious risk Low risk Moderate risk Moderate risk Serious risk Low risk Serious risk Low risk
Moore et al., 2017 [54] Low risk Low risk Low risk Low risk Low risk Moderate risk Moderate risk Low risk
Moore et al., 2018 [55] Low risk Low risk Low risk Low risk Serious risk Serious risk Serious risk Low risk
Underhill et al., 2015 [56] Moderate risk Low risk Moderate risk Serious risk Serious risk Low risk Moderate risk Low risk
Waltz et al., 2022 [61] Serious risk Low risk Moderate risk Serious risk Serious risk Moderate risk Moderate risk Low risk

Table 4.

RoB 2.0: Revised Cochrane risk of bias tool for randomized trials–randomized trial methodologies.

Randomization Process Intended Interventions Missing Outcome Data Measurement of the Outcome Selection of the Reported Result
Study Effect of Assignment to Intervention (Part 1 and 2) Effect of Adhering to Intervention
Cardoso et al., 2021 [62] Low risk Some concerns
Low risk
High risk Low risk High risk Low risk
D’Souza et al., 2021 [64] Low risk Some concerns
Low risk
High risk Low risk High risk Low risk
Koota et al., 2021 [65] Low risk Some concerns
Low risk
High risk Low risk High risk Low risk
Levin et al., 2011 [8] Some concerns Some concerns
Low risk
High risk Low risk High risk Some concerns

3.4. Description of the Interventions

The studies retained had each developed a tailored educational intervention focusing on their sample’s specific needs. Those interventions were divided into three categories: (i) multifaceted educational strategies incorporating mentoring and tutoring [8,52,54,57,60,66,69], (ii) single educational strategies [55,58,59,61], and (iii) multifaceted educational strategies using the five steps (five As) of EBP [56,68,70].

3.4.1. Multifaceted Educational Strategies Incorporating Mentoring and Tutoring

The different frameworks used in this category of interventions were the Advancing Research and Clinical practice through close Collaboration (ARCC) model, the Classification Rubric for EBP Assessment Tools in Education (CREATE), and the critical incident technique. The key component of the ARCC model is the central place of the mentor, an advanced practice nurse who assists nurses in honing their EBP skills and implementing projects to improve patient care and outcomes. Furthermore, the mentor actively participates in implementing strategies to overcome barriers within the healthcare environment by building a culture of EBP [68]. Guided by cognitive-behavioral theory, ARCC strategies are designed to encourage nurses to develop stronger beliefs about the value of EBP and more confidence in their ability to implement it consistently [67,69]. Levin et al. [8] proposed a multi-component ARCC model intervention program. The intervention started by clarifying the definition of EBP and the rationale for using it in clinical decision-making. In a second phase, the interventional group (IG) was taught how to explore searchable scientific databases to respond to clinical questions, and then participants were asked to find evidence by exploring appropriate scientific databases, registers, and websites. They were encouraged to explore the basic concepts of a systematic review method, especially reading and critically appraising meta-analyses. For 12 weeks, they had 1.5 h of focused follow-up with an EBP mentor to teach them how to implement the concepts and skills of EBP to address a clinical problem in daily practice. On the other hand, the control group (CG) received didactic content about adult physical assessment dispensed by an expert in EBP. Underhill et al. [56] investigated the beliefs and implementation of EBP after an institutional EBP education intervention that took place over 24 months. This consisted of four components, starting with face-to-face discussions with nurses to provide an introduction to EBP and EBP resources. The content of the EBP training was oriented toward quantitative research designs using the PICOT research question strategy (clinical population, intervention, comparison, outcome, time). It also included explanations of the levels of evidence, differences in quality improvement, nursing research and evidence-based definitions and methods. An overview of the content was presented on an EBP poster for guiding EBP projects (SPAWN projects). In 2012, a nursing scholarship day was organized to raise nursing staff’s awareness about the importance of providing evidence-based care. Finally, an online educational module on EBP was available. During each EBP session, conducted by a member of the Evidence-Based Practice and Innovation Committee who took on the role of mentor, participants could ask questions.

Gallagher et al. [50] evaluated the effects of a five-day EBP continuing education and skills-building program, which was attended by a total of 400 participants, including 377 RNs, between September 2014 and May 2016. The primary outcomes measured were: EBP attributes and competencies, beliefs, implementation, and knowledge using the ARCC model.

The final study to use the ARCC model was Singleton’s cohort design study [48]. It investigated the pre–post effects of an EBP teaching program using the EBP belief and implementation scales across seven years and five cohorts of the Doctor of Nursing Practice (DNP) program. The DNP program includes two education sessions on “Evidence-based Practice Methods and Techniques” involving 180 h of theory. Students also benefitted from a faculty mentor and a mentor from a clinical practice improvement project.

The second framework used was CREATE, a means of classifying EBP learner assessment tools. CREATE suggests that EBP knowledge should be assessed cognitively, using paper and pencil tests, as knowledge defines a learner’s retention of facts and concepts. Another assumption is that EBP skills should be assessed using performance tests, as skills reveal the application of knowledge. The assessment categories are: (a) Reaction to the Education Experience, (b) Attitude, (c) Self-Efficacy, (d) Knowledge, (e) Skills, (f) Behaviors, and (g) Benefit to Patients. The first three categories are assessed using self-reporting/opinion; knowledge and skills are assessed using performance testing; behaviors are tested using activity monitoring; and benefits to patients are assessed through patient-oriented outcomes [70]. Cardoso et al. [62] implemented an EBP educational program in a Portuguese nursing school in 2018. The educational program for the intervention cohort (IC) taught models for thinking about EBP and explained systematic review types, review question development, searching for studies, study selection processes, data extraction methods, and data synthesis. The last three sessions presented active methods through mentoring to groups of 2–3 students. The CG participants underwent their education as usual (theory, theory–practice, practice) delivered by the nursing educators.

Koota et al. [65] evaluated the effectiveness of an educational intervention on EBP for emergency nurses at two hospitals in Finland. The IG received the “Evidence-Based Practice Basics for Emergency Nurses” intervention, which included multifaceted educational strategies such as didactic lectures and discussions, small group tutorials, database search workshops, and a self-directed learning module. They also received support from the tutor teams, two emergency unit members (the researcher and a clinical nurse specialist), and a librarian for consultations. The CG completed a self-directed learning module entitled “Evidence-Based Practice Basics for Emergency Nurses”.

The study conducted by Mena-Tudela et al. [53] used the critical incident technique (CIT). The educational intervention was composed of two sessions: the first defined terms related to EBP and reflected on materials and using critical thinking, whereas the second involved practical exercises on information literacy. During their 12-week clerkship, student nurses had to identify a minimum of eight critical incidents, develop a PICO clinical question, and try to resolve the incident using a literature search. A lecturer was available to provide support and tutor feedback to groups of 8–10 students.

Different authors used the terms of mentor, EBP mentor, or faculty mentor [8,52,60,66], whereas others used tutoring or tutor teams. As Melnyk et al. stated [71], mentors work with nurses and are usually themselves advanced practice nurses (clinical nurse specialists or nurse educators) with superior knowledge and skills in EBP and organizational and individual behavior change. Additionally, some authors specified that mentors are opinion leaders within teams [7,21]. Mentoring involves a long-term relationship between two people, one of whom is usually older and/or more experienced than the other [72]. The aim of this relationship—based on mutual respect and compatible personalities—is to guide the mentee towards personal and professional growth [73]. Thus, a mentor’s role in an educational intervention is to provide support and social skills to the care teams involved and, more specifically, to assess how well they have acquired the skills taught. The tutoring described by Mena-Tudela et al. [53] was provided by nurse instructors, which is the traditional model of nurse tutoring [74]. Tutoring is a system of partnership throughout a learning process, placing the commitment and responsibility for the teaching and learning experience on the learner [75]. Even though the definitions and purposes of mentoring and tutoring are quite similar, tutoring requires the learner to be more proactive in the acquisition of new learning skills and does not require a one-to-one relationship, whereas mentoring does not always have a formal goal and does imply a one-to-one relationship.

3.4.2. Single Educational Strategies

Three [55,58,59,61,65] of the five studies classified in this second category used a computer-based education (CBE) strategy, the fourth used face-to-face and videoconferencing sessions [61], and the fifth used the Promoting Action on Research Implementation in Health Services Model (PARiHS) and Plan-Do-Check-Act (PDSA) frameworks [57]. Using computer-based education in the teaching process [76] is an alternative method of providing information in a reliable, uniform manner [77]. In the Hart et al. [51] study, education on EBP was delivered through three computer-based learning modules. Module 1 helped students develop a clinical question, module 2 trained their ability to read and understand a research article, and module 3 demonstrated how to transform theoretical evidence into nursing practice. The studies by Moore [54,55] delivered an online educational intervention on EBP using Rogers’ Diffusion of Innovation Theory. The IG studied a computer-based module composed of a self-paced PowerPoint presentation addressing the relevance of EBP, the steps of EBP, the resources available to facilitate the implementation of EBP, and examples of EBP. Control group 1 was assigned a computer-based PowerPoint learning module on pain management, with a delivery format and time available similar to the intervention group. Finally, control group 2 underwent no intervention. Wan’s study [57] was guided by the PARiHS and PDSA frameworks. The EBP education program came in four units of 2–4 lessons each and was delivered through lectures with printed handouts, small group discussions, and questioning and writing exercises. The components were the EBP process, the PICOT model, literature search skills, fundamental research critiques, levels of evidence, quality improvement strategy, and the PDSA rapid cycle method. Finally, the educational intervention in the study by Waltz et al. [61] consisted of a practice and an interactive workshop (of 1.5 h each) to improve EBP competencies. These were conducted by Ph.D. nurse educators. The research topic was an example selected because of its relevance to their care environment.

3.4.3. Multifaceted Educational Strategies Using the Five Steps of Evidence-Based Practice

The last three studies [56,64,67,70] retrieved were guided by the five basic steps of EBP: (a) Ask a question; (b) Search for information or evidence to answer the question; (c) Critically appraise that information or evidence; (d) Integrate the appraised evidence using clinical expertise and patients’ preferences; and (e) Evaluate [78]. The intervention proposed by Hsieh and Chen [52] consisted of an evidence-based school-nursing program that used multiple pedagogical strategies, such as a traditional in-class format, a combination of teaching strategies, a flipped classroom format, group activities, case studies and supervision, mobile learning technology, and online lessons. The RCT conducted by D’Souza et al. [64] involved a 30 h EBP training intervention over four days at a nursing education institution. Different strategies were used, such as lectures and discussions, small-group activities, critical-thinking exercises, scenario-based discussions, online library training, brainstorming sessions, critical appraisal sessions, activities on integrating EBP, EBP booster sessions, or sessions on literature searches. Finally, Chao et al. [60] developed an intervention consisting of two pedagogical strategies: a flipped e-learning intervention and traditional classroom teaching.

3.5. Clinical Outcomes, Instruments, and Results

3.5.1. Primary Outcomes

The 15 studies measured 11 primary outcomes—EBP beliefs, EBP self-efficacy, EBP attitudes, perceived EBP implementation, EBP competencies, EBP knowledge, EBP skills, EBP behaviors, EBP desire, EBP practice, and perceptions of organizational culture and readiness (EBP attributes)—using 15 qualitative and quantitative instruments.

  • (i)

    Evidence-Based Practice Belief Scale (EBP-B)

Six studies [8,52,54,60,61,69] used this instrument. In their longitudinal, pre-experimental study, Gallagher et al. [50] initially showed increasing mean scores from T0 to T1 (T0: M = 58.4 (SD = 8.7), T1: M = 68.1 (SD = 8.1)), and then decreases from T1 to T2 (M = 66.3 (SD = 9.1) and increases from T2 to T3 (M = 68.1 (SD = 7.4)). These scores were statistically significantly different between T0 and T1 (p < 0.001), between T0 and T2 (p < 0.001), and between T0 and T3 (p < 0.001).

In the study by Underhill et al. [56], oncology RNs and advanced practice nurses completed this quantitative tool at T1 and T2. The EBP-B median at T1 was 56.5 (range = 37–77, IQR1–3 = 50–61) and at T2 was 57 (range = 38–76, IQR1–3 = 51–63). There were no significant differences between T1 and T2 implementation scores or between those who had or had not participated in the SPAWN projects (p = 0.36). Finally, the level of nursing education was positively correlated with EBP-I (r = 0.32, p = 0.01), unlike years employed as a nurse (p = 0.16). The study by Levin et al. [8] demonstrated a major effect for its IG (F[1,15] = 10.39, p = 0.006), with nurses showing better results in EBP at T3 and T4. There was also a main effect of time (F[2,30] = 5.85, p = 0.007) since those authors noted a significant increase from T1 to T3 (T1: M = 12.89 and T3: M = 28.14; F[1,15] = 12.40, p = 0.003). Finally, there was also a significant time–group status interaction (F[2,30] = 3.625, p = 0.0039) since the IC showed a higher mean score on EBP implementation than the CG (M = 15.50 versus 10.33). These results were consistent with different studies [57,65]. Indeed, the CG’s emergency nurses revealed constant attitudes toward EBP throughout their follow-up. Nurses from the IG showed improved attitudes toward EBP at T2, with a decrease at T3 (M = 4.3 (SD = 0.23) versus M = 4.12 (SD = 0.49)), although still higher than at T0 (M = 3.85 (SD = 0.49)).

  • (ii)

    Perceived Evidence-Based Practice Implementation (EBP-I)

Five studies used the EBP-I, whereas two studies used the Adapted EBP-I. In Singleton’s study [48], DNP–FNP students showed a mean pre-intervention score of 2.68 (SD = 0.94) versus 3.61 (SD = 0.96) post intervention. Data analysis showed this to be statistically significant (t = 8.4 (52), p < 0.001), with an effect size ranging from 0.75 to 1.5 SD above the mean. Finally, gains were made by every cohort and across the curriculum (F ratio 20.01). Underhill et al. [56] asked oncology RNs and advanced practice nurses to complete the EBP-I at T1 and T2. No statistical differences were found between the scores of nurses who had and had not participated in SPAWN projects (p = 0.36). However, nurse leaders had higher perceived scores (p < 0.001). Finally, the level of nursing education was positively correlated with EBP-I (r = 0.32; p < 0.01). These results are consistent with Wan’s study [57]. Nurses in the IG who underwent the EBP education program had lower mean rank scores than CG nurses in the pre-test but higher ones in the post-test. Their scores increased between T0 and T1 (p = 0.025), unlike the CG (p = 0.257). Mean scores across the 18 implementation items used by Gallagher et al. [50] increased over time, although no data were reported for T1 (T0: M = 15.4 (SD = 13.9), T2: M = 20 (SD = 12.7), and T3: M = 25.4 (SD = 15.6)). Statistically significant differences were found between T0 and T2 (p = 0.0087) and between T0 and T3 (p > 0.0001). These results are consistent with Levin et al. [8]. The interventional group showed significantly higher scores from T1 to T3 (T1: M = 12.89, T3: M = 28.14) (F[1,15] = 12.40, p = 0.003). However, the study by Koota et al. [65] revealed that the IG’s emergency nurses’ EBP scores had improved at T2 but had decreased again by T3. Some studies showed conflicting findings on how EBP behaviors decreased following interventions [54,55], and the EBP scores of nurses in the CG had dropped below baseline levels by T3.

  • (iii)

    Evidence-Based Competencies Scale

Gallagher et al. [50] assessed nurses’ EBP competencies using the Evidence-Based Competencies Scale. Participants completed the scale at four time points, and mean scores for the 24 competency items increased over time (T0: M = 53.1 (SD = 18), T1: M = 75.7 (16.3), T2: M = 77.3 (SD = 17.7), and T3: M = 85 (SD = 19.9)). Statistically significant differences were found between T0 and T1, T0 and T2, and T0 and T3 (p < 0.001).

  • (iv)

    Evidence-Based Practice Knowledge Assessment Questionnaire (EBP-KAQ)

The EBP-KAQ was developed in 2015 (Gallagher-Ford et al.) but remained unpublished. Gallagher et al. [50] reported mean scores for the 38 EBP-KAQ questions increasing over the four time points (T0: M = 24 (SD = 6.1), T1: M = 30.2 (SD = 3.9), T2: M = 31.2 (SD = 3.5), T4: M = 31.7 (SD = 4.2)). Statistically significant differences were found between T0 and T1, T0 and T2, and T0 and T3 (p < 0.0001), providing strong evidence of a positive effect on participants.

  • (v)

    Adapted Fresno Test

Two studies used the Adapted Fresno test to evaluate the knowledge, skills, and EBP competency of nurse educators. In the study by Cardoso et al. [62], both the IG and the CG completed the Adapted Fresno Test at T0 and T1. The mean IG score improved from baseline (M = 6.85 (SD = 5.15)) to T1 (M = 12.47 (SD = 7.21)), as it did for the CG (T0: M = 7.25 (SD = 5.34), T1: M = 9.73 (SD = 5.56)). These improvements over time were statistically significant for both the IG and the CG (F(1.73) = 53.028, p < 0.001, and F(1.73) = 13.832, p < 0.001). Finally, students in both groups significantly improved their knowledge and skills in four items (1, 3, 4, 5), only IG improved in item 7 (Z = −2.543, p = 0.011 versus CG: Z = −1.941, p = 0.052), and only CG improved in item 7 (Z = −2.714, p = 0.007 versus IG: Z = −1.236, p = 0.216). D’Souza et al. [64] asked nurse educators to complete the Adapted Fresno Test at T0, T1, and T3. The mean IG score improved from T0 to T2 (M = 20.30 (SD = 13.13) to M = 103.45 (SD = 27.87)), whereas mean CG scores only improved a little (T0: M = 24.67 (SD = 11.39) and T2: M = 28.78 (SD = 10.86)). The differences between the IG and CG indicated that the EBP training program improved EBP competency (F[2,92.06] = 37.13, p < 0.05).

  • (vi)

    Evidence-Based Practice Questionnaire-FI (EBPQ-FI) or Evidence-Based Nursing Questionnaire

Koota et al. [65], D’Souza et al. [64], Wan [57], and Moore [54,55] all used the EBPQ. Hart et al. [51] used the same instrument but named it the Evidence-Based Nursing Questionnaire.

In Moore’s studies [54,55], RNs rated themselves moderate in attitudes to EBP (M = 5.28 (SD = 1.02)) and low in use (M = 4.80 (SD = 1.23)) and knowledge and skills (M = 4.83 (SD = 0.99)). Koota et al. [65] used the EBPQ to assess behaviors, attitudes, knowledge, and skills at four time points. At T2, the IG showed statistically significant improvements in EBP attitudes (p = 0.019) and knowledge (p = 0.005). Meanwhile, EBP behaviors had decreased among the CG. At T3, there was a statistically significant difference between the groups’ EBP attitudes (p = 0.010). EBP attitudes remained constant for the CG across the whole follow-up period, with an improvement at T2 but then a decrease at T3. The IG showed statistically significant improvements in their knowledge of EBP at every time point (p < 0.001, p < 0.001, and p < 0.001). Finally, the two CGs showed lower scores for EBP behaviors, whereas they improved at T2 and then decreased at T3 for the IG. Finally, the IG’s EBP skills showed a minor improvement at T2 (T0: M = 4.64 (SD = 0.76), T2: M = 5.05 (SD = 0.42)). In the study by Wan, the IG had lower mean ranks for attitudes to EBP than the CG in the pre-test scores but higher ones in the post-test. The IG did not increase attitude scores between T0 and T1, as CG did. D’Souza et al. [64] used the EBPQ to assess knowledge and use of EBP. Both groups’ knowledge increased gradually from T1 to T2. The 30 h EBP training intervention was effective in improving nurse educators’ knowledge (F[2,91.65] = 4.11, p < 0.05). After T1, the IG showed a greater increase in EBP use scores than the CG. IG nurse educators showed statistically significant improvements in their EBP use scores (F[2,94.88] = 6.21, p < 0.05). RNs in the Hart et al. [51] study completed the Evidence-Based Nursing Questionnaire before and after computer-based education sessions. Statistically significant pre- and post-intervention differences were found for knowledge of EBP (t (0.05, 312) = –2.296, p = 0.022), attitudes to EBP (t (0.05, 313) = –4.750, p < 0.001), skills in EBP (t (0.05, 313) = –4.322, p < 0.001), and organizational readiness for EBP (t (0.05, 313) = –8.601, p < 0.001) and its use in research.

  • (vii)

    Evidence-Based Practice Competence Questionnaire (EBP-COQ)

Mena-Tudela et al. [53] used the EBP-COQ to evaluate self-perceived levels of EBP competence among Spanish nursing students. Nearly all the items showed significant differences between the three measurement points, except for item 1 (p = 0.099), item 4 (p = 0.051), item 7 (p = 0.055), item 10 (p = 0.065), and item 11 (p = 0.441). Finally, statistically significant overall differences between the three time points were measured (p < 0.001), between T0 and T1 (p < 0.001) and between T0 and T2 (p < 0.001).

  • (viii)

    School Nurse Evidence-Based Practice Questionnaire (SNEBP)

Hsieh and Chen [52] used the SNEBP at three different time points. Knowledge section scores had increased by an average of 0.75 at post-training (p < 0.001) and by 0.37 at follow-up (p < 0.001). Self-efficacy section scores had increased by an average of 0.3 at post-training (p < 0.001) and 0.13 at follow-up (p < 0.001). Skills section scores significantly increased from baseline (M = 3.73; SD = 0.58) to post-training (M = 3.80; SD = 0.62) (p < 0.001), but they then significantly decreased (by an average of 0.1) between post-training and follow-up (p > 0.005). Finally, attitude section scores made no statistically significant changes over time.

  • (ix)

    Organizational Culture and Readiness for System-Wide Integration of Evidence-Based Practice Scale

Gallagher et al. [50] used the Organization Culture and Readiness for System-Wide Integration of Evidence-Based Practice Scale. This assesses cultural factors that can influence system-wide implementation of EBP and overall perceived readiness for EBP integration. The mean score for the 24 competency items increased across the three time points as no values were reported for T1 (immediately after the intervention) (T0: M = 77 (SD = 16.1), T1: M = 75.7 (16.3), T2: M = 80.4 (SD = 15.4), or T3: M = 84.6 (SD = 15.9)). Statistically significant differences between T0 and T2 (p = 0.0041) and between T0 and T3 (p > 0.0001) were described.

  • (x)

    Twenty-one-item scale

D’Souza et al. [64] developed a 21-item for their study in order to respond to the particularities of their population and assess their attitudes to EBP. Attitude scores were higher in the IG across time points T1 and T2. Significant changes in the attitudes of the IG nurse educators (F[2,91.07] = 3.55, p < 0.05) were described.

  • (xi)

    Thirty-item questionnaire

Chao et al. [60] used a 30-item questionnaire adapted from Lee et al.’s (2011) study [79]. It included the three dimensions of knowledge, attitudes, and practice. For the knowledge subscale, there was no significant difference in the two groups’ mean scores at T1 and T2 (p = 1.00 and p = 0.702). Both groups showed improved knowledge after the intervention. Both groups’ attitude scores (CG: p = 0.000; IG: p = 0.000) and practice scores (CG: p = 0.000; IG: p = 0.000) also improved after the intervention, and there were no differences between the two groups at the three different time points (T1: p = 0.573; T2: p = 0.792; T3: p = 0.153).

  • (xii)

    Healthcare EBP Assessment Tool (HEAT)

Waltz et al. [61] used HEAT to measure self-perceived competencies in EBP, including understanding, ability, desire, frequency, and barriers before and after the intervention. This consisted of an interactive workshop of 1.5 h each. The pre-test and post-test scores showed improvements for the EBP subscales of frequency (except for “Used research findings in practice”: p = 0.179), ability (except for “Participate in a research project”: p = 0.066), desire, and barriers (only “I have difficulty finding research or library materials with my searches” and “I have difficulty understanding research articles” were statistically significant; p = 0.019 and p = 0.035, respectively).

  • (xiii)

    Qualitative analysis of the monographs

Cardoso et al. [62] qualitatively analyzed 18 students’ monographs using an evaluation form with 11 guiding criteria (review questions, inclusion and exclusion criteria, methodology, the presentation of results, and the congruence between the review questions and the conclusion’s answers). The CG’s monographs were of a lower quality than the IG’s since the CG displayed difficulties synthesizing the data and providing clear answers to the review questions.

  • (xiv)

    Tailored satisfaction questionnaire

Underhill et al. [56] aimed to explore barriers to and facilitators of implementing EBP. The most common barriers experienced by participants were a lack of time, knowledge, and access to online journals or databases. That is why oncology nurses requested more time, resources, education, and awareness about EBP. On the other hand, the facilitating factors identified were nursing leadership and leaders from nursing research. Chao et al. [60] also used a questionnaire to discover levels of satisfaction with the learning process and course, with ratings from 1 (strongly disagree) to 5 (strongly agree). Higher scores indicated greater satisfaction in the learning process and course. There were no significant differences in the scores between the two groups (p = 0.001; CG: M = 38.49, SD = 0.99 and IG: M = 44.24, SD = 0.60). The IG showed significantly higher scores for each item and overall satisfaction than the CG. The item with the lowest score for the CG was “I could understand EBP through the course content” (M = 3.75, SD = 1.15), and the highest score was for the item “Practical experience improved my ability to search for empirical literature” (M = 3.96, SD = 0.96). The item with the lowest score for the CG was “The workshop helped me complete the EBP 5As and apply them to clinical care” (M = 4.35, SD = 0.52). The most favorably rated items were “I could understand EBP through the course content” (M = 4.46, SD = 0.66) and “The course strengthened my ability to search for empirical literature (M = 4.46, SD = 0.50).

3.5.2. Secondary Outcomes

Only two studies evaluated secondary outcomes, but they used eight instruments. Levin et al. [8] used a qualitative and/or quantitative approach to deepen their investigation of five secondary outcomes: group cohesion, job satisfaction, nurse productivity, nurse attrition/turnover, and professionalism and leadership. Finally, Koota et al. [65] examined learner satisfaction using instruments developed by those authors specifically for their study.

  • (i)

    Group Cohesion Scale

Levin et al. [8] used the Group Cohesion Scale developed by Good and Nelson in 1973 [80]. Scores for group cohesion were examined according to time point and group as well as time point–group interactions. No main effects were identified for group (F[1,16] = 0.016, p = 0.502), time point (F[2,32] = 2.50, p = 0.098), or time point–group interactions (F[2,32] = 0.824, p = 0.448). However, the IG did report better group cohesion at T3 than at T1 (F[1,21] = 5.580, p = 0.028).

  • (ii)

    Index of Work Satisfaction (IWS)

Levin et al. [8] use the IWS to measure work satisfaction according to time point, group, and time point–group interaction. Statistical analyses showed no main effects for group (F[1,22] = 0.057, p = 0.814) or time point–group status interaction (F[1,31] = 0.392, p = 0.537). However, it did show a main effect for time (F[1,22] = 0.016, p = 0.900).

  • (iii)

    Visiting Nurse Service of New York

Levin et al. [8] assessed the nurse productivity of the VNSNY. No main effects for group (F[3,52] = 0.422, p = 0.738) or time point (F[1,52] = 2.799, p = 0.100) were described. The overall mean productivity rate at T1 (M = 4.59) did not differ significantly from the rate at T3 (M = 4.59 versus M = 4.77). Finally, no significant time point–group status interaction (F[3,52] = 0.243, p = 0.866) was reported.

  • (iv)

    Visiting Nurse Service of New York–Human Resources

Levin et al. [8] used the VNSNY–Human Resources database to evaluate nurse attrition. In the IG, attrition rates decreased from 11% (n = 5) in 2004 to 5.7% (n = 3) in 2005, whereas the attrition rate in the CG was 35%.

  • (v)

    Focus group

Levin et al. [8] conducted focus groups to explore nurse managers’ and visiting staff nurses’ professionalism and leadership. No more details were given. These exchanges with study participants highlighted their greater sense of professionalism and increased ability to see managers as colleagues.

  • (vi)

    Learning questionnaire

Levin et al. [8] compared the intervention’s learning effects on the IG and the CG at T1 and T2 via a learning questionnaire. The IG answered more EBP questions correctly than the CG, but differences were not statistically significant.

  • (vii)

    Knowledge test and course assessment form

Koota et al. [65] developed these instruments because they could find no validated tools appropriate for their study. Groups differed significantly in their satisfaction with the teacher’s ability to encourage nurses to ask clinical questions (IG: M = 2.98 (SD = 0.62) versus CG: M = 2.75 (SD = 0.93)) (p > 0.001). In addition, groups significantly differed in their satisfaction with the usefulness of considering research evidence in nursing practice (IG: M = 3.65 (SD = 1.06) versus CG: M = 2.94 (SD = 0.83)) (p = 0.01).

4. Discussion

In increasingly complex healthcare systems, and given the inherent challenges of healthcare environments, it is becoming essential for nurses to have a solid foundation in EBP because of its proven links to improved patient and organizational outcomes [61]. This systematic review identified and summarized the most effective educational interventions to increase skills in EBP among nurses. It identified three types of key educational interventions: multifaceted educational strategies incorporating mentoring and tutoring, single educational strategies (often delivered online), and multifaceted educational strategies using the five basic steps of EBP. These pertinent findings will provide tangible support to researchers, clinicians, healthcare system managers, healthcare institutions, health policymakers, stakeholders, universities and their faculty members, and healthcare professionals because they reveal an assessment of the effectiveness of different educational interventions on eleven primary outcomes. The review covered studies involving 2712 RNs and licensed practical nurses (LPNs), bachelor’s degree students, Ph.D. students in Nursing Practice–Family Nurse Practice, nurse educators, emergency nurses, nurse managers, and visiting staff nurses.

EBP educational interventions should be a standard part of nurses’ professional development in clinical settings. Healthcare systems and nurses have been encouraged to adopt the principles of EBP in order to improve the delivery of safe, high-quality healthcare to patients. We might expect EBP to be well known and understood by nurses—and equally prevalent in their education programs and clinical practice—but although they generally have positive beliefs about EBP, many individual and organizational barriers remain [48], impeding them from turning their beliefs into action. These barriers include the lack of time, deficiencies in knowledge about EBP, access to evidence, autonomy, empowerment to change practice, resistance from colleagues or managers [57,60,68], and even inadequate numbers of EBP mentors [50]. To ensure EBP is used daily, it is necessary to identify the most efficient and effective EBP educational program to improve EBP beliefs, EBP self-efficacy, perceived EBP implementation, EBP competencies, EBP knowledge, EBP skills, EBP attitudes, EBP behaviors, EBP desire, EBP practice, and perceptions of organizational culture and readiness, to allow individuals, leaders, and healthcare organizations to build a culture and environment of EBP, and to make EBP’s adoption sustainable [35].

“EBP knowledge and skills can be acquired through formal and continuing education and self-paced online tutorials” [81]. Indeed, the 15 studies included in this systematic review present different strategies to implement the interventions described. Although the three different types of interventions identified had significant positive impacts on the 11 outcomes mentioned above, EBP cannot occur in isolation: it requires early adopters, champions, and teamwork. According to our findings, the best way to increase EBP skills among nurses and facilitate EBP implementation in their daily practice seems to be the use of a variety of educational strategies that integrate multiple learning pathways and techniques with regular follow-up and feedback from mentors and/or leaders. Computer-based education (CBE) currently appears to be the most cost-effective and efficient strategy, providing reliable, consistent delivery of information and the flexibility for caregivers to choose where and when to access the intervention. This type of educational strategy ensures that EBP’s adoption by nurses is sustainable in their daily practice.

Not all RNs have undergone training on EBP during their educational or career trajectory. Thus, it is important to develop educational strategies for both stages so as to standardize and homogenize EBP knowledge and skills: (1) during their student curricula, and (2) regularly, on site, where they practice. Indeed, mere theoretical input is not enough, and educational interventions in clinical settings are essential to promoting the use of EBP. The literature concludes that nurses’ EBP competencies are only low to moderate internationally, particularly in terms of their EBP knowledge, EBP skills, and their confidence in employing EBP. One potential explanation is the heterogeneity in the quality and content of EBP teaching programs in students’ curricula [29]. It is therefore important to homogenize nurses’ knowledge and skills while they are still in training. Only once nurses are competent in EBP will they be more likely to engage in EBP in their daily practice [11].

In addition to using the different strategies mentioned above, researchers must also examine the variables that might support or hinder the implementation of EBP, using reliable and valid measures [7,82]. In this context, the Consolidated Framework for Implementation Research (CFIR), developed by Damschroder et al. [82,83], sheds light on a wide range of factors that can support or hinder implementation. The CFIR is itself the result of a literature review; it synthesizes all the existing implementation models, bringing together the elements facilitating or constraining the implementation of new practices described in the literature. It is subdivided into five areas: (i) the external environment, which analyzes, among other things, cosmopolitanism; (ii) the internal environment, which looks at structural characteristics and culture; (iii) the personal characteristics that influence the course of a project; (iv) the characteristics of the intervention; and (v) the implementation process, at both the individual and organizational levels. According to its developers, the CFIR can be used at three time points. Firstly, during pre-implementation, to conduct a baseline survey (nurses’ sociodemographic characteristics, such as level of education, were positively correlated with EBP-I (r = 0.32, p = 0.01) in the studies by Gallagher et al. [50], Underhill et al. [56], and Wan [57]) including the medical specialty’s specific culture, leadership style, openness to change, learning climate, communication about the project, and an identification of its implementation process. Educational interventions are only one part of advancing EBP. Indeed, implementing EBP is also affected by nurses’ demographic characteristics, such as age [84], marital status [85], years of work experience, or educational level [86], as well as their learning environment [87], staffing levels, and leadership styles [88]. Secondly, during peri-implementation, CIFR can be used to enable monitoring (assessing the characteristics of an intervention, such as its clarity or adaptability, adapting the implementation process through PDSA cycles, regular communication, and feedback, responding to the specialty’s needs, and mobilizing the right leadership style). Thirdly, it can be used post implementation to ensure sustainability—including continuous adaptation based on feedback and active involvement from the specialty, regular communication, and the identification of mentors and/or opinion leaders [7,17]—and avoid a drop in buy-in as was observed in the study by Koota et al. [65]. In addition to the CFIR, there are other models that can guide the implementation of EBPs, such as Promoting Action on Research In Health Services Framework (PARiHS) [89], the Iowa Model of Evidence-Based Practice to Promote Quality Care [90], Advancing Research and Clinical practice through close Collaboration (ARCC) model [91], the JBI Model of Evidence-Based Healthcare [92], Johns Hopkins Nursing Evidence-Based Practice, or even the Clinical Scholar Model [93].

4.1. The Review’s Limitations

There are some limitations to this review that should be considered. First, despite a thorough literature search using recognized methodological guidelines and recommendations, and the use of many different terms to describe EBP, our review may have missed some studies that met all the selection criteria, either due to study search errors or investigators’ mistakes. Second, language and publication bias may be present despite the review’s scope being worldwide. Third, the effects of different types of healthcare system, systems of professional training, urban or rural settings, and socioeconomic factors, among others, were not analyzed. Finally, this review originally aimed to answer the question, “How effective are educational interventions to increase skills in EBP among nurses and physiotherapists delivering primary healthcare?” Despite a literature search that took place at three distinct time points (2019, 2021, and 2022), no relevant studies involving PTs were found. The study’s results can therefore only be generalized in relation to our sample’s heterogeneous nursing characteristics (RNs, LPNs, bachelor’s degree students, Ph.D. students in Nursing Practice–Family Nurse Practice, nurse educators, emergency nurses, nurse managers, and visiting staff nurses). However, the Gallagher-Ford et al. study goes beyond this sample to include 23 healthcare professionals [52]. Nevertheless, one of the review’s strengths is that it included four RCTs, despite their poor methodological quality, eight quasi-experimental studies, and one cohort study, all of which used validated instruments when measuring study outputs. Furthermore, we used highly recommended methodological norms and guidelines, which render our findings extremely reliable.

4.2. Recommendations

In view of the small number of studies published in our field of interest to date, future research should consider whether the present findings are applicable to other samples. Indeed, there is a need to include other primary HCPs to identify which interprofessional educational interventions will ensure that they too can provide the best available evidence-based healthcare. With a view to ensuring continuous improvements in safe, high-quality healthcare for patients, more research data on the most efficient, effective educational interventions for meeting HCPs’ needs and ensuring that EBP’s added value is delivered at the patient’s bedside also need to be collected. More data will also help to establish guidelines for EBP implementation. Finally, evidence from more high-quality studies involving larger sample sizes of HCPs would be welcome to test education interventions on EBP.

Our recommendations for implementing EBP in clinical practice and ensuring that implementation succeeds require a consideration of the specific characteristics and needs of the caregivers involved, as well as the uniqueness of the clinical setting, medical specialty, and the educational intervention before, during, and after implementation. Indeed, effective implementation of EBP comes through repeated exposure to and practice of EBP. This systematic review’s findings emphasize that strategic leadership, communication, and teaching strategies for EBP should rely heavily on the added value created by the mentors/tutors who informally and/or hierarchically take clinical leadership among their peers. However, there is a need to clearly identify the facilitators of and barriers to implementing educational interventions in healthcare settings. Our recommendations for future research also include analyzing and comparing the factors facilitating and constraining the pre-, peri-, and post-implementation phases of an educational intervention about EBP. This will help RNs, other HCPs, and nurse educators to adjust and improve their knowledge of implementation processes. That future research should use the CFIR during all three phases. It would also be important to consider whether different types of professional training, urban or rural settings, or socioeconomic factors had any impact on the effectiveness of educational interventions. This could be carried out during the CFIR’s pre-implementation phase.

5. Conclusions

To the best of our knowledge, this systematic review is the first on this topic to combine samples of RNs and LPNs, bachelor’s degree students, Ph.D. students in Nursing Practice–Family Nurse Practice, nurse educators, emergency nurses, nurse managers, and visiting staff nurses. Nurses corroborated the importance of evidence-based practice (EBP), but very few of them were able to implement it due to personal (lack of confidence, knowledge, and time, plus heavy workloads) and organizational (lack of material, human resources, support from the hierarchy, and implementation models) barriers. Systematically integrating best evidence into daily practice is challenging due to the inherent complexity of the EBP implementation process.

In conclusion, promoting the appropriate organizational culture and continuously developing the infrastructure, resources, and administrative support needed to advance the use of EBP is itself a process of moving from evidence to practice. However, this is not enough. Indeed, this paper highlighted the significant influence that educational strategies can have on the ability to master and implement EBP. The findings demonstrated that appropriate educational strategies could help EBP become a standardized part of the nursing student curriculum and of RNs’ daily practice. The most efficient, effective educational interventions will meet professionals’ needs and ensure that EBP’s added value is delivered at the patient’s bedside. These ongoing interventions seem to be relevant to continuing education based on multifaceted learning strategies with regular follow-up and feedback from mentors and/or leaders. Computer-based education is the most cost-effective and efficient strategy.

Acknowledgments

Collaboration with Tschopp Marielle, physiotherapist.

Author Contributions

Study design and concept: H.V., O.P.D.S., F.P., R.H., K.G. and E.P.; writing the study protocol: all authors; data acquisition: O.P.D.S., P.M., H.V. and F.P.; data analysis and interpretation: O.P.D.S. and H.V.; manuscript drafting: all authors; critical revisions of the manuscript for important intellectual content: all authors; statistical analysis: O.P.D.S. and H.V. All authors have read and agreed to the published version of the manuscript.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest. The funder was not involved in protocol design or the plan of analyses, nor will it be involved in the interpretation or publication of the systematic review’s results.

Funding Statement

This systematic review is funded by the Health Institute of the University of Applied Sciences and Arts Western Switzerland (funding reference number: Sagex 89265). This funding will support data collection and analysis.

Footnotes

Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

References

  • 1.Solomons N.M., Spross J.A. Evidence-based practice barriers and facilitators from a continuous quality improvement perspective: An integrative review. J. Nurs. Manag. 2011;19:109–120. doi: 10.1111/j.1365-2834.2010.01144.x. [DOI] [PubMed] [Google Scholar]
  • 2.Melnyk B.M., Fineout-Overholt E., Stillwell S.B., Williamson K.M. Evidence-based practice: Step by step: The seven steps of evidence-based practice. AJN Am. J. Nurs. 2010;110:51–53. doi: 10.1097/01.NAJ.0000366056.06605.d2. [DOI] [PubMed] [Google Scholar]
  • 3.Halfon P., Staines A., Burnand B. Adverse events related to hospital care: A retrospective medical records review in a Swiss hospital. Int. J. Qual. Health Care. 2017;29:527–533. doi: 10.1093/intqhc/mzx061. [DOI] [PubMed] [Google Scholar]
  • 4.Brown C.E., Wickline M.A., Ecoff L., Glaser D. Nursing practice, knowledge, attitudes and perceived barriers to evidence-based practice at an academic medical center. J. Adv. Nurs. 2009;65:371–381. doi: 10.1111/j.1365-2648.2008.04878.x. [DOI] [PubMed] [Google Scholar]
  • 5.Dawes M., Summerskill W., Glasziou P., Cartabellotta A., Martin J., Hopayian K., Porzsolt F., Burls A., Osborne J. Sicily statement on evidence-based practice. BMC Med. Educ. 2005;5:1. doi: 10.1186/1472-6920-5-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6.Melnyk B.M., Fineout-Overholt E., Gallagher-Ford L., Stillwell S.B. Evidence-Based Practice, Step by Step: Sustaining Evidence-Based Practice Through Organizational Policies and an Innovative Model. AJN Am. J. Nurs. 2011;111:57–60. doi: 10.1097/01.NAJ.0000405063.97774.0e. [DOI] [PubMed] [Google Scholar]
  • 7.Damschroder L.J., Aron D.C., Keith R.E., Kirsh S.R., Alexander J.A., Lowery J.C. Fostering implementation of health services research findings into practice: A consolidated framework for advancing implementation science. Implement. Sci. 2009;4:50. doi: 10.1186/1748-5908-4-50. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.Levin R.F., Fineout-Overholt E., Melnyk B.M., Barnes M., Vetter M.J. Fostering evidence-based practice to improve nurse and cost outcomes in a community health setting: A pilot test of the advancing research and clinical practice through close collaboration model. Nurs. Adm. Q. 2011;35:21–33. doi: 10.1097/NAQ.0b013e31820320ff. [DOI] [PubMed] [Google Scholar]
  • 9.Melnyk B.M., Gallagher-Ford L., Long L.E., Fineout-Overholt E. The establishment of evidence-based practice competencies for practicing registered nurses and advanced practice nurses in real-world clinical settings: Proficiencies to improve healthcare quality, reliability, patient outcomes, and costs. Worldviews Evid. Based Nurs. 2014;11:5–15. doi: 10.1111/wvn.12021. [DOI] [PubMed] [Google Scholar]
  • 10.Sackett D.L., Rosenberg W.M.C., Gray J.A.M., Haynes R.B., Richardson W.S. Evidence based medicine: What it is and what it isn’t. BMJ. 1996;312:71–72. doi: 10.1136/bmj.312.7023.71. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11.Saunders H., Gallagher-Ford L., Kvist T., Vehviläinen-Julkunen K. Practicing Healthcare Professionals’ Evidence-Based Practice Competencies: An Overview of Systematic Reviews. Worldviews Evid. Based Nurs. 2019;16:176–185. doi: 10.1111/wvn.12363. [DOI] [PubMed] [Google Scholar]
  • 12.Fernández-Domínguez J.C., De Pedro-Gómez J.E., Jiménez-López R., Romero-Franco N., Bays Moneo A.B., Oliva-Pascual-Vaca Á., Sesé-Abad A. Physiotherapists’ Evidence-Based Practice profiles by HS-EBP questionnaire in Spain: A cross-sectional normative study. PLoS ONE. 2022;17:e02694602022. doi: 10.1371/journal.pone.0269460. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.Scurlock-Evans L., Upton P., Upton D. Evidence-based practice in physiotherapy: A systematic review of barriers, enablers and interventions. Physiotherapy. 2014;100:208–219. doi: 10.1016/j.physio.2014.03.001. [DOI] [PubMed] [Google Scholar]
  • 14.Gudjonsdottir B., Arnadottir H.A., Gudmundsson H.S., Juliusdottir S., Arnadottir S.A. Attitudes Toward Adoption of Evidence-Based Practice Among Physical Therapists and Social Workers: A Lesson for Interprofessional Continuing Education. J. Contin. Educ. Health Prof. 2017;37:37–45. doi: 10.1097/CEH.0000000000000139. [DOI] [PubMed] [Google Scholar]
  • 15.Cleary-Holdforth J., O’Mathúna D., Fineout-Overholt E. Evidence-Based Practice Beliefs, Implementation, and Organizational Culture and Readiness for EBP Among Nurses, Midwives, Educators, and Students in the Republic of Ireland. Worldviews Evid. Based Nurs. 2021;18:379–388. doi: 10.1111/wvn.12543. [DOI] [PubMed] [Google Scholar]
  • 16.Melnyk B.M. The Role of Technology in Enhancing Evidence-Based Practice, Education, Heathcare Quality, and Patient Outcomes: A Call for Randomized Controlled Trials and Comparative Effectiveness Research. Worldviews Evid. Based Nurs. 2012;9:63–65. doi: 10.1111/j.1741-6787.2012.00245.x. [DOI] [PubMed] [Google Scholar]
  • 17.Kirk M.A., Kelley C., Yankey N., Birken S.A., Abadie B., Damschroder L. A systematic review of the use of the consolidated framework for implementation research. Implement. Sci. 2015;11:72. doi: 10.1186/s13012-016-0437-z. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18.Patelarou A.E., Patelarou E., Brokalaki H., Dafermos V., Thiel L., Melas C.D., Koukia E. Current evidence on the attitudes, knowledge and perceptions of nurses regarding evidence-based practice implementation in European community settings: A systematic review. J. Community Health Nurs. 2013;30:230–244. doi: 10.1080/07370016.2013.838501. [DOI] [PubMed] [Google Scholar]
  • 19.Dalheim A., Harthug S., Nilsen R.M., Nortvedt M.W. Factors influencing the development of evidence-based practice among nurses: A self-report survey. BMC Health Serv. Res. 2012;12:367. doi: 10.1186/1472-6963-12-367. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 20.Shorey S., Goh M.L., Ang S.Y., Ang L., Devi M.K., Ang E. The Progression and Future of Nursing in Singapore: A Descriptive Qualitative Study. J. Transcult. Nurs. 2019;30:512–520. doi: 10.1177/1043659618823909. [DOI] [PubMed] [Google Scholar]
  • 21.Lang N.M. Discipline-based approaches to evidence-based practice: A view from nursing. Jt. Comm. J. Qual. Improv. 1999;25:539–544. doi: 10.1016/S1070-3241(16)30468-0. [DOI] [PubMed] [Google Scholar]
  • 22.Dennis S. Secondary prevention of chronic health conditions in patients with multimorbidity: What can physiotherapists do? J. Comorb. 2016;6:50–52. doi: 10.15256/joc.2016.6.82. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 23.Johansson G., Eklund K., Gosman-Hedström G. Multidisciplinary team, working with elderly persons living in the community: A systematic literature review. Scand. J. Occup. 2010;17:101–116. doi: 10.3109/11038120902978096. [DOI] [PubMed] [Google Scholar]
  • 24.Brownstein J.N., Hirsch G.R., Rosenthal E.L., Rush C.H. Community health workers “101” for primary care providers and other stakeholders in health care systems. J. Ambul. Care Manag. 2011;34:210–220. doi: 10.1097/JAC.0b013e31821c645d. [DOI] [PubMed] [Google Scholar]
  • 25.Ubbink D.T., Vermeulen H., Knops A.M., Legemate D.A., Oude Rengerink K., Heineman M.J., Roos Y.B., Fijnvandraat C.J., Heymans H.S., Simons R., et al. Implementation of evidence-based practice: Outside the box, throughout the hospital. Neth. J. Med. 2011;69:87–94. [PubMed] [Google Scholar]
  • 26.Majid S., Foo S., Luyt B., Zhang X., Theng Y.-L., Chang Y.-K., Mokhtar I.A. Adopting evidence-based practice in clinical decision making: Nurses’ perceptions, knowledge, and barriers. J. Med. Libr. Assoc. JMLA. 2011;99:229–236. doi: 10.3163/1536-5050.99.3.010. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27.Saunders H., Vehviläinen-Julkunen K. The state of readiness for evidence-based practice among nurses: An integrative review. Int. J. Nurs. Stud. 2016;56:128–140. doi: 10.1016/j.ijnurstu.2015.10.018. [DOI] [PubMed] [Google Scholar]
  • 28.Stichler J.F., Fields W., Kim S.C., Brown C.E. Faculty knowledge, attitudes, and perceived barriers to teaching evidence-based nursing. J. Prof. Nurs. 2011;27:92–100. doi: 10.1016/j.profnurs.2010.09.012. [DOI] [PubMed] [Google Scholar]
  • 29.Albarqouni L., Hoffmann T., Straus S., Olsen N.R., Young T., Ilic D., Shaneyfelt T., Haynes R.B., Guyatt G., Glasziou P. Core Competencies in Evidence-Based Practice for Health Professionals: Consensus Statement Based on a Systematic Review and Delphi Survey. JAMA Netw. Open. 2018;1:e180281. doi: 10.1001/jamanetworkopen.2018.0281. [DOI] [PubMed] [Google Scholar]
  • 30.Zhao J., Bai W., Zhang Q., Su Y., Wang J., Du X., Zhou Y., Kong C., Qing Y., Gong S., et al. Evidence-based practice implementation in healthcare in China: A living scoping review. Lancet Reg. Health West. Pac. 2022;20:100355. doi: 10.1016/j.lanwpc.2021.100355. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 31.Malik G., McKenna L., Plummer V. Facilitators and barriers to evidence-based practice: Perceptions of nurse educators, clinical coaches and nurse specialists from a descriptive study. Contemp. Nurse. 2016;52:544–554. doi: 10.1080/10376178.2016.1188017. [DOI] [PubMed] [Google Scholar]
  • 32.Mohsen M.M., Safaan N.A., Okby O.M. Nurses’ Perceptions and Barriers for Adoption of Evidence Based Practice in Primary Care: Bridging the Gap. Am. J. Nurs. Res. 2016;4:25–33. [Google Scholar]
  • 33.World Health Organization . Global Strategic Directions for Strengthening Nursing and Midwifery 2016–2020. WHO; Geneva, Switzerland: 2016. [Google Scholar]
  • 34.Melender H.-L., Mattila L.-R., Häggman-Laitila A. A systematic review on educational interventions for learning and implementing evidence-based practice in nursing education: The state of evidence. Nord. J. Nurs. Res. 2016;36:3–12. doi: 10.1177/0107408315595161. [DOI] [Google Scholar]
  • 35.Melnyk B.M., Fineout-Overholt E. Evidence-Based Practice in Nursing & Healthcare: A Guide to Best Practice. Wolters Kluwer; Philadelphia, PA, USA: 2019. [Google Scholar]
  • 36.Moher D., Liberati A., Tetzlaff J., Altman D.G., PRISMA Group * Preferred reporting items for systematic reviews and meta-analyses: The PRISMA statement. Ann. Intern. Med. 2009;151:264–269. doi: 10.7326/0003-4819-151-4-200908180-00135. [DOI] [PubMed] [Google Scholar]
  • 37.Stroup D.F., Berlin J.A., Morton S.C., Olkin I., Williamson G.D., Rennie D., Moher D., Becker B.J., Sipe T.A., Thacker S.B. Meta-analysis of observational studies in epidemiology: A proposal for reporting. JAMA. 2000;283:2008–2012. doi: 10.1001/jama.283.15.2008. [DOI] [PubMed] [Google Scholar]
  • 38.Higgins J.P., Thomas J., Chandler J., Cumpston M., Li T., Page M.J., Welch V.A. Cochrane Handbook for Systematic Reviews of Interventions. John Wiley & Sons; Hoboken, NJ, USA: 2019. [Google Scholar]
  • 39.Verloo H., Melly P., Hilfiker R., Pereira F. Effectiveness of Educational Interventions to Increase Knowledge of Evidence-Based Practice Among Nurses and Physiotherapists in Primary Health Care: Protocol for a Systematic Review. JMIR Res. Protoc. 2020;9:e17621. doi: 10.2196/17621. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 40.Sedgwick P. What is a non-randomised controlled trial? BMJ. 2014;348:g41152014. doi: 10.1136/bmj.g4115. [DOI] [PubMed] [Google Scholar]
  • 41.Ferriter M., Huband N. Does the non-randomized controlled study have a place in the systematic review? A pilot study. Crim. Behav. Ment. Health. 2005;15:111–120. doi: 10.1002/cbm.43. [DOI] [PubMed] [Google Scholar]
  • 42.Manchikanti L., Singh V., Smith H.S., Hirsch J.A. Evidence-based medicine, systematic reviews, and guidelines in interventional pain management: Part 4: Observational studies. Pain Physician. 2009;12:73–108. doi: 10.36076/ppj.2009/12/73. [DOI] [PubMed] [Google Scholar]
  • 43.Glenton C., Lewin S., Downe S., Paulsen E., Munabi-Babigumira S., Agarwal S., Ames H., Cooper S., Daniels K., Houghton C., et al. Cochrane Effective Practice and Organisation of Care (EPOC) Qualitative Evidence Syntheses, Differences From Reviews of Intervention Effectiveness and Implications for Guidance. Int. J. Qual. Methods. 2022;21:16094069211061950. doi: 10.1177/16094069211061950. [DOI] [Google Scholar]
  • 44.Van Camp Y.P., Van Rompaey B., Elseviers M.M. Nurse-led interventions to enhance adherence to chronic medication: Systematic review and meta-analysis of randomised controlled trials. Eur. J. Clin. Pharmacol. 2013;69:761–770. doi: 10.1007/s00228-012-1419-y. [DOI] [PubMed] [Google Scholar]
  • 45.Osterberg L., Blaschke T. Adherence to Medication. N. Engl. J. Med. 2005;353:487–497. doi: 10.1056/NEJMra050100. [DOI] [PubMed] [Google Scholar]
  • 46.Page M.J., McKenzie J.E., Bossuyt P.M., Boutron I., Hoffmann T.C., Mulrow C.D., Shamseer L., Tetzlaff J.M., Akl E.A., Brennan S.E., et al. The PRISMA 2020 statement: An updated guideline for reporting systematic reviews. BMJ. 2021;372:n71. doi: 10.1136/bmj.n71. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 47.Peterson J., Welch V., Losos M., Tugwell P. The Newcastle-Ottawa scale (NOS) for assessing the quality of nonrandomised studies in meta-analyses. Ott. Ott. Hosp. Res. Inst. 2011;2:1–12. [Google Scholar]
  • 48.Singleton J.K. Evidence-Based Practice Beliefs and Implementation in Doctor of Nursing Practice Students. Worldviews Evid. Based. Nurs. 2017;14:412–418. doi: 10.1111/wvn.12228. [DOI] [PubMed] [Google Scholar]
  • 49.Jüni P., Loke Y., Pigott T., Ramsay C., Regidor D., Rothstein H., Sandhu L., Santaguida P., Schünemann H., Shea B. Risk of bias in non-randomized studies of interventions (ROBINS-I): Detailed guidance. Br. Med. J. 2016;2016:1–53. [Google Scholar]
  • 50.Gallagher-Ford L., Koshy Thomas B., Connor L., Sinnott L.T., Melnyk B.M. The Effects of an Intensive Evidence-Based Practice Educational and Skills Building Program on EBP Competency and Attributes. Worldviews Evid. Based Nurs. 2020;17:71–81. doi: 10.1111/wvn.12397. [DOI] [PubMed] [Google Scholar]
  • 51.Hart P., Eaton L., Buckner M., Morrow B.N., Barrett D.T., Fraser D.D., Hooks D., Sharrer R.L. Effectiveness of a Computer-Based Educational Program on Nurses’ Knowledge, Attitude, and Skill Level Related to Evidence-Based Practice. Worldviews Evid. Based Nurs. 2008;5:75–84. doi: 10.1111/j.1741-6787.2008.00123.x. [DOI] [PubMed] [Google Scholar]
  • 52.Hsieh P.-L., Chen S.-H. Effectiveness of an Evidence-Based Practice Educational Intervention among School Nurses. Int. J. Environ. Res. Public Health. 2020;17:4063. doi: 10.3390/ijerph17114063. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 53.Mena-Tudela D., González-Chordá V.M., Cervera-Gasch A., Maciá-Soler M.L., Orts-Cortés M.I. Effectiveness of an Evidence-Based Practice educational intervention with second-year nursing students. Rev. Lat.-Am. Enferm. 2018;26:1–9. doi: 10.1590/1518-8345.2502.3026. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 54.Moore L. Effectiveness of an Online Educational Module in Improving Evidence-Based Practice Skills of Practicing Registered Nurses. Worldviews Evid. Based Nurs. 2017;14:358–366. doi: 10.1111/wvn.12214. [DOI] [PubMed] [Google Scholar]
  • 55.Moore L.K. Improving Evidence-Based Practice Skills of Practicing Registered Nurses. Western Kentucky University; Bowling Green, KY, USA: 2018. [Google Scholar]
  • 56.Underhill M., Roper K., Siefert M.L., Boucher J., Berry D. Evidence-Based Practice Beliefs and Implementation Before and After an Initiative to Promote Evidence-Based Nursing in an Ambulatory Oncology Setting. Worldviews Evid. Based Nurs. 2015;12:70–78. doi: 10.1111/wvn.12080. [DOI] [PubMed] [Google Scholar]
  • 57.Wan L.P.A. Educational Intervention Effects on Nurses’ Perceived Ability to Implement Evidence-Based Practice. University of Phoenix; Phoenix, AZ, USA: 2017. [Google Scholar]
  • 58.Sterne J.A.C., Savović J., Page M.J., Elbers R.G., Blencowe N.S., Boutron I., Cates C.J., Cheng H.Y., Corbett M.S., Eldridge S.M., et al. RoB 2: A revised tool for assessing risk of bias in randomised trials. BMJ. 2019;366:l4898. doi: 10.1136/bmj.l4898. [DOI] [PubMed] [Google Scholar]
  • 59.Elliott J.H., Synnot A., Turner T., Simmonds M., Akl E.A., McDonald S., Salanti G., Meerpohl J., MacLehose H., Hilton J., et al. Living systematic review: 1. Introduction-the why, what, when, and how. J. Clin. Epidemiol. 2017;91:23–30. doi: 10.1016/j.jclinepi.2017.08.010. [DOI] [PubMed] [Google Scholar]
  • 60.Chao W.-Y., Huang L.-C., Hung H.-C., Hung S.-C., Chuang T.-F., Yeh L.-Y., Tseng H.-C. Effectiveness of Digital Flipped Learning Evidence-Based Practice on Nurses’ Knowledge, Attitude, and Practice: A Quasi-Experimental Trial. Healthcare. 2022;10:1257. doi: 10.3390/healthcare10071257. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 61.Waltz L.A., Munoz L., Miller R.A., Weber Johnson H. Measuring the Impact of an Educational Intervention on Evidence-Based Practice. J. Contin. Educ. Nurs. 2022;53:109–114. doi: 10.3928/00220124-20220210-05. [DOI] [PubMed] [Google Scholar]
  • 62.Cardoso D., Couto F., Cardoso A.F., Bobrowicz-Campos E., Santos L., Rodrigues R., Coutinho V., Pinto D., Ramis M.-A., Rodrigues M.A., et al. The Effectiveness of an Evidence-Based Practice (EBP) Educational Program on Undergraduate Nursing Students’ EBP Knowledge and Skills: A Cluster Randomized Control Trial. Int. J. Environ. Res. Public Health. 2021;18:293. doi: 10.3390/ijerph18010293. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 63.Koota E., Kääriäinen M., Melender H.-L. Educational interventions promoting evidence-based practice among emergency nurses: A systematic review. Int. Emerg. Nurs. 2018;41:51–58. doi: 10.1016/j.ienj.2018.06.004. [DOI] [PubMed] [Google Scholar]
  • 64.D’Souza P., George A., Nair S., Noronha J., Renjith V. Effectiveness of an Evidence-Based Practice Training Program for Nurse Educators: A Cluster-Randomized Controlled Trial. Worldviews Evid. Based Nurs. 2021;18:261–271. doi: 10.1111/wvn.12521. [DOI] [PubMed] [Google Scholar]
  • 65.Mickan S., Hilder J., Wenke R., Thomas R. The impact of a small-group educational intervention for allied health professionals to enhance evidence-based practice: Mixed methods evaluation. BMC Med. Educ. 2019;19:131. doi: 10.1186/s12909-019-1567-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 66.Koota E., Kääriäinen M., Kyngäs H., Lääperi M., Melender H.-L. Effectiveness of Evidence-Based Practice (EBP) Education on Emergency Nurses’ EBP Attitudes, Knowledge, Self-Efficacy, Skills, and Behavior: A Randomized Controlled Trial. Worldviews Evid. Based Nurs. 2021;18:23–32. doi: 10.1111/wvn.12485. [DOI] [PubMed] [Google Scholar]
  • 67.Melnyk B.M., Fineout-Overholt E. Key steps in implementing evidence-based practice: Asking compelling, searchable questions and searching for the best evidence. Pediatr. Nurs. 2002;28:262. [PubMed] [Google Scholar]
  • 68.Melnyk B.M. The evidence-based practice mentor: A promising strategy for implementing and sustaining EBP in healthcare systems. Worldviews Evid. Based Nurs. 2007;4:123–125. doi: 10.1111/j.1741-6787.2007.00094.x. [DOI] [PubMed] [Google Scholar]
  • 69.Melnyk B.M., Fineout-Overholt E., Fischbeck Feinstein N., Li H., Small L., Wilcox L., Kraus R. Nurses’ perceived knowledge, beliefs, skills, and needs regarding evidence-based practice: Implications for accelerating the paradigm shift. Worldviews Evid. Based Nurs. 2004;1:185–193. doi: 10.1111/j.1524-475X.2004.04024.x. [DOI] [PubMed] [Google Scholar]
  • 70.Tilson J.K., Kaplan S.L., Harris J.L., Hutchinson A., Ilic D., Niederman R., Potomkova J., Zwolsman S.E. Sicily statement on classification and development of evidence-based practice learning assessment tools. BMC Med. Educ. 2011;11:78. doi: 10.1186/1472-6920-11-78. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 71.Melnyk B.M., Fineout-Overholt E., Giggleman M., Cruz R. Correlates among cognitive beliefs, EBP implementation, organizational culture, cohesion and job satisfaction in evidence-based practice mentors from a community hospital system. Nurs. Outlook. 2010;58:301–308. doi: 10.1016/j.outlook.2010.06.002. [DOI] [PubMed] [Google Scholar]
  • 72.Andrews M., Wallis M. Mentorship in nursing: A literature review. J. Adv. Nurs. 1999;29:201–207. doi: 10.1046/j.1365-2648.1999.00884.x. [DOI] [PubMed] [Google Scholar]
  • 73.Marie Block L., Claffey C., Korow M.K., McCaffrey R. The Value of Mentorship Within Nursing Organizations. Nurs. Forum. 2005;40:134–140. doi: 10.1111/j.1744-6198.2005.00026.x. [DOI] [PubMed] [Google Scholar]
  • 74.Guerra-Martn M., Lima-Serrano M., Lima-Rodrguez J. Effectiveness of Tutoring to Improve Academic Performance in Nursing Students at the University of Seville. J. New Approaches Educ. Res. NAER J. 2017;6:93–102. doi: 10.7821/naer.2017.7.201. [DOI] [Google Scholar]
  • 75.Loke A.J.T.Y., Chow F.L.W. Learning partnership—The experience of peer tutoring among nursing students: A qualitative study. Int. J. Nurs. Stud. 2007;44:237–244. doi: 10.1016/j.ijnurstu.2005.11.028. [DOI] [PubMed] [Google Scholar]
  • 76.Tripathi A. Glossary of distance education and internet terminology. Retrieved April. 1998;21:2010. [Google Scholar]
  • 77.Reeves T. Evaluating what really matters in computer-based education. Comput. Educ. New Perspect. 1994:219–246. doi: 10.1007/BF02948597. [DOI] [Google Scholar]
  • 78.Ciliska D.K., Pinelli J., DiCenso A., Cullum N. Resources to enhance evidence-based nursing practice. AACN Adv. Crit. Care. 2001;12:520–528. doi: 10.1097/00044067-200111000-00008. [DOI] [PubMed] [Google Scholar]
  • 79.Lee C.Y., Wang W. F., Chang Y. J. The effects of evidence-based nursing training program on nurses’ knowledge, attitude, and behavior. New Taipei J. Nur. 2011;13:19–31. doi: 10.6540/ntjn.2011.1.003. [DOI] [Google Scholar]
  • 80.Good L.R., Nelson D.A. Effects of person-group and intragroup attitude similarity on perceived group attractiveness and cohesiveness: II. Psychol. Rep. 1973;33:551–560. doi: 10.2466/pr0.1973.33.2.551. [DOI] [Google Scholar]
  • 81.Fineout-Overholt E., Levin R.F., Melnyk B.M. Strategies for advancing evidence-based practice in clinical settings. JNY State Nurses Assoc. 2004;35:28–32. [PubMed] [Google Scholar]
  • 82.Damschroder L.J., Lowery J.C. Evaluation of a large-scale weight management program using the consolidated framework for implementation research (CFIR) Implement. Sci. 2013;8:51. doi: 10.1186/1748-5908-8-51. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 83.Fernandez M.E., Walker T.J., Weiner B.J., Calo W.A., Liang S., Risendal B., Friedman D.B., Tu S.P., Williams R.S., Jacobs S. Developing measures to assess constructs from the inner setting domain of the consolidated framework for implementation research. Implement. Sci. 2018;13:1–13. doi: 10.1186/s13012-018-0736-7. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 84.Penz K., D’Arcy C., Stewart N., Kosteniuk J., Morgan D., Smith B. Barriers to participation in continuing education activities among rural and remote nurses. J. Contin. Educ. Nurs. 2007;38:58–66. doi: 10.3928/00220124-20070301-03. [DOI] [PubMed] [Google Scholar]
  • 85.Weng Y.H., Chen C., Kuo K.N., Yang C.Y., Lo H.L., Chen K.H., Chiu Y.W. Implementation of evidence-based practice in relation to a clinical nursing ladder system: A national survey in Taiwan. Worldviews Evid. Based Nurs. 2015;12:22–30. doi: 10.1111/wvn.12076. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 86.Williams B., Perillo S., Brown T. What are the factors of organisational culture in health care settings that act as barriers to the implementation of evidence-based practice? A scoping review. Nurse Educ. Today. 2015;35:e34–e41. doi: 10.1016/j.nedt.2014.11.012. [DOI] [PubMed] [Google Scholar]
  • 87.Chan E.Y., Glass G.F., Phang K.N. Evaluation of a Hospital-Based Nursing Research and Evidence-Based Practice Mentorship Program on Improving Nurses’ Knowledge, Attitudes, and Evidence-Based Practice. J. Contin. Educ. Nurs. 2020;51:46–52. doi: 10.3928/00220124-20191217-09. [DOI] [PubMed] [Google Scholar]
  • 88.Sadeghi-Bazargani H., Tabrizi J.S., Azami-Aghdash S. Barriers to evidence-based medicine: A systematic review. J. Eval. Clin. Pr. 2014;20:793–802. doi: 10.1111/jep.12222. [DOI] [PubMed] [Google Scholar]
  • 89.Kitson A.L., Rycroft-Malone J., Harvey G., McCormack B., Seers K., Titchen A. Evaluating the successful implementation of evidence into practice using the PARiHS framework: Theoretical and practical challenges. Implement. Sci. 2008;3:1. doi: 10.1186/1748-5908-3-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 90.Titler M.G., Kleiber C., Steelman V.J., Rakel B.A., Budreau G., Everett L.Q., Buckwalter K.C., Tripp-Reimer T., Goode C.J. The Iowa model of evidence-based practice to promote quality care. Crit. Care Nurs. Clin. N. Am. 2001;13:497–509. doi: 10.1016/S0899-5885(18)30017-0. [DOI] [PubMed] [Google Scholar]
  • 91.Levin R., Vetter M., Fineout-Overholt E., Melnyk B., Barnes M. Advancing research and clinical practice through close collaboration (ARCC): A pilot test of an intervention to improve evidence-based care and patient outcomes in a community health setting; Proceedings of the 8th Annual National/International Evidence-Based Practice Conference Translating Research into Best Practice with Vulnerable Populations; Indianapolis, IN, USA. 14 November 2005. [Google Scholar]
  • 92.Pearson A., Wiechula R., Court A., Lockwood C. The JBI model of evidence-based healthcare. Int. J. Evid. Based Healthc. 2005;3:207–215. doi: 10.1111/j.1479-6988.2005.00026.x. [DOI] [PubMed] [Google Scholar]
  • 93.Newhouse R.P., Dearholt S.L., Poe S.S., Pugh L.C., White K.M. Johns Hopkins Nursing Evidence-Based Practice Model and Guidelines. Sigma Theta Tau International Honor Society of Nursing; Indianapolis, IN, USA: 2007. [Google Scholar]

Articles from Healthcare are provided here courtesy of Multidisciplinary Digital Publishing Institute (MDPI)

RESOURCES