Skip to main content
Implementation Science Communications logoLink to Implementation Science Communications
. 2021 Sep 8;2:99. doi: 10.1186/s43058-021-00201-1

Aligning implementation science with improvement practice: a call to action

Jennifer Leeman 1,, Catherine Rohweder 2, Matthew Lee 3, Alison Brenner 4, Andrea Dwyer 5, Linda K Ko 6,7, Meghan C O’Leary 8, Grace Ryan 9, Thuy Vu 6, Shoba Ramanadhan 10
PMCID: PMC8424169  PMID: 34496978

Abstract

Background

In several recent articles, authors have called for aligning the fields of implementation and improvement science. In this paper, we call for implementation science to also align with improvement practice. Multiple implementation scholars have highlighted the importance of designing implementation strategies to fit the existing culture, infrastructure, and practice of a healthcare system. Worldwide, healthcare systems are adopting improvement models as their primary approach to improving healthcare delivery and outcomes. The prevalence of improvement models raises the question of how implementation scientists might best align their efforts with healthcare systems’ existing improvement infrastructure and practice.

Main body

We describe three challenges and five benefits to aligning implementation science and improvement practice. Challenges include (1) use of different models, terminology, and methods, (2) a focus on generalizable versus local knowledge, and (3) limited evidence in support of the effectiveness of improvement tools and methods. We contend that implementation science needs to move beyond these challenges and work toward greater alignment with improvement practice. Aligning with improvement practice would benefit implementation science by (1) strengthening research/practice partnerships, (2) fostering local ownership of implementation, (3) generating practice-based evidence, (4) developing context-specific implementation strategies, and (5) building practice-level capacity to implement interventions and improve care. Each of these potential benefits is illustrated in a case study from the Centers for Disease Control and Prevention’s Cancer Prevention and Control Research Network.

Conclusion

To effectively integrate evidence-based interventions into routine practice, implementation scientists need to align their efforts with the improvement culture and practice that is driving change within healthcare systems worldwide. This paper provides concrete examples of how researchers have aligned implementation science with improvement practice across five implementation projects.

Supplementary Information

The online version contains supplementary material available at 10.1186/s43058-021-00201-1.

Keywords: Implementation science, Improvement practice, Quality improvement, Cancer prevention and control, Case studies


Contributions to the literature.

  • Tension continues to exist between the fields of implementation science and improvement practice. This paper presents both the challenges and benefits to aligning the two fields.

  • This paper argues that greater alignment between implementation science and improvement practice has the potential to strengthen the implementation of evidence-based interventions.

  • Five case studies provide concrete examples of how Cancer Prevention and Control Research Network members are achieving the benefits of aligning implementation science with improvement practice.

Background

In several recent articles, authors have called for aligning the fields of implementation and improvement science with the goal of strengthening both fields’ potential to improve care [13]. These authors contend that closer alignment would promote “cross-fertilization,” with each field learning from and building on the other’s expertise, methods, and frameworks [2]. While we recognize the value of closer alignment between implementation science and improvement science, we contend that implementation science may gain as much or more from greater alignment with improvement practice.

Implementation science and improvement science both aim to produce generalizable knowledge to guide the improvement of healthcare services and health outcomes but differ in their approach. Implementation science focuses on increasing the uptake of evidence-based interventions (EBIs) whereas improvement science focuses on iteratively measuring and improving care processes and systems (2). Improvement science has its foundation in the field of improvement practice, which also focuses on improving care processes and systems (1). Improvement practice is distinct from improvement science in that its goal is to produce the local (rather than generalizable) knowledge needed to improve care in a specific setting [1]. In this paper, we acknowledge other scholars’ observations on the value of aligning implementation science and improvement science: for example, the opportunity to share their respective expertise in assessing contextual factors and engaging practice-level stakeholders in the co-creation of solutions [2]. However, we move beyond this prior work by highlighting the need for greater alignment between implementation science and improvement practice, with improvement practice defined as practice-initiated efforts to design and redesign work processes and systems with the goal of improving local outcomes [4].

Across the USA and worldwide, healthcare systems are adopting Lean Six Sigma, quality improvement, and other improvement models as their primary approach to improving practice [510]. Table 1 provides a list of widely used improvement models [11].

Table 1.

Sample improvement modelsa

Quality improvement model
Continuous Quality Improvement (CQI)
Donabedian Model
European Foundation for Quality Management (EFQM) model
ISO 9000-9004 (International Organization for Standardization)
Joint Commission on Accreditation of Healthcare Organizations (JCAHO) Model
Kooperation für Transparenz und Qualität im Krankenhaus (KTQ)
Lean Management
Malcolm Baldrige Quality Award (MBQA)
Quality Improvement (QI)
Six Sigma
Strategic Collaborative Quality Management (SCQM)
Total Quality Management (TQM)

a Models were referenced at least 3 times according to [9, 11]

EBI evidence-based intervention, PDSA Plan-Do-Study-Act, HPV human papillomavirus, FIT fecal immunochemical test

The widespread adoption of improvement models is driven in part by requirements that health systems develop improvement programs to qualify for health insurance reimbursement and other sources of funding [1215]. Concurrently, a host of organizations have emerged to provide training and online resources to build improvement capacity [1620]. As a result, many healthcare systems are investing in improvement infrastructure, hiring personnel with expertise in improvement models, and training providers and staff to use tools such as root cause analysis, process flow diagramming, run charts, and Plan-Do-Study-Act (PDSA) cycles [2125].

A central tenet of implementation science is that implementation strategies will be most successful when they align with healthcare systems’ existing culture, infrastructure, and practices [26]. Consistent with this tenant, healthcare systems’ widespread adoption of improvement models raises the question of whether and how implementation scientists might best align their efforts with improvement practice. In this paper, we describe three challenges and five benefits of aligning implementation science with improvement practice. We then illustrate each of the potential benefits of alignment in five case studies from the Centers for Disease Control and Prevention (CDC)-funded Cancer Prevention and Control Research Network (CPCRN).

Main text

The challenges of aligning implementation science with improvement practice

As detailed below, three primary challenges have impeded the alignment of implementation science with improvement practice.

Challenge 1: Implementation science and improvement practice use different models, terminology, and approaches

Implementation science and improvement practice both promote specific approaches, referred to as “strategies” in implementation science and “interventions” or “tools” in improvement practice [1, 2]. While the approaches have similar objectives, the two fields use different terminology and apply different techniques. For example, both implementation science and improvement practice promote approaches for identifying contextual factors that contribute to a problem and/or inform a practice change. Both fields also promote strategies, interventions, and tools for use in developing and testing solutions to problems. The implementation science literature describes a multitude of strategies, with one widely cited taxonomy listing 73 distinct strategies [27]. In contrast, each improvement model offers its own distinct and relatively small list of interventions and tools, such as Fishbone diagrams, the Five Why’s, and process flow diagrams among others [28]. Although many of these interventions and tools are similar across models, they may vary in terminology, format, methods, and frameworks. These variations not only create barriers to communication among implementation scientists and improvement practice professionals, but also create confusion among the frontline providers and staff who are central to the hands-on work of implementing EBIs.

Challenge 2: Improvement practice produces locally specific, rather than generalizable, knowledge

As noted above, implementation science and improvement practice have different objectives. Implementation science aims to provide generalizable knowledge to improve healthcare services and outcomes. In contrast, improvement practice seeks to generate local knowledge to improve healthcare services and outcomes in a specific setting [1]. The emphasis given to the difference between science and practice contributes to the tendency to view improvement practice as outside of implementation science’s scope. For example, Mitchell and Chambers (2017) depict improvement science, implementation science, and other fields as forming an inter-related grouping of disciplines, and place quality improvement (a widely used improvement model) in a small oval completely outside that grouping [29].

Challenge 3: The evidence base for the effectiveness of improvement practice is limited

Systematic reviews of the literature have identified wide variation in how healthcare systems are using improvement models and tools [3032]. For example, Taylor and colleagues [30] reviewed studies that used Plan-Do-Study-Act (PDSA cycles), a widely used improvement tool that involves multiple, small, rapid tests of a local improvement. They found that 81% (59/73) of studies failed to report the conduct of iterative cycles of change, and only two of the studies reviewed demonstrated healthcare system compliance with all key principles of the PDSA cycle method [30]. Reed and Card (2016) determined that teams rarely completed the full PDSA cycle, putting most of their effort into “Doing” and far less into “Studying” and “Acting” [31]. In a systematic review of the literature, Hill et al. (2020) found mixed evidence for the effect of quality improvement on health services and little evidence for effects on health outcomes [32]. The limited evidence in support of effectiveness may lead implementation scientists to dismiss improvement practice as an approach that does not work. However, this is not the conclusion of Hill et al. who point to the poor quality of evaluation methods, complexity of the changes being evaluated, and the need for more well-designed studies to demonstrate quality improvement’s effectiveness [32].

The benefits of aligning implementation science with improvement practice

In this paper, we advocate for the benefits of implementation science and improvement practice aligning with and building on each other’s strengths. We do not argue for the value of generalizing the findings from improvement projects or for the effectiveness of improvement practice at improving health services and outcomes. Rather, we highlight five potential benefits of aligning with improvement practice as a means of strengthening research/practice partnerships, fostering local ownership of EBI implementation, generating in-depth knowledge of the practice context, and developing context-specific implementation strategies.

Benefit 1: Strengthen research/practice partnerships by supporting healthcare systems’ improvement projects

Implementation science typically requires that researchers engage with healthcare providers, staff, and multiple levels of system leadership in the healthcare systems where they plan to implement EBIs. Engaging these stakeholders is challenging due to competing demands for their time and attention. Furthermore, these stakeholders may resist the input of researchers outside their organizations, as reflected in a recent article where authors refer to implementation science as tending toward “scientific imperialism” and an “overall approach that disrespects and undervalues primary care as a co-producer of knowledge” [33]. This observation not only applies to primary care but also to other healthcare systems and providers. As noted above, many healthcare systems are adopting improvement models as their primary method for improving care processes. Recognizing and respecting this reality may be key to implementation scientists gaining entrée into healthcare systems and engaging stakeholders in the adoption, adaptation, and implementation of new EBIs [34]. In addition to supporting implementation research, engaged research/practice partnerships can strengthen healthcare systems’ improvement projects by contributing expertise in potential solutions to identified problems and methods for evaluating impact.

Benefit 2: Foster local ownership of EBI implementation and sustainment

In improvement teams, providers and staff work together to plan and test improvements in their unique healthcare setting, thereby promoting local ownership of the change in practice [35]. Partnerships between improvement teams and implementation scientists can yield opportunities to co-create EBIs as well as implementation strategies. Decades of research have shown that individuals are more likely to adopt and implement interventions when they view themselves as the owners of the interventions and are able to shape them to fit local needs and preferences [36]. Local ownership also is essential to sustaining an intervention over time. Once engaged, an improvement team has the potential to monitor implementation and respond to changes such as new regulations, new funding streams, or changes in staffing or in the population served [37].

Benefit 3: Employ improvement tools to generate practice-based evidence

Eliciting local stakeholders’ knowledge of their local context is key to ensuring that an EBI and its implementation align with a practice setting’s needs, assets, and priorities. Recent years have seen a growing emphasis on the importance of this type of “practice-based” evidence, which Leeman and Sandelowski defined as “evidence concerning the contexts, experiences, and practices of healthcare providers working in real-world practice settings” [38]. Improvement models (e.g., Lean Six Sigma, the Improvement Model) typically consist of a systematic, multistep approach that involves the use of specific tools to describe current processes (e.g., process maps) and identify the root causes of problems with or gaps in those processes (e.g., fishbone diagrams). These tools are designed to capture team members’ local knowledge of the factors that influence care delivery and outcomes within a specific setting and population. In this way, practical considerations drive the selection, testing, and refinement of candidate solutions for a given evidence-practice gap. By partnering with improvement teams on the use of process maps and other improvement tools, implementation scientists can elicit the practice-based evidence needed to select, adapt, and refine EBIs and implementation strategies [39]. This careful attention to local knowledge may be particularly important to efforts to promote health equity, as it is key to identifying factors that contribute to health disparities [40].

Benefit 4: Conduct Plan-Do-Study-Act (PDSA) cycles to develop and refine implementation strategies

Another primary component of many improvement models is the PDSA cycle. In conducting PDSA cycles, improvement teams design changes that they then rapidly test in multiple, small experiments with the goal of developing “fit-for-purpose solutions” to improve care [30]. In contrast, many of the current approaches to developing implementation strategies require substantial, upfront investments of time and resources (e.g., Implementation Mapping) [41, 42]. Furthermore, they often yield a multicomponent implementation strategy that is then “frozen” and distributed for replication across multiple settings [37]. While settings may be encouraged to adapt multicomponent implementation strategies, little guidance is provided on how to identify and address contextual factors that may influence implementation in new settings [43]. PDSA cycles offer an approach that implementation scientists might apply to engage local improvement teams in rapid, iterative testing of implementation strategies. Multiple examples exist of implementation scientists using PDSA cycles to develop, test, and refine implementation strategies, such as in the case of “implementation laboratories” [44] and multi-organization “learning evaluation” projects [45].

Benefit 5: Build healthcare systems’ overall capacity to implement EBIs and improve care

Strengthening healthcare systems’ use of improvement models and tools is important for both the improvement teams and implementation scientists to achieve their maximum potential. The literature includes examples of implementation scientists partnering with healthcare systems to bolster application of their improvement models and tools, usually in support of a specific EBI [46]. Efforts to strengthen a system’s use of improvement models and tools include the provision of training, tools, and other support to build improvement capacity as well as evaluations of improvement processes and outcomes. Of note, efforts to build improvement capacity may incorporate implementation science frameworks, strategies, and outcomes. For example, evaluations of improvement projects may draw on implementation frameworks and outcomes. In addition, programs intended to strengthen the use of improvement models and tools may include menus of implementation strategies to broaden the solutions available to overcome barriers to improvement.

Cancer Prevention and Control Research Network (CPCRN) case studies of implementation science/improvement practice alignment

As summarized in Fig. 1, we describe five case studies that illustrate how CPCRN researchers are achieving the benefits of aligning implementation science and improvement practice in their cancer prevention and control projects. Each of the case studies was selected to highlight one of the five benefits; however, in reality each case exemplifies multiple benefits.

Fig. 1.

Fig. 1

Five case studies illustrating the five benefits of aligning implementation science and improvement practice within cancer prevention and control

The CPCRN is a national network funded by the Centers for Disease Control and Prevention (CDC) and National Cancer Institute (NCI). Since 2002, CPCRN centers have conducted practice-engaged research to accelerate the adoption and implementation of cancer prevention and control interventions [47, 48]. In this current funding cycle (2019–2024), the CPCRN created a cross-center workgroup to study the alignment of implementation science and improvement practice. Workgroup members from the eight CPCRN member sites completed an initial survey of implementation research projects that incorporated improvement models and tools. A total of 17 projects were identified and described. Below, we highlight five case studies, each of which was selected to illustrate one of the benefits of aligning implementation science and improvement practice.

Case Study 1: Utilizing a research/practice partnership to improve colorectal cancer screening rates (Benefit: Strengthen research/practice partnerships)

Researchers at the University of North Carolina (UNC) Chapel Hill are partnering with a large, academic primary care practice that serves 13,000 adult patients and provides over 39,000 visits per year. The practice has extensive experience with quality improvement (QI) and participates in a system-wide improvement collaborative.

Recently, UNC researchers partnered on a clinic-initiated project to address a sudden drop in colorectal cancer (CRC) screening rates. In October 2018, the practice’s QI committee noted a steep decrease in CRC screening rates while reviewing standard clinical quality measures. A subgroup of the QI committee was formed to study the problem and implement potential solutions. The team conducted a root-cause analysis and found that prior interventions, in which CRC screening home test kits were distributed during annual clinic visits, increased CRC screening rates. However, the interventions were not sustained over time, resulting in a drop in screening rates. To identify a more sustainable solution, the team engaged UNC researchers with expertise in mailed CRC test kit outreach, a CRC screening intervention that does not require a clinic visit and has limited impact on clinic workflows. Researchers partnered with the team on multiple PDSA cycles to develop a protocol for mailing CRC test kits, targeting patients who had previously completed a screening kit but were now overdue (Fig. 2). From November 2018 to April 2019, eligible patients received (a) a mailed packet containing an introductory letter and a screening kit, and (b) a reminder message through the patient portal. A total of 160 screening kits were mailed across four waves. In the first three waves, adjustments were made between mailings to improve the materials (e.g., wave 2 included an edited introductory letter, wave 3 included a Spanish language letter). As illustrated in Fig. 2, the project achieved its goal of increasing CRC screening rates to above 77%. This case study illustrates the importance of researcher support for a practice-initiated project in developing strong research/practice partnerships. Through their decade-long partnership with this clinic, CPCRN researchers have engaged clinic stakeholders in multiple researcher-initiated projects [49, 50].

Fig. 2.

Fig. 2

Run chart of CRC test kit completion rates

Case Study 2: Tailoring lung cancer screening to fit clinic systems within the Colorado Cancer Screening Program for Patient Navigation (Benefit: Foster local ownership of EBI implementation)

Researchers and public health practitioners at the University of Colorado Cancer Center and the Colorado School of Public Health have developed and implemented the Colorado Cancer Screening Program (CCSP). With funding from state-level tobacco tax revenue, the CCSP has built extensive statewide partnerships, including partnerships with nearly 150 clinic systems and other organizations. The cornerstone of CCSP over the last decade has been the use of patient navigation to deliver multi-level EBIs to increase CRC screening. Patient navigation is an evidence-based intervention that employs patient navigators to guide patients through the health care system and reduce barriers to screening and treatment [51]. In 2018, CCSP partnered with National Jewish Health to develop training, materials, and tools to expand patient navigation to lung cancer screening. In 2019, CCSP invited members of their statewide network of clinic systems to pilot the materials and tools. Guided by the PRISM framework [52], the CCSP team conducted an assessment to determine clinic readiness and selected two community health centers with the organizational capacity (i.e., commitment to cancer screening and staffing resources and time) and environmental supports (i.e., access to hospital or radiology practices to provide low-dose computed tomography [CT] screening services) needed to pilot test the lung cancer screening materials and tools. One of the tools was a process flow template outlining the processes that clinics and patient navigators would use to identify patients eligible for screening, engage providers and patients in shared decision making, refer patients to CT lung cancer centers/hospitals, and track processes and outcomes [53, 54].

CCSP provided technical assistance to aid clinics in adapting the process flow template and creating a process flow diagram that was tailored to each clinic’s existing staffing, workflows, and referral arrangement with CT lung cancer centers/hospitals. For example, the more urban of the two pilot clinics shares an electronic health record (EHR) with its CT provider. This shared system allows their primary care team to have a streamlined process flow diagram for referring patients and tracking processes and outcomes (Fig. 3). In contrast, the rural clinic does not have a formal connection with its CT provider nor do they share the same EHR. As a result, a more complex process flow diagram was needed to outline the processes needed to refer patients and track outcomes (see Additional File 1: Lung Cancer Screening Workflow). The approach to customizing workflows, while maintaining fidelity to the guidelines, allows the decision makers within each system to embrace lung cancer screening and implement it within existing capacity and resources while also planning for longer-term sustainability of their efforts.

Fig. 3.

Fig. 3

Process flow diagram for lung cancer screening

Case Study 3: Diagramming process flow to scale up colorectal cancer screening (Benefit: Generate practice-based evidence)

UNC researchers are incorporating process flow diagramming throughout the design, implementation, and evaluation of their National Institutes of Health (NIH)-funded Scaling Colorectal Cancer Screening through Outreach, Referral, and Engagement (SCORE) study. In collaboration with two community health centers, SCORE is testing the effectiveness of an intervention that combines mailed CRC screening home test kits, patient reminders, and navigation to follow-up colonoscopy for patients with a positive test result. Process flow diagrams are being used to highlight key differences in the process steps and to delineate roles and resources needed based on feedback from clinic staff about their local context and standards of care. The research team iteratively developed the process maps through multiple stakeholder-engaged sessions. The maps reflect input from program directors and staff at the community health centers where SCORE is being implemented. Each diagram documents all steps in the process and identifies which individuals perform each step. The team created a separate diagram for each of the core components of the SCORE intervention, which include developing a registry of eligible patients, conducting a mailed screening program, and navigating patients with a positive screening test to follow-up colonoscopy.

Early iterations of the process flow diagrams were used to plan implementation of the intervention. Visualization of the planned steps helped identify outstanding questions about both the intervention and implementation, as well as prioritize decision-making. These decisions included (a) determining the total number of patient contacts (e.g., mail and phone reminders); (b) optimizing the order of intervention activities and the time intervals between activities; and (c) identifying the appropriate individual or organization to perform specific steps. The process flow diagrams were then used to develop a comprehensive set of implementation outcomes to be measured. For example, fidelity and cost measures were mapped onto each process step, noting which tool(s) would be used for measurement, the frequency of measurement, and where these data would be tracked. The ongoing assessment of these measures during implementation will provide insight into the time and resources required for implementation, potential challenges to implementing the intervention as intended, and opportunities to create efficiencies. The evidence gathered from the SCORE process flow diagrams will be used to inform the adaptation and sustainment of the intervention in community health centers across the state.

Case Study 4: Using Plan-Do-Study-Act cycles to implement human papillomavirus vaccination programs (Benefit: Develop implementation strategies)

Researchers at the NYU Grossman School of Medicine are partnering with six community health centers (three pediatric clinics and three school-based clinics) in Southwest Brooklyn to evaluate and support the implementation of their HPV vaccination programs for low-income adolescent populations. As a first step, semi-structured key informant interviews are being conducted with multi-level stakeholders, including medical directors and other organizational leaders, nurse practitioners, frontline staff, and administrators across all six clinics. Guided by the RE-AIM framework [55], these interviews serve the dual purposes of establishing and maintaining relationships with organizational leaders and staff across the community health centers and soliciting multi-level perspectives on organizational processes and outcomes regarding the adoption, implementation, effectiveness, and sustainability of their HPV vaccination efforts. Analyses of the interview data will include understanding and documenting: (a) improvements efforts and infrastructure that are currently in place at each clinic; (b) implementation strategies that are being used to support reach, retention, and equity; and (c) ideas for how to improve the programs moving forward both in general and in response to the COVID-19 pandemic.

Informed by findings from these key informant interviews, the NYU team will convene a virtual learning collaborative and provide technical assistance to support each community health center as they conduct iterative PDSA cycles to refine existing implementation strategies and develop new ones as needed to address barriers to implementation. In addition to PDSA cycles, clinics will learn to use additional improvement tools that guide root cause and strength/weakness/opportunity/threat (SWOT) analyses (e.g., fishbone diagrams) for use in identifying and addressing potential threats as community health centers seek to build capacity for and sustain high-quality HPV vaccination programs in the future.

Case Study 5: Building capacity to monitor the implementation of mailed CRC screening home test kits (Benefit: Build overall capacity)

Researchers at the University of Washington (UW) provide implementation and evaluation technical assistance to community health centers across Washington State that participate in the CDC’s Colorectal Cancer Control Program. In this role, UW researchers partnered with three community health centers to evaluate a mailed CRC screening home test kit intervention to increase CRC screening. The three community health centers operate 45 medical clinics that mail out more than 2000 CRC test kits annually. UW researchers met individually with community health centers to learn about their CRC screening workflows and identify data capture points for monitoring and improvement. During these visits, community health center partners reported that their current quality monitoring systems lacked the capacity and flexibility to comprehensively monitor and evaluate implementation of their mailed screening kit programs.

Researchers developed a simple Microsoft Excel tracking tool for partners to use; tool contents were tailored based on one-on-one consultations with each community health center to align with their needs and resources. The tool included patient-level implementation variables to track such as demographic data, key dates (e.g., initial mailing date, reminder date(s), date screening kit was returned), test result, referral to colonoscopy (if needed), and colonoscopy outcome. Researchers pre-programmed reports in the tool so that community health centers could generate reports to monitor implementation in real time. Feedback from community health center staff on the tool and its adaptations has been predominantly positive. One community health center was able to work with their EHR vendor to add some of the suggested implementation measures directly into their EHR. A second community health center made enhancements to the tool when it became clear that their system had significant gaps in closing the colonoscopy referral loop and recording follow-up data. The third community health center reported that the automated reports enhanced their ability to provide real-time reporting to leadership, whereas previous reports were less timely due to data lags caused by processing internal data requests. In turn, community health center partners have supported UW’s research efforts over the past 5 years by providing letters of support for research applications and participating in emergent research opportunities.

Conclusion

The number of EBIs available to prevent and control cancer continues to increase [56, 57], and healthcare systems need support to fully implement and sustain these life-saving interventions and programs over time. In this paper, we contend that the alignment of implementation science with improvement practice has the potential to improve EBI implementation and sustainment. We describe five benefits of this alignment and present five case studies to illustrate how CPCRN researchers are leveraging improvement practice to implement CRC screening, lung cancer screening, and HPV vaccination EBIs in diverse healthcare settings. While we aligned each case study with a single benefit for clarity, the projects actually achieved multiple, overlapping benefits. Most notably, all cases resulted in stronger research/practice partnerships and increased capacity to implement EBIs and improve care.

Investments are needed to build the capacity within both implementation science and improvement practice. Ideally, implementation scientists would include orientation to improvement practice in the multiple trainings being offered to build capacity in the field of implementation science [58] and would encourage attendees to participate in one or more of the many trainings on improvement practice [16]. To build practice-level capacity, implementation scientists might “embed” themselves within healthcare systems’ improvement efforts [59, 60]. As illustrated in this paper’s case studies, implementation scientists, often in partnership with other intermediary organizations (e.g., American Cancer Society, National Jewish Health), provide learning collaboratives, training, and technical assistance to support clinics’ use of EBIs and quality improvement models and tools. The case studies illustrate how researchers began by assessing partners’ capacity (or readiness) to use improvement models and tools and then tailored their training and other strategies accordingly. By doing so, they built practice partners’ capacity to use improvement tools and models to implement and sustain EBIs over time. In our examples, the improvement tools that CPCRN researchers supported included run charts, process flow diagrams, root cause analysis, quality monitoring systems, and PDSA cycles. Most importantly, the case studies illustrate how positive outcomes can be achieved when both implementation scientists and improvement practitioners center their efforts on a common improvement goal, and draw on the full breadth of their respective expertise to achieve that goal.

Supplementary Information

Additional file1:. (102.2KB, docx)

Acknowledgements

The authors would like to thank Christina McMillan and Dr. Shana Ratner at UNC for contributing their CRC screening case study; Mary Wangen for assisting with graphics; and Dr. Stephanie Wheeler and Dr. David Chambers for commenting on drafts of the manuscript.

Abbreviations

ACS

American Cancer Society

CCSP

Colorado Cancer Screening Program

CDC

Centers for Disease Control and Prevention

CPCRN

Cancer Prevention and Control Research Network

CQI

Continuous quality improvement

CRC

Colorectal cancer

CT

Computed tomography

EBI

Evidence-based intervention

EFQM

European Foundation for Quality Management

EHR

Electronic health record

HPV

Human papillomavirus

ISO

International Organization for Standardization

JCAHO

Joint Commission on Accreditation of Healthcare Organizations

KTQ

Kooperation für Transparenz und Qualität im Krankenhaus

MBQA

Malcolm Baldrige Quality Award

NCI

National Cancer Institute

NIH

National Institutes of Health (PDSA) Plan-Do-Study-Act

QI

Quality improvement

SCORE

Scaling Colorectal Cancer Screening through Outreach, Referral, and Engagement

SCQM

Strategic Collaborative Quality Management

SWOT

Strength/Weakness/Opportunity/Threat

TQM

Total Quality Management

UI

University of Iowa

UNC

The University of North Carolina

UW

University of Washington

Authors’ contributions

JL conceptualized and drafted the manuscript; CR substantively contributed to and revised the manuscript; ML, AB, AD, MO, GR, and TV contributed case studies and revised the manuscript; LK and SR revised the manuscript. All authors read and approved the final manuscript.

Author’s information

JL (Lead and Corresponding Author) is Professor at the UNC School of Nursing at The University of North Carolina at Chapel Hill, and Principal Investigator of the Comprehensive Cancer Control Collaborative of North Carolina (4CNC), a Collaborating Center of the Cancer Prevention and Control Research Network (CPCRN).

Funding

This manuscript is supported by the Centers for Disease Control and Prevention of the U.S. Department of Health and Human Services (HHS) as part of a financial assistance award with 100 percent funded by CDC/HHS. The contents are those of the author(s) and do not necessarily represent the official views of, nor an endorsement, by CDC/HHS, or the U.S. Government. Individual authors were supported by the following Cooperative Agreement Numbers: Jennifer Leeman, Catherine Rohweder, Alison Brenner - U48 DP006400; Matthew Lee - U48 DP006396; Andrea Dwyer - U48 DP006399; Linda Ko, Thuy Vu - U48 DP006398; Grace Ryan - U48 DP006389. Meghan O'Leary was supported by the Cancer Care Quality Training Program, University of North Carolina at Chapel Hill, (Grant No. T32-CA-116339, PIs: Ethan Basch, Stephanie Wheeler).

Availability of data and materials

Not applicable to this article as no datasets were generated or analyzed during the current study.

Declarations

Ethics approval and consent to participate

Not applicable, this manuscript does not include any human subjects research.

Consent for publication

Not applicable, this manuscript does not contain data from any individual person.

Competing interests

The authors declare that they have no competing interests.

Footnotes

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

References

  • 1.Nilsen P, Bender M, Thor J, Leeman J, Andersson G, Sevdalis N. Improvement science. In: Nilsen P, Birken S, editors. Handbook on Implementation Science. UK: Edward Elgar Publishing; 2020. pp. 389–408. [Google Scholar]
  • 2.Koczwara B, Stover AM, Davies L, Davis MM, Fleisher L, Ramanadhan S, Schroeck FR, Zullig LL, Chambers DA, Proctor E. Harnessing the synergy between improvement science and implementation science in cancer: a call to action. J Oncol Pract. 2018;14(6):335–340. doi: 10.1200/jop.17.00083. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3.Check DK, Zullig LL, Davis MM, Davies L, Chambers D, Fleisher L, Kaplan SJ, Proctor E, Ramanadhan S, Schroeck FR, Stover AM, Koczwara B. Improvement science and implementation science in cancer care: identifying areas of synergy and opportunities for further integration. J Gen Intern Med. 2021;36(1):186–195. doi: 10.1007/s11606-020-06138-w. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4.Silvester K, Harriman P, Downes T. Introducing quality improvement as an intrinsic part of healthcare. Future Hosp J. 2016;3(3):188–190. doi: 10.7861/futurehosp.3-3-188. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5.Mousavi Isfahani H, Tourani S, Seyedin H. Features and results of conducted studies using a Lean management approach in emergency department in hospital: a systematic review. Bull Emerg Trauma. 2019;7(1):9–20. doi: 10.29252/beat-070102. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6.Amaratunga T, Dobranowski J. Systematic review of the application of Lean and Six Sigma quality improvement methodologies in radiology. J Am Coll Radiol. 2016;13(9):1088–95.e7. doi: 10.1016/j.jacr.2016.02.033. [DOI] [PubMed] [Google Scholar]
  • 7.Mason SE, Nicolay CR, Darzi A. The use of Lean and Six Sigma methodologies in surgery: a systematic review. Surgeon. 2015;13(2):91–100. doi: 10.1016/j.surge.2014.08.002. [DOI] [PubMed] [Google Scholar]
  • 8.Nicolay CR, Purkayastha S, Greenhalgh A, Benn J, Chaturvedi S, Phillips N, Darzi A. Systematic review of the application of quality improvement methodologies from the manufacturing industry to surgical healthcare. Br J Surg. 2012;99(3):324–335. doi: 10.1002/bjs.7803. [DOI] [PubMed] [Google Scholar]
  • 9.Glasgow JM, Scott-Caziewell JR, Kaboli PJ. Guiding inpatient quality improvement: a systematic review of Lean and Six Sigma. Jt Comm J Qual Patient Saf. 2010;36(12):533–540. doi: 10.1016/s1553-7250(10)36081-8. [DOI] [PubMed] [Google Scholar]
  • 10.The Breakthrough Series: IHI’s Collaborative Model for Achieving Breakthrough Improvement. IHI Innovation Series white paper. Boston: Institute for Healthcare Improvement; 2003. (Available on www.IHI.org).
  • 11.Maritz R, Scheel-Sailer A, Schmitt K, Prodinger B. Overview of quality management models for inpatient healthcare settings. A scoping review. Int J Qual Health Care. 2018;31(6):404–410. doi: 10.1093/intqhc/mzy180. [DOI] [PubMed] [Google Scholar]
  • 12.Swensen SJ, Dilling JA, McCarty PM, Bolton JW, Harper CM., Jr The business case for health-care quality improvement. J Patient Saf. 2013;9(1):44–52. doi: 10.1097/PTS.0b013e3182753e33. [DOI] [PubMed] [Google Scholar]
  • 13.Goldfield N. Potentially preventable events: an actionable set of measures for linking quality improvement and cost savings. Qual Manag Health Care. 2012;21(4):213–219. doi: 10.1097/QMH.0b013e31826d1d3a. [DOI] [PubMed] [Google Scholar]
  • 14.Schumacher H. Funding cancer quality improvement: payer's perspective. J Oncol Pract. 2015;11(3):180–181. doi: 10.1200/jop.2015.003913. [DOI] [PubMed] [Google Scholar]
  • 15.Wen L, Divers C, Lingohr-Smith M, Lin J, Ramsey S. Improving quality of care in oncology through healthcare payment reform. Am J Manag Care. 2018;24(3):e93–ee8. [PubMed] [Google Scholar]
  • 16.Institute for Healthcare Improvement: Eduation. http://www.ihi.org/education/. Accessed 30 Jan 2021.
  • 17.Joint Commission Center for Transforming Healthcare. https://www.centerfortransforminghealthcare.org/. Accessed 30 Jan 2021.
  • 18.The Quality Coaching Company. https://the-qcc.com/. Accessed 30 Jan 2021.
  • 19.Public Health Foundation. http://www.phf.org. Accessed 30 Jan 2021.
  • 20.Quality Improvement Organizations (CMS). https://www.cms.gov/Medicare/Quality-Initiatives-Patient-Assessment-Instruments/QualityImprovementOrgs. Accessed 30 Jan 2021.
  • 21.Alexander JA, Weiner BJ, Shortell SM, Baker LC, Becker MP. The role of organizational infrastructure in implementation of hospitals' quality improvement. Hosp Top. 2006;84(1):11–20. doi: 10.3200/HTPS.84.1.11-21. [DOI] [PubMed] [Google Scholar]
  • 22.Daugherty JD, Blake SC, Kohler SS, Culler SD, Hawley JN, Rask KJ. Quality improvement training: experiences of frontline staff. Int J Health Care Qual Assur. 2013;26(7):627–641. doi: 10.1108/IJHCQA-10-2011-0056. [DOI] [PubMed] [Google Scholar]
  • 23.Sarff L, O'Brien R. Evidence-based quality improvement training programs: building staff capability and organizational capacity. J Nurs Care Qual. 2020;35(2):95–101. doi: 10.1097/NCQ.0000000000000416. [DOI] [PubMed] [Google Scholar]
  • 24.Silow-Carroll S AT, Meyer J. Hospital quality improvement: strategies and lessons from U.S. hospitals. Commonwealth Fund pub. no. 1009: Health Management Associates, 2007.
  • 25.Newton WP, Lefebvre A, Donahue KE, Bacon T, Dobson A. Infrastructure for large-scale quality-improvement projects: early lessons from North Carolina improving performance in practice. J Contin Educ Health. 2010;30(2):106–113. doi: 10.1002/chp.20066. [DOI] [PubMed] [Google Scholar]
  • 26.Aarons GA, Hurlburt M, Horwitz SM. Advancing a conceptual model of evidence-based practice implementation in public service sectors. Adm Policy Ment Health. 2011;38(1):4–23. doi: 10.1007/s10488-010-0327-7. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27.Powell BJ, Waltz TJ, Chinman MJ, Damschroder LJ, Smith JL, Matthieu MM, Proctor EK, Kirchner JAE. A refined compilation of implementation strategies: results from the Expert Recommendations for Implementing Change (ERIC) project. Implement Sci. 2015;10(1):21. doi: 10.1186/s13012-015-0209-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 28.Institute for Healthcare Improvement. How to Improve. 2017. http://www.ihi.org/resources/Pages/HowtoImprove/default.aspx. Accessed 30 Jan 2021.
  • 29.Mitchell SA, Chambers DA. Leveraging implementation science to improve cancer care delivery and patient outcomes. J Oncol Pract. 2017;13(8):523–529. doi: 10.1200/jop.2017.024729. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 30.Taylor MJ, McNicholas C, Nicolay C, Darzi A, Bell D, Reed JE. Systematic review of the application of the plan-do-study-act method to improve quality in healthcare. BMJ Qual Saf. 2014;23(4):290–298. doi: 10.1136/bmjqs-2013-001862. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 31.Reed JE, Card AJ. The problem with Plan-Do-Study-Act cycles. BMJ Qual Saf. 2016;25(3):147–152. doi: 10.1136/bmjqs-2015-005076. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 32.Hill JE, Stephani A-M, Sapple P, Clegg AJ. The effectiveness of continuous quality improvement for developing professional practice and improving health care outcomes: a systematic review. Implement Sci. 2020;15(1):23. doi: 10.1186/s13012-020-0975-2. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 33.Miller WL, Rubinstein EB, Howard J, Crabtree BF. Shifting implementation science theory to empower primary care practices. Ann Fam Med. 2019;17(3):250–256. doi: 10.1370/afm.2353. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 34.Ramanadhan S, Davis MM, Armstrong R, Baquero B, Ko LK, Leng JC, Salloum RG, Vaughn NA, Brownson RC. Participatory implementation science to increase the impact of evidence-based cancer prevention and control. Cancer Causes Control. 2018;29(3):363–369. doi: 10.1007/s10552-018-1008-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 35.Kothari A, Wathen CN. Integrated knowledge translation: digging deeper, moving forward. J Epidemiol Commun H. 2017;71(6):619–623. doi: 10.1136/jech-2016-208490. [DOI] [PubMed] [Google Scholar]
  • 36.Rogers EM. Diffusion of innovations. New York: Free Press; 2003. [Google Scholar]
  • 37.Chambers DA, Glasgow RE, Stange KC. The dynamic sustainability framework: addressing the paradox of sustainment amid ongoing change. Implement Sci. 2013;8(1):117. doi: 10.1186/1748-5908-8-117. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 38.Leeman J, Sandelowski M. Practice-based evidence and qualitative inquiry. J Nurs Scholarship. 2012;44(2):171–179. doi: 10.1111/j.1547-5069.2012.01449.x. [DOI] [PubMed] [Google Scholar]
  • 39.Kononowech J, Landis-Lewis Z, Carpenter J, Ersek M, Hogikyan R, Levy C, Phibbs C, Scott W, Sales AE. Visual process maps to support implementation efforts: a case example. Implement Sci Commun. 2020;1(1):105. doi: 10.1186/s43058-020-00094-6. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 40.Chinman M, Woodward EN, Curran GM, Hausmann LRM. Harnessing implementation science to increase the impact of health equity research. Med Care. 2017;55(Suppl 9 2):S16–S23. doi: 10.1097/MLR.0000000000000769. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 41.Powell BJ, Beidas RS, Lewis CC, Aarons GA, McMillen JC, Proctor EK, et al. Methods to improve the selection and tailoring of implementation strategies. J Behav Health Ser R. 2017;44(2):177–194. doi: 10.1007/s11414-015-9475-6. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 42.Fernandez ME, ten Hoor GA, van Lieshout S, Rodriguez SA, Beidas RS, Parcel G, et al. Implementation mapping: using intervention mapping to develop implementation strategies. Front Public Health. 2019;7(7):158. doi: 10.3389/fpubh.2019.00158. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 43.Escoffery C, Lebow-Skelley E, Haardoerfer R, Boing E, Udelson H, Wood R, Hartman M, Fernandez ME, Mullen PD. A systematic review of adaptations of evidence-based public health interventions globally. Implement Sci. 2018;13(1):125. doi: 10.1186/s13012-018-0815-9. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 44.Grimshaw JM, Ivers N, Linklater S, Foy R, Francis JJ, Gude WT, Hysong SJ, Audit and Feedback MetaLab Reinvigorating stagnant science: implementation laboratories and a meta-laboratory to efficiently advance the science of audit and feedback. BMJ Qual Saf. 2019;28(5):416–423. doi: 10.1136/bmjqs-2018-008355. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 45.Balasubramanian BA, Cohen DJ, Davis MM, Gunn R, Dickinson LM, Miller WL, Crabtree BF, Stange KC. Learning Evaluation: blending quality improvement and implementation research methods to study healthcare innovations. Implement Sci. 2015;10(1):31. doi: 10.1186/s13012-015-0219-z. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 46.Rohweder C, Wangen M, Black M, Dolinger H, Wolf M, O'Reilly C, Brandt H, Leeman J. Understanding quality improvement collaboratives through an implementation science lens. Prev Med. 2019;129S:105859. doi: 10.1016/j.ypmed.2019.105859. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 47.Fernandez ME, Melvin CL, Leeman J, Ribisl KM, Allen JD, Kegler MC, et al. The cancer prevention and control research network: an interactive systems approach to advancing cancer control implementation research and practice. Cancer Epidemiol Biomarkers Prev. 2014;23(11):2512–2521. doi: 10.1158/1055-9965.EPI-14-0097. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 48.Ribisl KM, Fernandez ME, Friedman DB, Hannon PA, Leeman J, Moore A, Olson L, Ory M, Risendal B, Sheble L, Taylor VM, Williams RS, Weiner BJ. Impact of the Cancer Prevention and Control Research Network: accelerating the translation of research into practice. Am J Prev Med. 2017;52(3 Suppl 3):S233–SS40. doi: 10.1016/j.amepre.2016.08.026. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 49.Reuland DS, Cubillos L, Brenner AT, Harris RP, Minish B, Pignone MP. A pre-post study testing a lung cancer screening decision aid in primary care. BMC Med Inform Decis Mak. 2018;18(1):5. doi: 10.1186/s12911-018-0582-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 50.Brenner AT, Cubillos L, Birchard K, Doyle-Burr C, Eick J, Henderson L, Jones L, Massaro M, Minish B, Molina P, Pignone M, Ratner S, Rivera MP, Reuland DS. Improving the implementation of lung cancer screening guidelines at an academic primary care practice. J Healthc Qual. 2018;40(1):27–35. doi: 10.1097/jhq.0000000000000089. [DOI] [PubMed] [Google Scholar]
  • 51.Centers for Disease Control and Prevention. Patient Navigation. 2020. https://www.cdc.gov/screenoutcancer/patient-navigation.htm. Accessed 24 June 2021.
  • 52.Feldstein AC, Glasgow RE. A practical, robust implementation and sustainability model (PRISM) for integrating research findings into practice. Jt Comm J Qual Patient Saf. 2008;34(4):228–243. doi: 10.1016/s1553-7250(08)34030-6. [DOI] [PubMed] [Google Scholar]
  • 53.U.S. Preventive Services Task Force. Final Recommendation Statement: Lung Cancer Screening. December 31, 2013. https://www.uspreventiveservicestaskforce.org/uspstf/document/RecommendationStatementFinal/lung-cancer-screening. Accessed 30 Jan 2021.
  • 54.Centers for Medicare & Medicaid Services. National Coverage Determination (NCD) for Lung Cancer Screening with Low Dose Computed Tomography (LDCT) (210.14). February 5, 2015. https://www.cms.gov/medicare-coverage-database/details/ncd-details.aspx?NCDId=364&ncdver=1&bc=AAAAEAAAAAAA&. Accessed 30 Jan 2021.
  • 55.Glasgow RE, Harden SM, Gaglio B, Rabin B, Smith ML, Porter GC, et al. RE-AIM planning and evaluation framework: adapting to new science and practice with a 20-year review. Front Public Health. 2019;7(64). doi:10.3389/fpubh.2019.00064. [DOI] [PMC free article] [PubMed]
  • 56.National Cancer Institute. Evidence-Based Cancer Control Programs (EBCCP). https://ebccp.cancercontrol.cancer.gov/index.do. Accessed 30 Jan 2021.
  • 57.The Community Preventive Services Task Force. The Community Guide. https://www.thecommunityguide.org/topic/cancer. Accessed 30 Jan 2021.
  • 58.Davis R, D'Lima D. Building capacity in dissemination and implementation science: a systematic review of the academic literature on teaching and training initiatives. Implement Sci. 2020;15(1):97. doi: 10.1186/s13012-020-01051-6. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 59.Churruca K, Ludlow K, Taylor N, Long JC, Best S, Braithwaite J. The time has come: embedded implementation research for health care improvement. J Eval Clin Pract. 2019;25(3):373–380. doi: 10.1111/jep.13100. [DOI] [PubMed] [Google Scholar]
  • 60.Robinson TE, Janssen A, Harnett P, Museth KE, Provan PJ, Hills DJ, Shaw T. Embedding continuous quality improvement processes in multidisciplinary teams in cancer care: exploring the boundaries between quality and implementation science. Aust Health Rev. 2017;41(3):291–296. doi: 10.1071/AH16052. [DOI] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Additional file1:. (102.2KB, docx)

Data Availability Statement

Not applicable to this article as no datasets were generated or analyzed during the current study.


Articles from Implementation Science Communications are provided here courtesy of BMC

RESOURCES