Abstract
Objective
HealthyHearts NYC was a stepped wedge randomized control trial that tested the effectiveness of practice facilitation on the adoption of cardiovascular disease guidelines in small primary care practices. The objective of this study was to identify was to identify attributes of small practices that signaled they would perform well in a practice facilitation intervention implementation.
Methods
A mixed methods multiple-case study design was used. Six small practices were selected representing 3 variations in meeting the practice-level benchmark of >70% of hypertensive patients having controlled blood pressure. Inductive and deductive approaches were used to identify themes and assign case ratings. Cross-case rating comparison was used to identify attributes of high performing practices.
Results
Our first key finding is that the high-performing and improved practices in our study looked and acted similarly during the intervention implementation. The second key finding is that 3 attributes emerged in our analysis of determinants of high performance in small practices: (1) advanced use of the EHR; (2) dedicated resources and commitment to quality improvement; and (3) actively engaged lead clinician and office manager.
Conclusions
These attributes may be important determinants of high performance, indicating not only a small practice’s capability to engage in an intervention but possibly also its readiness to change. We recommend developing tools to assess readiness to change, specifically for small primary care practices, which may help external agents, like practice facilitators, better translate intervention implementations to context.
Keywords: case study, qualitative methods, mixed methods, primary care, practice transformation
Introduction
Despite declines in number, small practices (<5 providers) still employ 35.7% of physicians in the US1 and provide care to a majority of the population.2 These practices face multiple challenges in meeting quality improvement goals, in part because improvement initiatives do not often account for the limited resources and constraints of the small practice setting.3 In 2014, the Agency for Healthcare Research and Quality set out to provide support to small practices, launching EvidenceNOW: Advancing Heart Health,4 a $112 million initiative to strengthen small practices’ capacity to adopt new models of care to improve cardiovascular disease (CVD) prevention and treatment.
Our study draws on data from HealthyHearts NYC, 1 of the EvidenceNOW grantees. HealthyHearts NYC was a cluster stepped-wedge randomized control trial that evaluated the use of practice facilitation (PF) to increase adoption of CVD guidelines.[citation redacted for blinded review] PF is the “deliberate and valued process of interactive problem-solving and support”5 through an external coach, and it is an emergent implementation strategy for supporting primary care practices’ efforts to meet quality improvement (QI) goals.6-14 In HealthyHearts NYC, the intervention included 13 on-site facilitator visits over 1 year, in which facilitators worked with small practices to: educate them on the evidence-based CVD guidelines; redesign workflows to support CVD prevention and treatment; and optimize electronic health record (EHR) use for capturing and tracking patients with high risk for CVD. One of the key clinical outcomes of HealthyHearts NYC was practice-level blood pressure (BP) control.
The objective of this study was to identify attributes of small practices that signaled they would perform well in a practice facilitation intervention implementation. We use a mixed methods, multiple-case study approach15 to examine small practices from HealthyHearts NYC. Even though PF is an increasingly common implementation strategy for interventions in primary care settings,6 there are no studies examining what attributes of primary care practices may be determinants of high performance in PF interventions. This study builds on the primary care and PF literature by reporting on attributes of small primary care practices that may determine high performance during a PF intervention.
Methods
The study was approved by the NYU School of Medicine and New York City Department of Health and Mental Hygiene Institutional Review Boards. Our qualitative reporting adheres to the COREQ checklist.16
Study Setting and Study Design
HealthyHearts NYC is a partnership among the NYU School of Medicine and the Primary Care Information Project, a bureau of the New York City Department of Health and Mental Hygiene, which included 257 small practices.17 We use a mixed methods, convergence design, which is the concurrent analysis and interpretation of quantitative and qualitative data to explore the same phenomenon.18 We use the multiple-case study approach to focus on 6 small practices that participated in HealthyHearts NYC. The multiple-case study approach is often used to study complex implementations, as it allows for data triangulation of multiple sources of evidence.15 By design, multiple-case studies include a small number of distinct cases, with 3 to 4 as the recommended number of cases; small numbers allow for in-depth exploration of a phenomenon.15,19
Case Selection
We purposively sampled cases to represent wide variation in performance during the intervention implementation. To optimize data quality, we targeted practices with a patient volume of at least 500 and that used the same EHR vendor. Performance was defined using the Centers for Disease Control and Prevention’s “Million Hearts” target, in which high performance was having at least 70% of its hypertensive patients with BP <140/90.20 We create 3 BP performance categories: “high” performing practices (those that started above 70% and were now above 70%); “improved” practices (started below and were now above 70%); and “low” practices (started below and were still below 70%). About 9 months into the 1-year intervention, we assessed practices’ BP data to recruit 2 cases to correspond with each performance category, resulting in a final sample of 6 cases. Once cases were recruited, we collected qualitative data (described below).
Quantitative Data Sources and Measures
Practice-level BP data and number of patients came from the practices’ EHR. Practice-level characteristics were captured at baseline through the Practice Survey (completed by practice staff), including: number of providers, number of staff, and Patient-Centered Medical Home status. We extracted data on location in a medically underserved area from the U.S. Health and Human Services website.21 Clinician-level characteristics were captured post-intervention through the Clinician Survey (completed by the lead clinician), including: race, ethnicity, sex, age, whether born in the U.S., and whether graduated from a U.S. medical school.8
Qualitative Data Sources
Qualitative data sources included 3 semi-structured interviews with practice stakeholders (lead clinician, staff, and facilitator), facilitator visit notes, and ethnographic site observations. All interviews were 1-h, semi-structured, in-person (ie, in the practice), included only the interviewer and the participant, audio-recorded, and transcribed. The interview guides were tested internally and not shared with participants in advance. The research team did not have existing relationships with the practices and approached the sites by email and phone with the research objectives and requests for participation.
Interviews with the clinician and staff took place September 2016 to July 2017 were conducted by CB, DS, ER, and SK (all female; faculty with qualitative research experience). Interviews began with brief introductions about the interviewers’ roles and research goals and focused broadly about the implementation.22 All clinicians participated (n = 6) and 83.3% of staff members (n = 5), with 1 declining due to scheduling challenges; each was paid $50. Facilitator interviews took place July to August 2017 and were conducted by AC (female; project manager) and DS, began with brief introductions about the interviewers’ roles and research goals, and focused on experiences with the specific case study site. About 100% of facilitators agreed to participate (n = 6). Facilitators were offered a $25 honorarium but not required to participate. Transcripts were not returned to participants for comment.
Three researchers with formal ethnography training (AC, PN [female, researcher], LJ [female, project coordinator]) conducted site observations of a facilitator visit to each site in March-May 2017. Researchers observed for 2 to 3 h, took field notes by hand during the visit, typed up field notes immediately after the visit, and then debriefed with the research team. All practices agreed to participate (n = 6) and were paid $100.
Analytical Approach
We used inductive and deductive approaches in the context of an immersion/crystallization process, in which the qualitative team looked closely at the data (immersion) and then took a step back to identify patterns across the cases (crystallization).23 First, the full qualitative team (AC, AN [female, researcher with mixed methods training], DS, ER, NVD [female, faculty with mixed methods training]) read documents for each case, blinded of the performance category, and met iteratively to develop a codebook (see Supplement A. Codebook). Two researchers (AN & AC) then used the codebook to independently code the documents using Dedoose® software,24 meeting weekly to review codes and create case memos. The full team met to discuss the memos and identified 3 categories for examination across cases: (1) use of EHR, (2) resources for QI, and (3) leadership engagement. Next, we used a cross-case rating comparison. The 2 coders used the criteria shown in Table 125 to review quotes in each document and assign ratings. Each case’s overall rating was calculated as the average of its documents’ ratings.26 The full team met to discuss the ratings, identifying attributes (major themes and minor themes) that were determinants of high performance.
Table 1.
Rating Criteria for Qualitative Documents.
| Rating | Criteria |
|---|---|
| −1 | There is lack of evidence of the theme. This rating indicates that the theme does not have a positive influence on the practice or does not facilitate implementation efforts. The interviewee makes general statements about the theme manifesting in a negative way but without concrete examples: The theme is mentioned only in passing or at a high level without examples or evidence of actual, concrete descriptions; There is a mixed effect of different aspects of the theme but with a general overall negative effect; There is sufficient information to make an indirect inference about the generally negative influence; and/or Judged as weakly negative by the absence of the theme |
| 0 | The influence of the theme was unclear or neutral: It appears to have neutral effect (purely descriptive) or is only mentioned generically without valence; There is no evidence of positive or negative influence; Interviewee contradicts his/herself; There are positive and negative influences at different levels of the practice that balance each other out; and/or different aspects of the theme have positive influence while others have negative influence and overall, the effect is neutral |
| +1 | There is evidence of the theme. This rating indicates that the theme has a positive influence on the practice or facilitated implementation efforts. The interviewee describes explicit examples of how the theme manifests itself in a positive way using concrete examples: The theme is mentioned with examples or evidence of actual, concrete descriptions of how that construct manifests; There is a general overall positive effect; and/or There is sufficient information to make an indirect inference about the generally positive influence. |
| Missing | Interviewee(s) were not asked about the presence or influence of the construct; or if asked, their responses did not correspond to the intended theme and were instead coded to another construct. Interviewee(s)’ lack of knowledge about a theme does not necessarily indicate missing data and may indicate the absence of the theme. |
Source: Adapted from Damschroder and Lowery.25
Results
Table 2 shows the characteristics of the 6 cases, referred to by monikers such as High #1 and High #2. All were small, independent practices that completed the intervention as prescribed (at least 13 facilitator visits, indicating fidelity to the implementation plan). The low-performing cases had more staff than the high-performing and improved cases. The practice with the highest ratio of patients to providers was Improved #2, followed by Low #2. Both high-performing cases and Improved #1 were located in a medically underserved area. The lead clinicians’ characteristics were similar across all cases.
Table 2.
Profile of Cases.
| High #1 | High #2 | Improved #1 | Improved #2 | Low #1 | Low #2 | |
|---|---|---|---|---|---|---|
| Practice characteristics | ||||||
| # of FTE providers* | 1 | 2 | 1 | 1 | 2 | 2 |
| # of FTE staff | 2 | 2 | 3 | 2 | 8 | 6 |
| # of patients | 943 | 2044 | 727 | 2384 | 1303 | 4424 |
| Patient-centered medical home status | Recognized | Recognized | Recognized | Recognized | Recognized | Recognized |
| Located in a medically underserved area | Yes | Yes | Yes | No | No | No |
| Lead clinician characteristics | ||||||
| Race | Asian | Black/African American | Black/African American | Asian | Black/African American | Black/African American |
| Ethnicity | Non-Hispanic | Non-Hispanic | Non-Hispanic | Non-Hispanic | Hispanic | Non-Hispanic |
| Sex | Male | Female | Male | Female | Male | Male |
| Age | 55 to 64 | 65+ | 55 to 64 | 45 to 54 | 55 to 64 | 55 to 64 |
| Born in U.S. | No | No | No | No | Yes | No |
| Graduated from medical school in U.S. | No | Yes | No | Yes | Yes | Yes |
Note. FTE = Full-time equivalent; *Providers were defined as physicians, nurse practitioners, and physician assistants who provided primary care. We did not collect data on specialty type.
Our first key finding is: the high-performing and improved cases in our study looked and acted similarly during the intervention implementation. The 2 high-performing cases and the 2 improved cases have nearly identical characteristics (Table 1) and case ratings across the 3 code categories, “use of the EHR,” “resources for QI,” and “leadership engagement” (see Supplement B. Case ratings). These characteristics and ratings were distinctly different from the 2 low-performing cases.
The second key finding is that the high-performing and improved cases exhibited the following attributes (major themes) that may have been determinants of their high BP performance: (1) advanced use of the EHR; (2) dedicated resources and commitment to QI; and (3) actively engaged lead clinician and office manager. Given the similar nature of the high-performing and improved cases, below, we describe these 4 cases together, contrasting them with the 2 low-performing cases. Minor attributes (minor themes) are also described throughout.
Attribute 1: Advanced Use of EHR
The high-performing and improved cases all demonstrated advanced use of their EHR systems (major attribute) and a belief in the value of EHRs to facilitate high quality care. A facilitator described:
“This practice is a top spot on my list because they are using the HIT (health information technology). They’re strongly relying on and believe in HIT, and they’re using the HIT. That takes out a lot of my EHR teaching time.” (Improved #1: Facilitator)
These practices were able generate their own QI reports, which allowed them to engage with facilitators on more advanced EHR activities (minor attribute), such as discussing the use of reports to track intervention progress. To give an example, a staff member at High #2 and facilitator were observed reviewing a BP report together, where the staff member asked about how to compare data, rather than how to create the report: “How did we do on this measure last month? How are we in relation to other practices?” (High #2: Observation)
In comparison, the low cases reported having technical issues that they were unable to or did not have time to address themselves. During the intervention, facilitators spent many visits troubleshooting EHR issues to ensure the practice could document and assess the study’s measures accurately. A facilitator described a common experience in low-performing practices:
“Trying to connect to the EHR is sometimes challenging. If they need a ticket put into [EHR vendor], I actually put the ticket in for them with my name, so I can make sure it gets fixed.” (Low #2: Facilitator)
Attribute 2: Dedicated Resources and Commitment to QI
The high-performing and improved cases demonstrated dedicated resources and commitment to QI (major attribute), evidenced by protected time for specified staff to conduct QI activities. For example, High #2 specified a nurse as the point person who consistently met with the facilitator and monitored QI projects. The facilitator described the value having designated QI staff:
“Having a dedicated person in a QI role helps move the work along. It gives consistency. It’s hard to make progress when you don’t always meet with same person, especially if they don’t have the same level of EHR knowledge. You end up repeating what you did at the last visit.” (High #2: Facilitator)
Further, clinicians in these cases recognized the role and importance of data in QI (minor attribute). For example, High #1 described how they routinely used data to improve care quality:
“We get monthly or quarterly data from our independent practice association and from insurance companies. We look at the data on each doctor. We look at the performance. We have quality markers.” (High #1: Clinician)
The combination of having dedicated resources and a practice-level commitment to QI were indicative of a high-performing practice.
In comparison, the low-performing cases did not have sufficient nor consistent dedicated QI resources and commitment, despite having more staff. A lack of clarity in role definition for QI responsibilities as well as staff turnover may have contributed to reduced QI capacity, as exemplified:
“The practice had a glitch at one point [during the intervention] when there was a staff change. I had to speak with the admin during the time, not the office manager, because she was busy running the front desk.” (Low #1: Facilitator)
Attribute 3: Actively Engaged Lead Clinician and Office Manager
In the high-performing and improved cases, both the lead clinician and office manager were actively engaged in the intervention (major attribute). Office managers met with the facilitators every visit, with the lead clinicians also participating in most. One facilitator described the impact of high engagement: “You really need to have someone that’s really interested and engaged with you in order to be successful.” (High #1: Facilitator) The lead clinician and office manager were also able to recognize how the intervention goals were aligned with their own practice’s QI goals (minor attribute):
“I’m constantly involved in the HealthyHearts NYC project because it’s not just a HealthyHearts process. It’s the Patient-Centered Medical Home process.” (Improved #1: Clinician)
For the low-performing cases, the lead clinicians were not actively engaged with the intervention; facilitators documented meeting primarily with the office manager and/or other staff. These clinicians shared that they wanted to be more engaged with QI initiatives but did not have sufficient time, as they were “busy seeing patients” (Low #1: Clinician). Notably, Low #2 had 1 of the highest provider to patient ratios. This led to the clinicians “leaving everything to the office manager” (Low #2), and office managers subsequently feeling “overwhelmed” (Low #2: Facilitator) from QI activities on top of their usual responsibilities.
Discussion
This case study provided insight on attributes of small practices that may be associated with high performance in a practice facilitation-led guideline implementation. High-performing and improved cases demonstrated advanced use of the EHR, dedicated resources and commitment to QI, and an actively engaged lead clinician and office manager. Due these attributes, the practices that were already high-performing at baseline may have been able to quickly integrate the new CVD guidelines into their workflows, and the practices that improved performance over the intervention may have had greater readiness to leverage facilitator support to make changes. Our findings thus suggest that the identified attributes indicate the capability of a small practice to engage in the intervention.
There are no studies on attributes that determine high performance among small practices and thus we turn to the literature on large healthcare organizations to see if there are synergies. The attributes we identified in our study of small practices are surprisingly consistent with work that have identified attributes that determine high-performing large healthcare organizations. For example, in the study by D’Aunno et al. (2016), 3 factors that determined high-performing accountable care organizations in the Medicare Shared Savings program align with the attributes in our study: “sophisticated use of information systems,” “trusted, long-standing physician leaders focused on improving performance,” and “embedded care coordinators.”27 In the study by Mannion et al. (2005), 2 factors that determined high-performing hospitals included: “leadership and management orientation” and “accountability and information systems.”28 In Damschroder and Lowery (2013), 2 of the identified CFIR constructs that determined effective VA facilities align with ours: learning climate and leadership engagement.25 Even though accountable care organizations, hospitals, and VA facilities are very different types of organizations than small primary care practices, it was not surprising that these studies all reported health IT, QI resources, and leadership engagement as core to better performance—these attributes are fundamental to measuring, monitoring, and improving care quality. Furthermore, we posit that these attributes speak to the need for an organization to have high levels of readiness to change to successfully implement a QI initiative.
While we did not measure readiness to change, the attributes identified in our study have similarities to the conceptual literature on organizational readiness to change.29 For example, prior studies have recommended that facilitators be equipped with a baseline assessment tool to assess organizational readiness for engagement in QI.30,31 However, there are currently no publicly-available measures of readiness for change for small practices.32 There is thus an opportunity to develop an organizational readiness tool specifically for small practices that could draw upon the findings of our study. Two existing tools serve as strong starting points: (1) the Quality Improvement Capacity Assessment (QICA), which is based on medical home principles and assesses capacity for QI in small practices, a component of readiness for change33 and (2) the Organizational Readiness to Change Assessment instrument (ORCA), which has been widely used to assess implementation of evidence-based practices in large health organizations.34 Neither assess EHR skills and competencies, however, which our study suggests is important to the success of QI projects. Prior work has also demonstrated that an EHR system’s limited capabilities can hinder QI in primary care practices.35 We recommend research to create functional measures of EHR readiness that go beyond asking practices to self-report the presence of EHR functions or their perceived skill level. Such a tool could be utilized by implementation scientists and researchers to inform their knowledge translation strategies36 and by practitioners to determine small primary care practices’ potential for change.
Limitations
Due to the New York City setting, findings may have limited generalizability to other settings. The multiple-case study approach, however, is not intended for wide generalizability, but rather, to gain insight on a specific phenomenon.19 Due to the case study design and resources, we were limited in the data we could use to assess thematic saturation; however, we believe thematic saturation was achieved by following case study best practices, which includes obtaining data from multiple sources (survey data, EHR data, interviewers with 3 stakeholders types, and site observations for every case) in a bounded time and space15,19 and triangulating the data.37-39 For the interviews and site observations, we minimized potential bias due to the interviewers’ knowledge of the primary care literature by utilizing a semi-structured interview/observation guide. For analyses, we used a multidisciplinary team and deliberately included co-author NVD, who had not been involved in the study up to this point to minimize groupthink, and the HPS and SS from the Department of Health and Mental Hygiene for external validation.
Conclusion
High-performing small practices exhibit attributes that may signify their ability and readiness for change during an intervention. For implementation scientists and practitioners, our study highlights the importance of assessing organizational readiness for change and developing functional measures of EHR readiness, specifically for the small practice setting. Appropriate readiness assessment may help external agents, like practice facilitators, better tailor intervention implementations to context and ultimately help small practices improve their care quality.
Supplemental Material
Supplemental material, sj-docx-1-jpc-10.1177_2150132720984411 for Attributes of High-Performing Small Practices in a Guideline Implementation: A Multiple-Case Study by Ann M. Nguyen, Allison M. Cuthel, Erin S. Rogers, Nancy Van Devanter, Hang Pham-Singer, Sarah Shih, Carolyn A. Berry and Donna R. Shelley in Journal of Primary Care & Community Health
Acknowledgments
The authors would like to thank Nina Siman, MA MSed and Thomas Gepts, BA for their support processing the quantitative data, creating case definitions, and generating figures; Sue Kaplan, JD for conducting interviews; and Paulomi Niles, CNM LM MPH and Lily Jordahl for conducting site observations.
Footnotes
Declaration of Conflicting Interests: The author(s) declared no potential conflicts of interest with respect to the research, authorship, and/or publication of this article.
Funding: The author(s) disclosed receipt of the following financial support for the research, authorship, and/or publication of this article: This project was funded under grant number 1R18HS023922 from the Agency for Healthcare Research and Quality (AHRQ), U.S. Department of Health and Human Services (HHS). The authors are solely responsible for this document’s contents, findings, and conclusions, which do not necessarily represent the views of AHRQ. Readers should not interpret any statement in this report as an official position of AHRQ or of HHS.
ORCID iD: Ann M. Nguyen
https://orcid.org/0000-0002-7903-0223
Supplemental Material: Supplemental material for this article is available online.
References
- 1. Kane CK. Updated Data on Physician Practice Arrangements: For the First Time, Fewer Physicians Are Owners Than Employees. Chicago, IL; 2019. https://www.ama-assn.org/system/files/2019-07/prp-fewer-owners-benchmark-survey-2018.pdf. [Google Scholar]
- 2. Rui P, Okeyode T. National Ambulatory Medical Care Survey: 2016 National Summary Tables; 2016. Accessed October 30, 2019 https://www.cdc.gov/nchs/data/ahcd/namcs_summary/2016_namcs_web_tables.pdf.
- 3. Miller WL, Crabtree BF, McDaniel R, Stange KC. Understanding change in primary care practice using complexity theory. J Fam Pract. 1998;46:369-376. [PubMed] [Google Scholar]
- 4. U.S. Department of Health & Human Services. EvidenceNOW: Advancing Heart Health in Primary Care. https://www.ahrq.gov/evidencenow/index.html.
- 5. Stetler CB, Legro MW, Rycroft-Malone J, et al. Role of “external facilitation” in implementation of research findings: a qualitative evaluation of facilitation experiences in the Veterans Health Administration. Implement Sci 2006;1. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 6. Baskerville NB, Liddy C, Hogg W. Systematic review and meta-analysis of practice facilitation within primary care settings. Ann Fam Med. 2012;10:63-74. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 7. Liddy C, Hogg W, Singh J, et al. A real-world stepped wedge cluster randomized trial of practice facilitation to improve cardiovascular care. Implement Sci. 2015;10:150. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8. Shelley DR, Ogedegbe G, Anane S, et al. Testing the use of practice facilitation in a cluster randomized stepped-wedge design trial to improve adherence to cardiovascular disease prevention guidelines: HealthyHearts NYC. Implement Sci. 2015;11:88. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 9. Chase SM, Nutting PA, Crabtree BF. How to solve problems in your practice with a new meeting approach. Fam Pract Manag. 2010;17:31-34. [PMC free article] [PubMed] [Google Scholar]
- 10. Hemler JR, Hall JD, Cholan RA, et al. Practice facilitator strategies for addressing electronic health record data challenges for quality improvement: EvidenceNOW. J Am Board Fam Med 2018;31:398-409. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 11. Laferriere D, Liddy C, Nash K, Hogg W. Navigating change: how outreach facilitators can help clinicians improve patient outcomes. J Am Board Fam Med. 2012;25:232-237. [DOI] [PubMed] [Google Scholar]
- 12. Harvey G, Lynch E. Enabling Continuous quality improvement in practice: the role and contribution of facilitation. Front Public Heal. 2017;5:27. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 13. Wang A, Pollack T, Kadziel LA, et al. Impact of practice facilitation in primary care on chronic disease care processes and outcomes: a systematic review. J Gen Intern Med. 2018;33:1968-1977. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14. Nguyen AM, Cuthel A, Padgett DK, et al. How practice facilitation strategies differ by practice context. J Gen Intern Med. 2020;35:824-831. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15. Yin RK. Case Study Research: Design and Methods (Applied Social Research Methods). 5th ed. Sage; 2013. [Google Scholar]
- 16. Tong A, Sainsbury P, Craig J. Consolidated criteria for reporting qualitative research (COREQ): a 32-item checklist for interviews and focus groups. Int J Qual Heal Care. 2007;19:349-357. [DOI] [PubMed] [Google Scholar]
- 17. Shelley DR, Ogedegbe G, Anane S, et al. Testing the use of practice facilitation in a cluster randomized stepped-wedge design trial to improve adherence to cardiovascular disease prevention guidelines: HealthyHearts NYC. Implement Sci. 2015;11(1):88. doi: 10.1186/s13012-016-0450-2 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 18. Palinkas LA, Aarons GA, Horwitz S, Chamberlain P, Hurlburt M, Landsverk J. Mixed method designs in implementation research. Adm Policy Ment Heal Ment Heal Serv Res. 2011;38:44-53. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 19. Schoch K. Research design and methods: an applied guide for the scholar-practitioner. In: Burkholder GJ, Cox KA, Crawford LM, Hitchcock JH, Patton MQ, eds. Selected Research Designs and Approaches. SAGE Publications, Inc; 2020: 245-258. [Google Scholar]
- 20. Centers for Disease Control and Prevention. Million Hearts: Strategies to reduce the prevalence of leading cardiovascular disease risk factors—United States, 2011. MMWR Morb Mortal Wkly Rep. 2011;60:1248-1251. Accessed April 25, 2019 https://www.cdc.gov/mmwr/preview/mmwrhtml/mm6036a4.htm. [PubMed] [Google Scholar]
- 21. Health Resources and Services Administration. MUA Find. data.HRSA.gov. Accessed August 2, 2019 https://data.hrsa.gov/tools/shortage-area/mua-find.
- 22. Damschroder LJ, Aron DC, Keith RE, Kirsh SR, Alexander JA, Lowery JC. Fostering implementation of health services research findings into practice: a consolidated framework for advancing implementation science. Implement Sci. 2009;4:50. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 23. Borkan J. Immersion/crystallization. In: Crabtree B, Miller W, eds. Doing Qualitative Research. SAGE Publications; 1999: 179-194. [Google Scholar]
- 24. SocioCultural Research Consultants. Dedoose. 2019. http://www.dedoose.com.
- 25. Damschroder LJ, Lowery JC. Evaluation of a large-scale weight management program using the consolidated framework for implementation research (CFIR). Implement Sci. 2013;8:1-17. Accessed September 17, 2018 https://implementationscience.biomedcentral.com/track/pdf/10.1186/1748-5908-8-51. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 26. Miles MB, Huberman AM, Saldaña J. Qualitative Data Analysis: A Methods Sourcebook. 3rd ed. Thousand Oaks, CA: SAGE Publications; 2013. [Google Scholar]
- 27. D’Aunno T, Broffman L, Sparer M, Kumar SR. Factors that distinguish high-performing accountable care organizations in the medicare shared savings program. Health Serv Res. 2018;53:120-137. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 28. Mannion R, Davies HTO, Marshall MN. Cultural characteristics of “high” and “low” performing hospitals. J Health Organ Manag. 2005;19:431-439. [DOI] [PubMed] [Google Scholar]
- 29. Lehman WEK, Greener JM, Simpson DD. Assessing organizational readiness for change. J Subst Abuse Treat. 2002;22:197-209. [DOI] [PubMed] [Google Scholar]
- 30. Liddy C, Rowan M, Valiquette-Tessier S-C, Drosinis P, Crowe L, Hogg W. Experiences of practice facilitators working on the Improved Delivery of Cardiovascular Care project. Can Fam Physician. 2018;64:e23-e32. [PMC free article] [PubMed] [Google Scholar]
- 31. Ritchie MJ, Dollar KM, Miller CJ, et al. Using Implementation Facilitation to Improve Care in the Veterans Health Administration (Version 2); 2017. Accessed June 11, 2019 https://www.queri.research.va.gov/tools/implementation/Facilitation-Manual.pdf.
- 32. Weiner BJ, Amick H, Lee S-YD. Conceptualization and measurement of organizational readiness for change: a review of the literature in health services research and other fields. Med Care Res Rev. 2008;65:379-436. [DOI] [PubMed] [Google Scholar]
- 33. Parchman ML, Anderson ML, Coleman K, et al. Assessing quality improvement capacity in primary care practices. BMC Fam Pract. 2019;20:103. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 34. Helfrich CD, Li Y-F, Sharp ND, Sales AE. Organizational readiness to change assessment (ORCA): Development of an instrument based on the Promoting Action on Research in Health Services (PARIHS) framework. Implement Sci. 2009;4:38. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 35. Cohen DJ, Dorr DA, Knierim K, et al. Primary care practices’ abilities and challenges in using electronic health record data for quality improvement. Health Aff. 2018;37:635-643. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 36. Peters S, Bussières A, Depreitere B, et al. Facilitating guideline implementation in primary health care practices. J Prim Care Community Health. 2020;11:215013272091626. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 37. Kimchi J, Polivka B, Stevenson JS. Triangulation: operational definitions. Nurs Res. 1991;40:364-366. [PubMed] [Google Scholar]
- 38. LeRouge C, Dickhut K, Lisetti C, Sangameswaran S, Malasanos T. Engaging adolescents in a computer-based weight management program: avatars and virtual coaches could help. J Am Med Informatics Assoc. 2016;23:19-28. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 39. Rogers ES, Cuthel AM, Berry CA, Kaplan SA, Shelley DR. Clinician perspectives on the benefits of practice facilitation for small primary care practices. Ann Fam Med. 2019;17(suppl 1):S17-S23. [DOI] [PMC free article] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Supplementary Materials
Supplemental material, sj-docx-1-jpc-10.1177_2150132720984411 for Attributes of High-Performing Small Practices in a Guideline Implementation: A Multiple-Case Study by Ann M. Nguyen, Allison M. Cuthel, Erin S. Rogers, Nancy Van Devanter, Hang Pham-Singer, Sarah Shih, Carolyn A. Berry and Donna R. Shelley in Journal of Primary Care & Community Health
