Abstract
Learning health systems (LHSs) are designed to systematically integrate external evidence of effective practices with internal data and experience to put knowledge into practice as a part of a culture of continuous learning and improvement. Researchers embedded in health systems are an essential component of LHSs, with defined competencies. However, many of these competencies are not generated by traditional graduate/post-graduate training programs; evaluation of new LHS training programs has been limited. This commentary reviews and extends results of an evaluation of early career outcomes of fellows in one such program designed to generate impact-oriented career pathways embedded in healthcare systems. Discussion considers the need for increasingly rigorous evaluation methods to ensure production of high-quality professionals ready for system engagement, the importance of training and preparing other LHS stakeholders as effective partners and evidence users, and the promise and challenges in advancing the science and practice of embedded research in LHSs.
Keywords: Learning Health Systems, Embedded Research, Implementation Science, Training Programs, Research Impacts
The return on investment of health research has long been challenged by the “voltage drop” in outcomes from clinical research focused on demonstrating efficacy to that actually realized in actual healthcare settings and experienced by their patients.1 This “research to real world” knowledge gap is at the heart of efforts to lift up and further build health systems research and implementation science capacity and ensure their relevance and impacts through partnerships with healthcare delivery systems.2 The promise of learning health systems (LHSs), where such multilevel and engaged partnerships are organized and supported, is in their ability to use research and evaluation evidence linked to operational priorities to transform care and outcomes.3
While LHS competencies have been defined, multifaceted programs described, and career trajectories postulated,4,5 little is known about effective strategies for training this next generation of applied and embedded scholars. Kasaai et al, recently reported on the outcomes of just such a program, focused on early career trajectories of postdoctoral research fellows after participation in a Health System Impact Fellowship.6 The Fellowship was designed to confer LHS competencies prioritized by potential employers (eg, leadership, change management, negotiation) through relevant didactics, experiential training embedded in a health system (organized around a project of high priority to the organization), and co-mentorship from a university-based academic and health system leader, in the context of a 2-year cohort-based approach to training.6 Overall, they found that all graduated fellows were employed and in diverse sectors (eg, academic, public, healthcare delivery, and private sectors), chiefly in alignment with their career aspirations at program onset. Self-reported career preparedness and readiness were high.
Evaluations of efforts to build embedded research competencies in service of achieving LHS outcomes are critical at the same time they raise important questions (Figure 1). In Kasaai et al, full employment post-fellowship was laudable, as was trainee multisectoral distribution, and yet the majority ended up in academia. The absence of prior employment patterns and comparison groups makes it difficult to know if their results are attributable to an existing time trend (no program effect), a positive programmatic outcome (as suggested), or worse, a program failure, if the goal was to primarily embed graduates in delivery systems. Consistency between initial career aspirations and employment outcomes is similarly challenging to interpret. While only about half of fellows who aspired to become an embedded researcher did so, the program’s focus on multidisciplinary collaboration, health system embedding, and networking may have yielded opportunities not previously considered. Academic mentors may also have lacked embedded research experience themselves, so fellows may have modeled themselves toward academia, whereas being health system leaders (who served as co-mentors) was neither a career aspiration nor expected program outcome. Information on curricular details (eg, appropriate theories, systems science, implementation science, policy evaluation) and engagement content, quality, or time spent by different mentor types was not available, though these graduates likely blazed new trails across sectors, armed with skills and experience not common among their predecessors. To the authors’ credit, they acknowledged many of their evaluation limitations, including the lack of comparison group, small sample size, reliance on online data sources, and limited time since graduation (2-4 years). Training impacts need to be assessed for longer term career outcomes (eg, retention, promotion, career trajectories), traditional indicators of productivity (eg, grants funded, papers published), and very importantly, on the health systems that hire them (eg, operations projects approved, levels of multilevel engagement achieved, impacts on care delivery, patient outcomes, cost savings). The authors also noted a significant loss to follow-up for post-program surveys, which could have contributed to bias in measures of readiness, preparedness, and satisfaction with supervisors’ support of their career prospects. Less attention was paid to the need for more validated measures, lack of a mixed methods approach, and a more rigorous post-program survey that might have illuminated what aspects of the program fellows thought were most effective. Information about the nature of the projects that fellows designed and conducted in their respective health systems would also be useful in gauging the quality and caliber of their embedded work, in addition to the multilevel engagement and partnerships they garnered and the outcomes they achieved during their relatively brief tenures. Evaluation of target health system contexts is also key, including the level of LHS implementation, availability of organizational supports to conduct embedded research, and the cultural milieu that either helps or hinders the kinds of learning loops and multilevel stakeholder engagement that differentiates these kinds of models.7 Evaluation of their competencies in practice from their new employers would also be an important contribution, as would assessments of their training needs over the long term, while also tracking their mentorship of the next generation of embedded researchers.5 The challenge is how to obtain adequate funding to pursue rigorous evaluations that address these and other methodological considerations on the path to improved outcomes of embedded research and furthering the business case for adoption and effective integration of embedded researchers.8
Advancing LHS functions and tenets is obviously not just about building embedded research capacity and will require going far beyond training researchers to work effectively in health systems, especially given how variably health systems have actually adopted essential LHS elements.9 Instead, the path to LHS adoption and maturity requires reciprocally priming the health system by building and supporting the necessary infrastructure, processes, and training (eg, numeracy) to be an effective “host” if not “receptor” for embedded researchers (ie, organizational symbiosis, co-production, synergies), ideally where the whole is greater than the sum of its parts.10 Health system leaders and managers need training on how to pose questions to their own data systems with and through embedded staff, how to appraise evidence worth implementing (ie, when is there enough evidence that warrants changing existing practice to new routines), how to balance evidence and improvement value in relation to the costs of change (and potential costs of inertia), and how to optimally integrate embedded researchers in health system workflows and operations. The extent to which employers and staff are themselves ready for LHS implementation and engagement with embedded researchers is not known, though experience suggests substantial work is needed on all levels of health system organizations.
While, in some respects, the business case for training and employing embedded researchers in a LHS has never been clearer, initiatives that clearly and consistently communicate their value to stakeholders within and outside the health sector are needed.8 Some of the push-pull to adequately adopt LHS functions in health system organizations may be a result of historical failures of in-house continuous quality improvement initiatives to achieve hoped-for improvements in care.11 In contrast, evidence-based quality improvement, where researchers partner with health system leaders is an effective embedded research strategy, which may confer greater impacts if not a competitive advantage in enhancing LHS functions (eg, self-efficacy implementing new care models, change-readiness, team-based communication, team function), while also reducing burnout.12,13 Better training programs for all LHS stakeholders are essential, but at the core of some of these challenges is the persistent lack of a shared language, common frameworks, aligned incentives, and meaningful multilevel and interdisciplinary engagement anchored in diverse contexts of health systems and the patients they serve.2,7
Ultimately, the science of embedded research needs to more strongly meld with the pragmatics and evidence needs of health system practice and policy in an LHS context. Nowhere is that more apparent than in starting with the nature of the roles between health system decision-makers and researchers. Historically, research has been almost exclusively in the purview of university-based academic faculty who may study a system and whose external vantage point has been lauded as central to objective scientific inquiry as independent agents of “truth.” However, this model of external evidence generation, while broadly useful and important in its own right, typically treats decision-makers as passive recipients of evidence rather than partners in agenda-setting let alone real-world application of evidence-based improvements, thus limiting the potential for LHS development let alone self-reliance. In contrast, the US Department of Veterans Affairs (VA) has focused on an embedded research-operations partnership model, with an intramural research program aligned with system priorities and Veteran needs.VA health systems researchers, many of whom also have academic affiliations, are oriented to the importance of operations leader engagement before, during, and after each project. Funding may be predicated on system partners demonstrating commitment to implementing research results, while Veteran engagement is also a scored element.Many national VA program offices overseeing practice and policy in specific clinical and policy areas (eg, primary care, women’s health, health equity) also directly fund evaluation, frequently relying on the embedded research workforce given their training, experience, system knowledge (eg, many also deliver VA clinical care), and attention to contexts of the system, its providers and staff, and patients, moving discovery to system-wide change.14 In yet another model, the World Health Organization (WHO) has invested in an initiative that actually requires decision-makers to be the research leaders, given their system-level authority over practice and policy; early evidence suggests promising outcomes across diverse international exemplars, though such models may require an uncommon degree of research acumen among system leaders.7 Both partnership models, as well as many others underway, require recognition and optimization of the roles of researchers as sources of independent and objective evidence, including when that evidence highlights problems. Lessons across these and other diverse LHS experiments suggest the vital importance of recognizing the value of (1) embedded researchers trained and supported to be responsive to system and patient needs, (2) their routine engagement of and collaboration with multilevel decision-makers, as well as frontline providers and staff, in all aspects of embedded research, (3) their explicit attention to community and system contexts in the design, conduct, and use of research and data, with evidence and communication/coordination tailored to local needs, and (4) an embrace of cyclical knowledge generation and use of rigorous, practice-based evidence in deploying innovations and implementing evidence-based practices.7,14
In reality, it is very difficult to achieve the tenets of an LHS without embedded researchers, and their value proposition spans broad LHS principles3 (eg, science, informatics, incentives, culture) (Figure 2). For science, they serve as enablers of high-quality evidence generation, rigorous evaluation, and implementation of evidence-based practice, while also serving as internal experts and trainers of other LHS stakeholders. For informatics, their training may help ensure health systems rely on valid data and continuously improve encounter data, while adding other data (eg, provider/staff, organizational, area data) that refine the questions decision-makers may pose and researchers may help answer. Others may be able to advance data tools and electronic health record based improvements at the point of care or for practice-wide improvement, while their role in improved statistical modeling and machine learning will enhance health systems’ ability to explore the value of artificial intelligence in improving care quality and efficiency. For incentives, some may be anchored in traditional markers toward academic promotion (eg, funding, peer-reviewed papers) but others likely reflect the culture of embedded researchers to have their research make a difference on system performance. And, for culture, embedded researchers often become central to an array of improvement initiatives, supporting rapid cycle improvement, adapting evidence to local contexts, and enhancing collaboration in ways that empower teams and lowers burnout.13 When researchers are integral to the delivery system, enhancing productive interactions with leaders and frontline providers and staff alike, within and across care delivery settings, they begin to embody the iterative clinical and research lifecycle that is central to LHS functions.
While the relevant literature in this area has grown at a fast pace, increasingly drawing from diverse disciplines and fields of inquiry (eg, community engaged research, participatory action research, learning theories, systems science, informatics/data science, engagement science, implementation science, policy evaluation),4,7,8 rigorous evaluation of novel training approaches that take these lessons and needs into account and explore different models of embedded research in support of advancing LHSs remain critical to advancing the field.15
Ethical issues
Not applicable.
Conflicts of interest
Author declares that she has no conflicts of interest.
Disclaimer
The views expressed in this article are those of the author and do not necessarily reflect the position or policy of the US Department of Veterans Affairs or the US Government.
Citation: Yano EM. Training an embedded workforce to realize health system impacts and the promise of learning health systems: Comment on "Early career outcomes of embedded research fellows: an analysis of the health system impact fellowship program." Int J Health Policy Manag. 2024;13:8667. doi:10.34172/ijhpm.8667
Funding Statement
Dr. Yano’s effort is covered by a VA Health Systems Research Senior Research Career Scientist Award (RCS #05-195) from the VA Office of Research & Development, Veterans Health Administration, US Department of Veterans Affairs.
References
- 1.Yano EM, Green LW, Glanz K, et al. Implementation and spread of interventions into the multilevel context of routine practice and policy: implications for the cancer care continuum. J Natl Cancer Inst Monogr. 2012;2012(44):86–99. doi: 10.1093/jncimonographs/lgs004. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 2.Kilbourne AM, Braganza MZ, Bowersox NW, et al. Research lifecycle to increase the substantial real-world impact of research: accelerating innovations to application. Med Care. 2019;57(10 Suppl 3):S206–S212. doi: 10.1097/mlr.0000000000001146. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 3. Smith M, Saunders R, Stuckhardt L, McGinnis JM. Best Care at Lower Cost: The Path to Continuously Learning Health Care in America. Washington, DC: The National Academies Press; 2013. [PubMed]
- 4.Forrest CB, Chesley FD Jr, Tregear ML, Mistry KB. Development of the learning health system researcher core competencies. Health Serv Res. 2018;53(4):2615–2632. doi: 10.1111/1475-6773.12751. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 5.Yano EM, Resnick A, Gluck M, Kwon H, Mistry KB. Accelerating learning healthcare system development through embedded research: Career trajectories, training needs, and strategies for managing and supporting embedded researchers. Healthc (Amst) 2021;8 Suppl 1:100479. doi: 10.1016/j.hjdsi.2020.100479. [DOI] [PubMed] [Google Scholar]
- 6.Kasaai B, Thompson E, Glazier RH, McMahon M. Early Career outcomes of embedded research fellows: an analysis of the health system impact fellowship program. Int J Health Policy Manag. 2023;12:7333. doi: 10.34172/ijhpm.2023.7333. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 7. Sheikh K, Abimbola S. Learning Health Systems: Pathways to Progress. Flagship Report of the Alliance for Health Policy and Systems Research. Geneva: World Health Organization; 2021.
- 8.Ghaffar A, Gupta A, Kampo A, Swaminathan S. The value and promise of embedded research. Health Res Policy Syst. 2021;19(Suppl 2):99. doi: 10.1186/s12961-021-00744-8. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 9.Somerville M, Cassidy C, Curran JA, Johnson C, Sinclair D, Elliott Rose A. Implementation strategies and outcome measures for advancing learning health systems: a mixed methods systematic review. Health Res Policy Syst. 2023;21(1):120. doi: 10.1186/s12961-023-01071-w. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 10.Lannon C, Schuler CL, Seid M, et al. A maturity grid assessment tool for learning networks. Learn Health Syst. 2021;5(2):e10232. doi: 10.1002/lrh2.10232. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 11.Hill JE, Stephani AM, Sapple P, Clegg AJ. The effectiveness of continuous quality improvement for developing professional practice and improving health care outcomes: a systematic review. Implement Sci. 2020;15(1):23. doi: 10.1186/s13012-020-0975-2. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 12.Shojania KG, Grimshaw JM. Evidence-based quality improvement: the state of the science. Health Aff (Millwood) 2005;24(1):138–150. doi: 10.1377/hlthaff.24.1.138. [DOI] [PubMed] [Google Scholar]
- 13.Yano EM, Than C, Brunner J, et al. Impact of evidence-based quality improvement on tailoring VA’s patient-centered medical home model to women veterans’ needs. J Gen Intern Med. 2024;39(8):1349–1359. doi: 10.1007/s11606-024-08647-4. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14.Atkins D, Kilbourne AM, Shulkin D. Moving from discovery to system-wide change: the role of research in a learning health care system: experience from three decades of health systems research in the Veterans Health Administration. Annu Rev Public Health. 2017;38:467–487. doi: 10.1146/annurev-publhealth-031816-044255. [DOI] [PubMed] [Google Scholar]
- 15.Rajit D, Reeder S, Johnson A, Enticott J, Teede H. Tools and frameworks for evaluating the implementation of learning health systems: a scoping review. Health Res Policy Syst. 2024;22(1):95. doi: 10.1186/s12961-024-01179-7. [DOI] [PMC free article] [PubMed] [Google Scholar]