Skip to main content
BMJ Global Health logoLink to BMJ Global Health
. 2019 Mar 7;4(2):e001257. doi: 10.1136/bmjgh-2018-001257

How is implementation research applied to advance health in low-income and middle-income countries?

Olakunle Alonge 1,, Daniela Cristina Rodriguez 1, Neal Brandes 2, Elvin Geng 3, Ludovic Reveiz 4, David H Peters 1
PMCID: PMC6441291  PMID: 30997169

Abstract

This paper examines the characteristics of implementation research (IR) efforts in low-income and middle-income countries (LMICs) by describing how key IR principles and concepts have been used in published health research in LMICs between 1998 and 2016, with focus on how to better apply these principles and concepts to support large-scale impact of health interventions in LMICs. There is a stark discrepancy between principles of IR and what has been published. Most IR studies have been conducted under conditions where the researchers have considerable influence over implementation and with extra resources, rather than in ‘real world’ conditions. IR researchers tend to focus on research questions that test a proof of concept, such as whether a new intervention is feasible or can improve implementation. They also tend to use traditional fixed research designs, yet the usual conditions for managing programmes demand continuous learning and change. More IR in LMICs should be conducted under usual management conditions, employ pragmatic research paradigm and address critical implementation issues such as scale-up and sustainability of evidence-informed interventions. This paper describes some positive examples that address these concerns and identifies how better reporting of IR studies in LMICs would include more complete descriptions of strategies, contexts, concepts, methods and outcomes of IR activities. This will help practitioners, policy-makers and other researchers to better learn how to implement large-scale change in their own settings.

Keywords: implementation, delivery, science, research, low and middle income countries, literature review


Summary box.

  • Implementation research (IR) in low-income and middle-income countries has mainly focused on evaluation of whether strategies for implementing evidence-informed health interventions can work; little of the IR addresses problems of scale-up and sustainability, which are key issues for health interventions.

  • Most of the publications on IR are not conducted under routine conditions for management and financing. If IR is to make an impact on policy and practice or inform the scale-up of programmes, more research needs to be conducted under the conditions in which interventions are expected to be implemented.

  • Most IR publications do not describe implementation characteristics completely; future research should more consistently provide complete descriptions of the implementation strategies, report on implementation variables and the context under which implementation occurs.

  • IR uses a full range of quantitative, qualitative and mixed methods approaches, but more rigorous and adaptive research designs are needed to address how to scale up and sustain interventions

Introduction

Implementation research (IR) efforts are not new, but attention is growing,1 particularly for the potential of IR to support evidence-informed interventions needed for achieving the Sustainable Development Goals.2 3 The growing attention to IR has revealed a gap in understanding of how widely IR is used and how its concepts and methods could be applied to achieve widespread health impact in low-income and middle-income countries (LMICs).4 In 2004, a call was made at the WHO Ministerial Summit on Health Research in Mexico for more IR around health systems strengthening strategies and evidence-informed interventions addressing major public health problems.5 The health research landscape in many LMICs has, however, changed little since this call.6

While there are ongoing debates concerning the definitions and boundaries of IR, there is a consensus on certain principles that should apply in IR.2 7–10 For instance, the need to conduct such research under real world conditions, and for it to respond to implementation problems, often in real time, is clear.2 8 9 The need to include inquiry about the context and for a team comprising diverse stakeholders to support implementation is also important.2 8–10 There are also concepts and methods that have been identified as appropriate for studying implementation problems.7–9 11–13 For instance, the use of implementation outcome variables that describe the results of intentional actions to deliver a programme,7 8 and a pragmatic research paradigm that allows flexibility in studying implementation strategies and contextual factors has been strongly recommended as being suitable for IR.8 11–15

There is also a consensus on key characteristics of IR, which we refer to as implementation descriptors, which should be reported in the peer-reviewed literature involving IR (box 1).16–18 These criteria are based on frameworks described under different published reporting guidelines for IR,16–18 but made broader to capture the plurality of concepts, methods and approaches used in global health research. These criteria have been described as important descriptors of IR studies because they provide insight into the how implementation works and indicate factors necessary for evaluating the external validity of evidence from IR studies.16

Box 1. Implementation descriptors: what you need to learn from others’ implementation experience.

Context and intervention

  • Clear description of the planned intervention, the evidence that justifies its choice and the main implementation strategies.

  • Contextual factors that affect implementation.

  • Clear description of who is implementing the intervention and strategies.

  • Any deviations from the planned intervention design and activities.

Implementation results

  • Measurement of implementation outcome variables, preferably showing changes over time

  • Clear discussion of the policy and practice implications of the study findings.

In this paper, we examine how IR is being operationalised to advance health in LMICs, drawing attention to how the IR principles and methods in health studies are applied in efforts to enhance the potential impact of interventions, which include policies, programme and individual behaviours. The goal of this paper is to describe how IR and evidence from such studies could be better applied to support large-scale impact of health interventions in LMICs.

Characteristics of IR health studies from LMICs

To characterise IR studies from LMICs, we conducted a systematic review of the relevant literature (box 2), examining the implementation descriptors identified in box 1. The number of IR publications over a 18 years review period (between 1998 and 2016) increased substantially. However, few papers reported on these factors, limiting the lessons that can be learnt from their efforts to enhance implementation in LMICs. For example, only 791 articles (8% of the pool of 10 292 relevant peer-reviewed IR literature) described the evidence-informed interventions and the set of implementation strategies that accompanied them. A majority of the 791 articles described who was doing the implementation (mainly civil society organisations [CSO]); however, less than 15% described whether there was any deviation from the initial planned intervention, which limits the understanding of how these interventions were adapted to fit their context (table 1). Adaptation is often a necessary step for achieving large-scale impact of interventions in varied settings.

Box 2. Selection of relevant peer-reviewed IR literature.

  • We conducted a systematic review of implementation research (IR) literature between January 1998 and December 2016 (see online supplementary file 1 for the full search string, PRISMA flow chart and methodology).

  • We identified relevant peer-reviewed records from the systematic review (see online supplementary file 2 for the bibliography of the relevant IR publications identified). Relevant peer-reviewed IR literature was broadly defined as research or evaluation articles that describe the implementation of an intervention to improve health8 and are set in low-income and middle-income countries.

  • We conducted additional in-depth review on relevant peer-reviewed records that satisfied three or more of the implementation descriptors described in box 1 (see online supplementary file 3 for links to the relevant IR publications selected for in-depth review).

Table 1.

Prevalence of key implementation descriptors reported in IR studies—context and intervention

Context and intervention N=791
n (%)
Contextual factors that influence IR described 415 (52.4)
Description of who is implementing the key activities 702 (88.7)
Deviation from initial planned intervention described 101 (12.8)
Beneficiaries described 706 (89.2)
Target audience for study findings described 366 (46.3)
Implementation period identified 368 (46.4)
Mean (SE)
Mean duration of implementation period (in months) 24.4 (1.3)

Source: Authors’ systematic review of the literature.

IR, implementation research.

Supplementary data

bmjgh-2018-001257supp001.pdf (87.8KB, pdf)

Supplementary data

bmjgh-2018-001257supp002.pdf (555KB, pdf)

Supplementary data

bmjgh-2018-001257supp003.pdf (48.4KB, pdf)

Only a small set of articles report more completely on the implementation descriptors (28 studies from our review reported on three or more descriptors and are shown in table 2 and online supplementary file 3).19–46 All but 1 of the 28 articles reported contextual factors,45 but 15 did so partially,19 20 28–32 34 35 38–40 43 46 usually focusing on external factors only (mainly, relevant sociopolitical factors that are external to the project settings and implementing organisations, but with implications for the set of implementation strategies), excluding discussion on the internal factors (social factors within the project settings and implementing organisations, including the organisation’s structure and culture). Fourteen articles measured changes in implementation outcome variables,19 20 22–28 30 31 37 38 46 including acceptability (perception among stakeholders that an intervention is agreeable), fidelity (extent to which an intervention was implemented as described), uptake (intention, decision or action to use an intervention) and coverage or reach (degree to which eligible population for an intervention received the intervention).7 8 Only five articles used an explicit or published implementation research model or theory such as the Consolidated Framework for Advancing Implementation Research (CFIR) and the theory of organisational readiness for change.15 20 28 36 37 42 Twenty-five articles reported who was conducting the implementation,19–31 33–41 44–46 but most only focused on the front-line implementers (eg, community health workers) without describing the overarching institutions supporting activities. Only nine articles reported on any adaptation or deviations from planned intervention,19–22 30 31 33–35 and it is unclear if this reflects a reporting issue or a lack of deviations actually taking place. For the reporting criteria emerging from this series,2 47 48 embeddedness (of the IR study in existing programmes and the larger health system) and incentives (of involved stakeholders to conduct the IR study) were reported very infrequently, seven and four articles, respectively. However, more than half of these articles (15) discussed policy and/or practice implications in some way.19–28 32 33 38 42 43

Table 2.

In-depth review of IR articles with more complete reporting of implementation descriptors

Article (Lead author and year) Research/evaluation article Implementation of intervention/strategy Context Changes in IR variables measured Who implements Deviations from initial protocol Discussion of policy/practice implications
Included all implementation descriptors
 McDougal 201219 Y Y Y (partially) Y Y Y Y
 Rajaraman 201220 Y Y Y (partially) Y Y Y Y
 Sherr 201321 Protocol Y Y N/A Y Y Y
 Belaid 201522 Y, process evaluation Y Y Y Y Y Y
Included all implementation descriptors, except deviation from initial protocol
 Liao 201123 Y Y Y Y Y N Y
 Georgeu 201224 Y Y Y Y Y N Y
 Nahar 201225 Y Y Y Y Y (partially) N Y
 Teklehaimanot 201326 Y, case study Y Y Y Y N Y
 Shei 201327 Y Y Y Y Y N Y
 Ansbro 201528 Y Y Y (partially) Y Y N Y
Included all implementation descriptors, except discussion of policy/practice implication
 Abramsky 201229 Protocol Y Y (partially) N/A Y N/A N
 Ngana 201230 Y Y Y (partially) Y Y Y N
 DeCelles 201631 Y, process evaluation Y (partially) Y (partially) Y Y Y N
Included all implementation descriptors, except who implements
 Bernabe-Ortiz 201432 Protocol Y Y (partially) N/A N N/A Y
Included all implementation descriptors, except changes in IR variables measured
 Cooper 201533 Y Y Y N Y Y Y
Included all implementation descriptors, except changes in IR variables measured and discussion of policy/practice implication
 Miri 199834 Y, process evaluation Y Y (partially) N Y Y N
 Christensen 201535 Y Y Y (partially) N Y Y N
Included all implementation descriptors, except deviation from initial protocol and discussion of policy/practice implication
 Findley 201336 Y Y Y Y Y (partially) N N
 Shelley 201537 Protocol Y Y Y Y N N
Included all implementation descriptors, except changes in IR variables measured and deviation from initial protocol
 Hirschhorn 201538 Y Y Y (partially) N Y N Y
Included only three implementation descriptors
 Hopkins 199839 No, programme report Y Y (partially) N Y N/A N
 Brune 200940 Y Y Y (partially) N Y N N
 Carlo 201041 Y Y Y N Y N N
 Shrestha 201342 Y Y Y N N N Y
 Kim 201543 Y Y Y (partially) N N N Y
 Church 201544 Y Y Y N Y N N
 Mahvu 201545 Y Y N Y Y N N
 Babiarz 201646 Y Y Y (partially) N Y N N

Application of key IRDS principles and concepts to IR health studies from LMICs

Key IR principles and concepts can be summarised under six major themes (box 3).1 2 7–9 11–13 These themes reflect the convergence in literature on the distinguishing features of IR activities that are expected to produce impact at scale and are relevant when examining IR studies that may yield evidence of large-scale impact.

Box 3. Key principles and concepts of IR activities.

  • Implementation research (IR) activities should focus on questions around implementing evidence-informed interventions.

  • IR activities should be conducted under real world conditions, with the types of resources, incentives and operational support they would have under routine situations (where provision of additional resources is an implementation strategy, and then the study should indicate how such provisions will be sustained and integrated into the system).

  • IR studies should provide evidence for the ‘how’ and ‘why’ evidence-informed interventions led to health impact through use of implementation outcome variables and a pragmatic research paradigm.

  • Conducting IR studies should be fit to purpose and balance the need to address the immediate implementation problems and support broader and longer term learning.

  • IR activities require multistakeholder collaboration and partnerships.

  • Key characteristics of IR activities should be reported to facilitate learning and action.

Conducting research that focuses on questions about implementing evidence-informed interventions

Questions about implementation are often concerned with how to produce expected results in a particular setting from an intervention that has been previously shown to work elsewhere.1 2 7–9 Such questions may address the fit of an intervention to new settings, how to introduce, scale-up or sustain the intervention among broader populations or address real-time operational issues that may occur in the process of implementation.7 8 IR studies that focused on scale-up and sustainability are particularly essential for understanding the large-scale impact of health interventions. However, less than 5% of 791 Implementation Research (IR) studies reviewed for this paper addressed objectives concerning issues such as scaling up (n=32) or sustainability (n=25). The BetterBirth project,38 which focused on implementing a Safe Childbirth Checklist (SCC) to promote safe childbirth practices and reduce maternal and neonatal mortality in Uttar Pradesh, India, is an example of a research project that did address questions around implementing an evidence-informed intervention. Using an iterative learning process, researchers involved in the BetterBirth project adapted the SCC intervention to operational issues that occur in the process of introducing the SCC to health workers in different settings from where it was originally tested. Over two learning phases, they added a peer-to-peer coaching strategy to support the introduction of the checklist, in part to respond to leadership challenges and health worker’s lack of motivation that were identified during an initial implementation phase. They used physician coaches to motivate health facility leaders and provide critical leadership for the implementation and used nurse coaches to motivate behavioural changes in birth attendants to use the checklist. The study was able to demonstrate large increases in the correct use of the SCC intervention over time and provide evidence for a set of implementation strategies to support the scale-up of the SCC.

Whereas the focus of most IR studies are around evaluation, key implementation questions in most settings are concerned with how to scale up or sustain an intervention within a practice area or population.49 Issues around scaleup of interventions and their sustainability occur within complex health systems that are constantly evolving.8 9 The lack of studies addressing issues on scale-up and sustainability may perhaps point to the practical difficulties in how to conduct research that addresses complexity. Indeed, the capacity of public health researchers to address questions examining complex and adaptive processes has been limited until recently.50 There are, however, examples of implementation models for describing and studying scale-up and innovations for sustaining implementation activities and outcomes from other fields that are being extended to IR in public health.51–53 For example, Aarons and colleagues describe one model for implementing public sector services and included four different implementation phases (exploration, adoption/preparation, implementation and sustainment), highlighting specific factors that are important at each phase.51 These included factors affecting the outer context (service environment, interorganisational environment, consumer support/advocacy), the inner context (intraorganisational characteristics, individual adopter characteristics), innovation characteristics and the fit with existing systems.51 There are additional methods related to systems thinking that are also useful for examining questions of implementation within complex systems.50 54

Conducting research on implementation of evidence-informed interventions under real world conditions

Whereas what constitute ‘real-world’ conditions for a set of implementation activities may be debatable, criteria such as whether implementation was led by usual implementing agencies and without additional funding (apart from usual budget) or management support for implementation during the study1 2 8 may be useful to gauge the extent to which conditions for implementing an intervention depart from normal routine. Where provision of additional resources is used as an implementation strategy, then this should be clear, and the study should indicate how such provisions will be sustained and integrated into the system. This principle—to conduct research on implementation under real world conditions—is perhaps one the most distinguishing characteristics of IR studies.2 8 10 Unfortunately, most IR studies are conducted under more controlled settings, which limits the extent to which learning from those studies could be applied to commonly found conditions.2 8–10 Our review of the literature also suggests that this critical approach may be lacking in published IR studies in LMICs. In examining management support, funding and who is doing the implementation, only 26% of IR studies reviewed did not have additional funding or provided plans for sustaining any additional funding provided as an implementation strategy, and implementation of interventions of these studies was identified as being led by the usual implementing agency in about 44% of the studies (table 3).

Table 3.

Application of key IR principles to IRDS health studies for large-scale impact in LMICs

Management and financing arrangements: real world conditions N=791
n (%)
Implementation led by usual implementing agencies
 Yes 344 (43.5)
 No 249 (31.5)
 Not described 198 (25.0)
No additional funding (apart from usual budget) or plans for sustaining any additional funding provided
 Yes 206 (26.0)
 No (additional funding provided) 361 (45.6)
 Not described 225 (28.4)
No additional management support or plans for sustaining any additional management provided
 Yes 215 (27.2)
 No (additional management support provided) 365 (46.1)
 Not described 211 (26.7)
Measurement of implementation
 Included implementation outcome variables (eg, fidelity, acceptability, coverage) 426 (53.9)
 Included effectiveness outcome variables (eg, morbidity and mortality) 395 (49.9)
 Included both effectiveness and implementation outcome variables 200 (24.9)
 Measured change in implementation outcome variables 79 (10.0)
Multistakeholder collaboration in IRDS studies
 Academics, implementers/CSO, Govt./policy-makers and donors 17 (2.2)
 Academics, implementers/CSO and Govt./policy-makers 130 (16.4)
 Academics and Implementers/CSO 411 (51.9)
 Academics and Govt./Policy-makers 176 (22.2)
 Academics and donors 60 (7.6)

Source: Authors’ literature review.

CSO, civil society organisations; IR, implementation research; LMICs, low-income and middle-income countries.

Some authors have suggested that implementation efforts and relevant IR studies should be led by dedicated implementation teams (including key personnel enacting the implementation strategy) under routine conditions.10 55 56 Such teams have been espoused as critical for maintaining the implementation activities under real-world practices and also aiding the scale-up and dissemination of such practices to other settings.10 55 56 The formation of teams is critical also for addressing systemic issues and contributing to sustainability. One example is the Population Health and Implementation Training (PHIT) Partnership in Mozambique, which has built a network of institutions to strengthen primary healthcare and is integrated into the Ministry of Health (MOH) at the subnational level.21 Partners include provincial health government, academic institutions (local and external), NGOs, and the research arm of the MOH.21 Each PHIT partner has a specific technical assistance and/or research role to support provincial health authorities design and implement activities. Another example is the proof of concept study in Liberia that assessed integration of family planning and immunisation. In this case, the MOH worked with an implementing partner NGO and USAID to design a project that would generate lessons to inform implementation at-scale. The three partners worked together to engage with a broader stakeholder group, identify intervention study areas, pilot intervention and share results (see Cooper in table 1).33

Research under real-world conditions may also incorporate readiness assessments or involve organisational and system level changes for optimising the benefit and impact of evidence-informed interventions.56 For example, the BetterBirth project recognised leadership gap at health facility levels in Uttar Pradesh, India, and the implication of this for the successful rollout of the SCC intervention.38 The project engaged with district health level leadership and an application-based technology to motivate organisational and system level changes that provided real-time data feedback to district teams on observations and the availability of essential birth supplies at health facilities.38 Using the data, the district health leadership were able to strengthen the real-time availability of supplies necessary for essential birth practices.41 The PHIT Partnership in Mozambique has integrated targeted operations research studies to address systems bottlenecks on a continuous basis; these are implemented in partnership with local level managers.21 Examples of their studies include assessments of the effect of introducing shift work to extend outpatient care clinic hours and quality improvement activities for reducing loss to follow-up among paediatric patients with HIV.21 These strategies involved organisational changes to improve work processes to support implementation of evidence-informed interventions.

IR conducted under real-world conditions also allows for identification and potentially the modification of incentives faced by different actors. A project that expanded community mobilisation for maternal and newborn health through women’s groups in rural Bangladesh faced challenges of frequent staff turnover of group facilitators due to unexpected constraints on facilitators’ movement or offers of better employment, which were compounded by recruitment requirements for minimum levels of literacy. Project implementers were aware of how important familiarity with local communities and customs was to effective facilitation, so they adapted the recruitment criteria and processes to be more flexible and responsive to their context.25 Another example is the STRETCH trial in South Africa, which was an intervention aimed at supporting nurse-initiated and managed antiretroviral therapy. As part of the trial, clinical support to improve quality of care and self-confidence was offered to nurses administering antiretroviral therapy through several mechanisms, including training, telephone and in-person assistance and management and logistics support. Unfortunately, clinical support was insufficient and nurses felt that STRETCH trainers were unable to fill the gap.24

IR studies should provide evidence for ‘how’ and ‘why’ interventions led to health impact through use of implementation outcome variables and a pragmatic research paradigm

The primary objective of IR studies is often about answering how and why interventions produce their desired effects on health in a given setting.8 9 There is almost always the assumption that some evidence exists a priori on what interventions have worked for a particular health problem in some (often controlled) setting; this is why they are selected as ‘evidence-informed’ interventions. What is often missing, however, is an understanding of the implementation pathway for these interventions; how they are carried out in real-life contexts to address existing health problems and how they may be used to achieve sustainable impact at scale.7 One way to describe the implementation pathway is to link changes in implementation outcome variables and contexts to specific implementation strategies.7 Implementation outcome variables include measures of implementation fidelity, acceptability of interventions by potential beneficiaries, uptake of the interventions, reach or coverage and cost.7 8 The changes in implementation outcome variables and contextual factors could in turn be linked to measures of programme effectiveness and impact.7 For example, an evaluation of the Brazil’s Bolsa Familia programme—a conditional cash transfer programme to reduce poverty and improve health outcomes among poor families—showed that an increase in the reach of the programme (an implementation outcome variable) was associated with increased utilisation of preventive child health services, and this in turn was associated with a decline in postneonatal deaths and infant mortality over a 5 years period.27 The study was able to show the implementation pathway leading to impact for the evidence-informed intervention by explicitly linking changes in implementation outcome variables to indicators of programme impact.

Our review, however, found that the majority of IR studies from LMICs neither described changes in implementation outcome variables nor described how these changes are related to implementation strategies and desired outcomes in a given setting (table 3). While such descriptions may not be necessary for all types of IR studies, they are essential for IR studies that yield evidence of large-scale impact. Without such links, IRDS studies perhaps become undistinguishable from more traditional efficacy and effectiveness studies57 and the utility of IRDS for fulfilling the dual role of being useful for generating knowledge while addressing health problem in real-time may not be realised.57

Use of a pragmatic research paradigm and relevant study designs such as implementation-effectiveness hybrid designs (which combines measurement of outcomes of clinical effectiveness and implementation), mixed methods designs (which strategically combines qualitative and quantitative methods) and systems science methods (which characterises nonlinear processes within complex phenomena) allow IR researchers to model the iterative nature of implementation efforts without compromising the rigour of such studies.8 11 12 Such research designs are also useful for modelling changes in the intervention as well as implementation outcome variables and programme effectiveness in the same study. Our review suggested that mixed methods were indeed commonly used in IR health studies in LMICs; more than one-third of IR papers reviewed used some form of mixed methods. However, when we examined these mixed method IR studies closely, very few described the explicit mixed method design strategies used,58 which have implication for the rigour and validity of findings from such studies. Furthermore, only 14 of the relevant 791 IR papers we reviewed used other designs informed by a pragmatic research paradigm, such as implementation-effectiveness hybrid designs or system science methods. One example involved studying the adaptations to the implementation of the Seguro Popular de Salud programme in Mexico (a health financing reform package to accelerate universal health coverage and financial risk protection) using a system dynamics approach to inform a better understanding of real-life implementation variations in the programme in response to different incentives.59 While the impact of the reform on reducing catastrophic health expenditure was previously established,60 the application of a system dynamics approach to the evaluation data provided useful lessons on how the reform was tailored to various contextual factors and the features of organisational and system level changes that contributed to the impact observed under various settings.61

Conducting IR studies should be fit to purpose and balance the need to address the immediate implementation problems and support broader and longer term learning

Whereas IR studies conducted under real-world conditions may be more responsive to immediate implementation problems, lessons learnt from such studies may not always be readily generalisable to support broader and longer term learning, resulting in a common trade-off of IR studies.61 Suggested approaches for balancing this trade-off include use of specific research designs (as described above) that allow IR studies to adapt and study particular real-world conditions while still ensuring that research parameters are studied and analysed objectively.8 11 12 Others have suggested that well-designed studies conducted within multisetting and multicountry contexts help to balance this trade-off and improve the generalisability of IR study findings.56 61 For example, a feasibility study to evaluate the potential for establishing a platform for bringing together relevant health data from the public and private sectors to promote data use at the district health level was initially implemented in five different districts of India, Nigeria and Ethiopia.62 This study captured potential facilitators and barriers to the introduction of the platform and provided generalisable learning on the lack of standardised processes for data-based decision-making at district level across all three countries.62 Less than 10% of studies we reviewed were conducted within multiple settings and multicountry contexts. This may in fact be reflective of the financial and logistical constraints of undertaking such complex research efforts.4 While it may be difficult, and perhaps inappropriate, to answer relevant IR questions using multiple settings and countries, the need to optimise the trade-off between internal and external validity of IR studies is another reason to use of adaptive research designs.8 11 12 56

IR requires multistakeholder collaboration and partnerships

Collaboration and partnerships among multiple stakeholders (such as academics, implementers, policy-makers and donors) across various influence domains (research, programme, policy and funding) is important for any IR enterprise to achieve large-scale impact.1 2 4 5 8 10

Aligning research, programme, policy and funding cycles has been suggested as critical to address health problems at scale using IR, especially in LMICs.2 Such alignments are often more easily accomplished when there is collaboration and partnership to facilitate decisions and cross-learning among actors involved in these various cycles.2 One example is a community mobilisation project of women’s groups in Bangladesh, which made considerable efforts to engage and build partnerships with local government and NGOs, and also with community stakeholders involved in the women’s groups, such as traditional birth attendants and community health workers.25 These efforts were made in recognition that trust was needed to ensure an effective intervention, but also to eventually address service gaps and influence policymaking.25 Similarly, an intervention to enhance social capital in Nicaragua worked with governmental, non-governmental and community stakeholders at the project’s outset to identify, based on their collective experience, the communities which would be targeted in the programme, but further interactions between the stakeholders was unclear.40 In the integrated family planning and immunisation project in Liberia, project team recommendations were supplemented by additional recommendations from the MOH, suggesting strong engagement from policy-makers.33 The degree of alignments may also reflect the extent of embeddedness of such IR activities within context-specific programme, practice or policy and could infer the policy and practice relevance of findings from such IR activities.34 Some teams develop relationships, either by design or that build organically over time, that put researchers in a position to support implementers’ roles and activities. For example, Ngana and colleagues point out how by implementing a participatory action research model to address health reporting challenges, health staff developed problem-solving skills that allowed them to better respond and adapt to systemic challenges.30 Under the STRETCH trial in South Africa, the trial coordinator contributed to improving communications between clinical staff to strengthen relationships between physicians, nurses and other staff, which were critical to successful task-shifting.24 However, our review showed that efforts to involve key stakeholders across various domains of influence are still infrequent, with only 16 out of the 28 articles included for in-depth review reporting on the policy or practice implications of their findings (table 3).

Reporting of key characteristics of IR studies

Our review uncovered the lack of consistent reporting of implementation characteristics, including the adequate descriptions of the context, the intervention itself, including deviations from planned interventions and changes in implementation variables. Several authors have noted that implementation of evidence-informed interventions in health has been hampered by the lack of adequate reporting of IR studies.16–18 As a result, implementers have difficulty in translating the IR publication findings into practice, and the findings of such research cannot be easily used outside of their original settings.16 17 Adequate reporting of implementation strategies has been particularly emphasised in the literature because of its centrality to implementation science.17 Indeed, some authors have referred to implementation strategy as the ‘intervention’ in IR studies,8 11 and several guidelines have been proposed on how to report implementation strategy and IR studies.16–18 All the guidelines recommend that an implementation strategy should be described in details and the implementation outcome variable that describes the proximal effect of such strategy should also be described.16–18

Other key aspects that have been proposed for reporting include description of the background (including implementation gaps and supporting evidence for selecting an intervention), the settings and contexts (including relevant factors within the internal and external environment of the implementing organisations and beneficiaries) and how these factors changed over time to facilitate or hinder implementation activities.16 Such information may be required to inform practitioners on how to adapt the strategy or implementation activities to their own local contexts. It is possible that issues, such as journal word limits and writing requirements, can restrict what is included in publications.16 Reporting of IR studies could be improved if deliberate attention is paid to highlighting key aspects of IR given journal constraints. Our review suggested that articles published in certain journals (eg, Global Health: Science and Practice, PLoS One, Implementation Science) had higher probability of reporting more completely on the implementation descriptors assessed under this study. This goes to suggest that journal constraints may be a much bigger factor in determining extent to which key IR characteristics are reported. Our review further suggested that involvement of multiple stakeholders in an IR study was associated with more complete reporting on the implementation descriptors assessed. For example, articles involving academics, CSO, government and donors had the highest percentage of reporting on implementation outcome variables (69%), much higher than observed with single stakeholder or other multistakeholder groups for example, academics/CSO/government (51%), academic/CSO (50%), academics/government (57%), academic/donor (67%).

Conclusion

IR has promise in supporting large-scale impact of health interventions in LMICs, but to maximise its potential, more IR work needs to take place under usual management conditions, employ a pragmatic research paradigm and address critical implementation issues such as scale-up and sustainability as part of complex systems. The increasing trend of publications in IR from LMICs is indicative of the promise and usefulness of IR approaches in addressing critical health problems in the developing world. Better reporting of IR studies in LMICs are needed that more clearly describe the strategies, contexts, concepts, methods and outcomes of IR activities and more peer-reviewed journals should encourage such reporting. More high quality peer-reviewed IR publications are needed to reflect the full range of problems, contexts, methods and results, which will serve to enhance learning and implementation of health interventions more widely.

Footnotes

Handling editor: Seye Abimbola

Contributors: OA conceptualised the paper, led the review, was the principal author and managed all revisions of drafts of the paper. DCR contributed to the review and writing drafts of the paper. DHP provided guidance to the review and writing drafts of the paper. NB, EG, LR and DHP revised the paper for intellectual content.

Funding: The study has been funded by USAID, Alliance for Health Policy and Systems Research, and the World Bank.

Competing interests: None declared.

Patient consent for publication: Not required.

Provenance and peer review: Not commissioned; externally peer reviewed.

Data sharing statement: All data relevant to the study are included in the article or uploaded as supplementary information.

References

  • 1. Alliance for Health Policy and Systems Research within the World Health Organization, the United States Agency for International Development (USAID), and the World Bank Group Health systems global, 2014. Available: http://www.healthsystemsglobal.org/the-irds-statement/ [Accessed 14 Jan 2019].
  • 2. Theobald S, Brandes N, Gyapong M, et al. . Implementation research: new imperatives and opportunities in global health. The Lancet 2018;392:2214–28. 10.1016/S0140-6736(18)32205-0 [DOI] [PubMed] [Google Scholar]
  • 3. United Nations Transforming our world: the 2030 agenda for sustainable development, 2015. Available from: https://sustainabledevelopment.un.org/post2015/transformingourworld [Accessed 14 Jan 2019].
  • 4. Sanders D, Haines A. Implementation research is needed to achieve international health goals. PLoS Med 2006;3:e186 10.1371/journal.pmed.0030186 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5. Ministerial Summit on Health Research The Mexico statement on health research. Mexico City, 2004. Available from: http://www.wpro.who.int/health_research/policy_documents/mexico_statement_20nov04.pdf [Accessed 18 Aug 2018].
  • 6. Bennett S, Adam T, Zarowsky C, et al. . From Mexico to Mali: progress in health policy and systems research. The Lancet 2008;372:1571–8. 10.1016/S0140-6736(08)61658-X [DOI] [PubMed] [Google Scholar]
  • 7. Proctor E, Silmere H, Raghavan R, et al. . Outcomes for implementation research: conceptual distinctions, measurement challenges, and research agenda. Adm Policy Ment Health 2011;38:65–76. 10.1007/s10488-010-0319-7 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8. Peters DH, Adam T, Alonge O, et al. . Implementation research: what it is and how to do it. BMJ 2013;347:f6753 10.1136/bmj.f6753 [DOI] [PubMed] [Google Scholar]
  • 9. Lobb R, Colditz GA. Implementation science and its application to population health. Annu Rev Public Health 2013;34:235–51. 10.1146/annurev-publhealth-031912-114444 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10. Fixsen DL, Naoom SF, Blase KA, et al. . Implementation research: a synthesis of the literature. [Google Scholar]
  • 11. Curran GM, Bauer M, Mittman B, et al. . Effectiveness-implementation hybrid designs: combining elements of clinical effectiveness and implementation research to enhance public health impact. Med Care 2012;50 10.1097/MLR.0b013e3182408812 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12. Brown CH, Curran G, Palinkas LA, et al. . An overview of research and evaluation designs for dissemination and implementation. Annu Rev Public Health 2017;38:1–22. 10.1146/annurev-publhealth-031816-044215 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13. Palinkas LA, Aarons GA, Horwitz S, et al. . Mixed method designs in implementation research. Adm Policy Ment Health 2011;38:44–53. 10.1007/s10488-010-0314-z [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14. Alonge O, Frattaroli S, Davey-Rothwell M, et al. . A Transdisciplinary approach for teaching implementation research and practice in public health. Pedagogy in Health Promotion 2016;2:127–36. 10.1177/2373379915618215 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15. Damschroder LJ, Aron DC, Keith RE, et al. . Fostering implementation of health services research findings into practice: a consolidated framework for advancing implementation science. Implementation Science 2009;4 10.1186/1748-5908-4-50 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16. Pinnock H, Epiphaniou E, Sheikh A, et al. . Developing standards for reporting implementation studies of complex interventions (STARI): a systematic review and e-Delphi. Implement Sci 2015;10 10.1186/s13012-015-0235-z [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17. Proctor EK, Powell BJ, McMillen JC. Implementation strategies: recommendations for specifying and reporting. Implementation Science 2013;8 10.1186/1748-5908-8-139 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18. Hales S, Lesher-Trevino A, Ford N, et al. . Reporting guidelines for implementation and operational research. Bull World Health Organ 2016;94:58–64. 10.2471/BLT.15.167585 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19. McDougal L, Moteetee MM, Mohai F, et al. . Lesotho's minimum PMTCT package: lessons learned for combating vertical HIV transmission using co-packaged medicines. J Int AIDS Soc 2012;15 10.7448/IAS.15.2.17326 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 20. Rajaraman D, Travasso S, Chatterjee A, et al. . The acceptability, feasibility and impact of a lay health counsellor delivered health promoting schools programme in India: a case study evaluation. BMC Health Services Research 2012;12 10.1186/1472-6963-12-127 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21. Sherr K, Cuembelo F, Michel C, et al. . Strengthening integrated primary health care in Sofala, Mozambique. BMC Health Services Research 2013;13 10.1186/1472-6963-13-S2-S4 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22. Belaid L, Ridde V. Contextual factors as a key to understanding the heterogeneity of effects of a maternal health policy in Burkina Faso? Health Policy and Planning 2015;30:309–21. 10.1093/heapol/czu012 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 23. Liao S, Weeks MR, Wang Y, et al. . Female condom use in the rural sex industry in China: analysis of users and non-users at post-intervention surveys. AIDS Care 2011;23:66–74. 10.1080/09540121.2011.555742 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 24. Georgeu D, Colvin CJ, Lewin S, et al. . Implementing nurse-initiated and managed antiretroviral treatment (NIMART) in South Africa: a qualitative process evaluation of the stretch trial. Implementation Science 2012;7 10.1186/1748-5908-7-66 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25. Nahar T, Azad K, Aumon BH, et al. . Scaling up community mobilisation through women's groups for maternal and neonatal health: experiences from rural Bangladesh. BMC Pregnancy and Childbirth 2012;12 10.1186/1471-2393-12-5 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 26. Teklehaimanot HD, Teklehaimanot A. Human resource development for a community-based health extension program: a case study from Ethiopia. Human Resources for Health 2013;11 10.1186/1478-4491-11-39 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27. Shei A. Brazil’s conditional cash transfer program associated with declines in infant mortality rates. Health Affairs 2013;32:1274–81. 10.1377/hlthaff.2012.0827 [DOI] [PubMed] [Google Scholar]
  • 28. ÉM A, Gill MM, Reynolds J, et al. . Introduction of syphilis point-of-care tests, from pilot study to national programme implementation in Zambia: A qualitative study of healthcare workers’ perspectives on testing, training and quality assurance. PloS one 2015;10:e0127728. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 29. Abramsky T, Devries K, Kiss L, et al. . A community mobilisation intervention to prevent violence against women and reduce HIV/AIDS risk in Kampala, Uganda (the Sasa! study): study protocol for a cluster randomised controlled trial. Trials 2012;13 10.1186/1745-6215-13-96 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 30. Rambu Ngana F, Myers BA, Belton S. Health reporting system in two subdistricts in eastern Indonesia: highlighting the role of village midwives. Midwifery 2012;28:809–15. 10.1016/j.midw.2011.09.005 [DOI] [PubMed] [Google Scholar]
  • 31. DeCelles J, Hershow RB, Kaufman ZA, et al. . Process evaluation of a Sport-Based voluntary medical male circumcision Demand-Creation intervention in Bulawayo, Zimbabwe. J Acquir Immune Defic Syndr 2016;72(Suppl 4):S304–S308. 10.1097/QAI.0000000000001172 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 32. Bernabe-Ortiz A, Diez-Canseco F, Gilman RH, et al. . Launching a salt substitute to reduce blood pressure at the population level: a cluster randomized stepped wedge trial in Peru. Trials 2014;15 10.1186/1745-6215-15-93 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 33. Cooper CM, Fields R, Mazzeo CI, et al. . Successful proof of concept of family planning and immunization integration in Liberia. Global Health: Science and Practice 2015;3:71–84. 10.9745/GHSP-D-14-00156 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 34. Miri ES. Problems and perspectives of managing an onchocerciasis control programme: a case study from plateau state, Nigeria. Ann Trop Med Parasitol 1998;92 Suppl 1:S121–S128. 10.1080/00034983.1998.11813375 [DOI] [PubMed] [Google Scholar]
  • 35. Christensen G, Dentz HN, Pickering AJ, et al. . Pilot cluster randomized controlled trials to evaluate adoption of water, sanitation, and hygiene interventions and their combination in rural Western Kenya. Am J Trop Med Hyg 2015;92:437–47. 10.4269/ajtmh.14-0138 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 36. Findley SE, Uwemedimo OT, Doctor HV, et al. . Early results of an integrated maternal, newborn, and child health program, Northern Nigeria, 2009 to 2011. BMC Public Health 2013;13 10.1186/1471-2458-13-1034 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 37. Shelley D, VanDevanter N, Cleland CC, et al. . Implementing tobacco use treatment guidelines in community health centers in Vietnam. Implement Sci 2015;10 10.1186/s13012-015-0328-8 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 38. Hirschhorn LR, Semrau K, Kodkany B, et al. . Learning before leaping: integration of an adaptive study design process prior to initiation of BetterBirth, a large-scale randomized controlled trial in Uttar Pradesh, India. Implement Sci 2015;10 10.1186/s13012-015-0309-y [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 39. Hopkins AD. Mectizan delivery systems and cost recovery in the Central African Republic. Ann Trop Med Parasitol 1998;92:S97–S100. 10.1080/00034983.1998.11813372 [DOI] [PubMed] [Google Scholar]
  • 40. Brune NE, Bossert T. Building social capital in post-conflict communities: evidence from Nicaragua. Soc Sci Med 2009;68:885–93. 10.1016/j.socscimed.2008.12.024 [DOI] [PubMed] [Google Scholar]
  • 41. Carlo WA, McClure EM, Chomba E, et al. . Newborn care training of midwives and neonatal and perinatal mortality rates in a developing country. PEDIATRICS 2010;126:e1064–71. 10.1542/peds.2009-3464 [DOI] [PubMed] [Google Scholar]
  • 42. Shrestha S, Petrini M, Turale S. Newborn care in Nepal: the effects of an educational intervention on nurses' knowledge and practice. International Nursing Review 2013;60:205–11. 10.1111/inr.12017 [DOI] [PubMed] [Google Scholar]
  • 43. Kim MH, Ahmed S, Hosseinipour MC, et al. . Implementation and operational research: the impact of option B+ on the antenatal PMTCT cascade in Lilongwe, Malawi. J Acquir Immune Defic Syndr 2015;68 10.1097/QAI.0000000000000517 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 44. Church K, Wringe A, Lewin S, et al. . Exploring the feasibility of service integration in a low-income setting: a mixed methods investigation into different models of reproductive health and HIV care in Swaziland. Plos One 2015;10:e0126144 10.1371/journal.pone.0126144 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 45. Mavhu W, Larke N, Hatzold K, et al. . Implementation and operational research: a randomized noninferiority trial of AccuCirc device versus Mogen clamp for early infant male circumcision in Zimbabwe. J Acquir Immune Defic Syndr 2015;69:e156 10.1097/QAI.0000000000000694 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 46. Babiarz KS, Mahadevan SV, Divi N, et al. . Ambulance service associated with reduced probabilities of neonatal and infant mortality in two Indian states. Health Affairs 2016;35:1774–82. 10.1377/hlthaff.2016.0564 [DOI] [PubMed] [Google Scholar]
  • 47. Nhan T, Reveiz LH, Bercerra F, et al. . Embedded Implementation Research & Delivery Science: improving programme performance and facilitating scale up of effective interventions. BMJ Global Health. Forthcoming 2019. [Google Scholar]
  • 48. Reichenbach L, Jones K, Madhavan S. Delivering Improved Health Outcomes: Recognizing Stakeholder’s Roles and Leveraging Incentives and Disincentives in Implementation Research and Delivery Science. BMJ Global Health 2019. [Google Scholar]
  • 49. Greenhalgh T, Robert G, Macfarlane F, et al. . Diffusion of innovations in service organizations: systematic review and recommendations. Milbank Q 2004;82:581–629. 10.1111/j.0887-378X.2004.00325.x [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 50. Luke DA, Stamatakis KA. Systems science methods in public health: dynamics, networks, and agents. Annual Review of Public Health 2012;33:357–76. 10.1146/annurev-publhealth-031210-101222 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 51. Aarons GA, Hurlburt M, Horwitz SM. Advancing a conceptual model of evidence-based practice implementation in public service sectors. Adm Policy Ment Health 2011;38:4–23. 10.1007/s10488-010-0327-7 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 52. Fixsen A. Defining scale-up across disciplines: an annotated bibliography, 2009. Available: https://nirn.fpg.unc.edu/resources/defining-scaling-across-disciplines-annotated-bibliography [Accessed cited 2019 January 14].
  • 53. Fixsen DL, Blase KA, Duda M, et al. . Sustainability of evidence-based programs in education. Journal of Evidence-Based Practices for Schools 2010;11:30–46. [Google Scholar]
  • 54. Hammond R. Considerations and best practices in Agent-Based modeling to inform policy. in US Institute of medicine, editors. Assessing the use of Agent-Based models for tobacco regulation. Washington DC: National Academies Press, 2015. [PubMed] [Google Scholar]
  • 55. Higgins M, Weiner J, Young L. Implementation teams: a new lever for organizational change. J Organ Behav 2012;33:366–88. 10.1002/job.1773 [DOI] [Google Scholar]
  • 56. Fixsen DL, Blase KA, Naoom SF, et al. . Core implementation components. Res Soc Work Pract 2009;19:531–40. 10.1177/1049731509335549 [DOI] [Google Scholar]
  • 57. US Institute of Medicine Implementation and Dissemination of Prevention Programs : O'Connell M, Boat T, Warner K, National research Council (US) and Institute of Medicine (US) Committee on the prevention of mental disorders and substance abuse among children, youth, and young adults: research advances and promising interventions. Washington DC: National Academies Press, 2009. [Google Scholar]
  • 58. Creswell JW, Clark VLP. Designing and conducting mixed methods research. Thousand Oaks: Sage, 2011. [Google Scholar]
  • 59. Nigenda G, González-Robledo LM, Juárez-Ramírez C, et al. . Understanding the dynamics of the Seguro popular de Salud policy implementation in Mexico from a complex adaptive systems perspective. Implementation Science 2015;11 10.1186/s13012-016-0439-x [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 60. Galárraga O, Sosa-Rubí SG, Salinas-Rodríguez A, et al. . Health insurance for the poor: impact on catastrophic and out-of-pocket health expenditures in Mexico. Eur J Health Econ 2010;11:437–47. 10.1007/s10198-009-0180-3 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 61. Green L, Nasser M, Brownson R, et al. . Furthering dissemination and implementation research: the need for more attention to external validity. Dissemination and implementation research in health: Translating science to practice 2012;19:305–26. [Google Scholar]
  • 62. Avan BI, Berhanu D, Umar N, et al. . District decision-making for health in low-income settings: a feasibility study of a data-informed platform for health in India, Nigeria and Ethiopia. Health Policy Plan 2016;31(suppl 2):ii3–11. 10.1093/heapol/czw082 [DOI] [PMC free article] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Supplementary data

bmjgh-2018-001257supp001.pdf (87.8KB, pdf)

Supplementary data

bmjgh-2018-001257supp002.pdf (555KB, pdf)

Supplementary data

bmjgh-2018-001257supp003.pdf (48.4KB, pdf)


Articles from BMJ Global Health are provided here courtesy of BMJ Publishing Group

RESOURCES