Skip to main content
BMJ Global Health logoLink to BMJ Global Health
. 2019 Oct 31;4(5):e001638. doi: 10.1136/bmjgh-2019-001638

Realist evaluations in low- and middle-income countries: reflections and recommendations from the experiences of a foreign researcher

Brynne Gilmore 1,
PMCID: PMC6830045  PMID: 31749993

Abstract

Realist evaluation, a methodology for exploring generative causation within complex health interventions to understand ‘how, why and for whom’ programmes work, is experiencing a surge of interest. Trends indicate that the proliferation in the use of this methodology also applies to research in low- and middle-income countries (LMICs). The value of using realist evaluation for project evaluation is also being noticed by non-governmental organisations (NGOs) and other programme implementers within such contexts. Yet, there is limited exploration of the use of realist evaluations in LMICs, especially their use by foreign researchers. This paper draws on the author’s experience of conducting two realist evaluations across three different sub-Saharan African settings: Mundemu, Tanzania; Kabale, Uganda and Marsabit, Kenya. The realist evaluations were used as an operations research methodology to study two NGO community health programmes. This paper highlights four main challenges experienced by the author throughout the methodological process: (1) power imbalances prevalent during realist interviews, (2) working through translation and what this means for identfying Context-Mechanism-Outcome Configurations, (3) limited contextual familiarity and being an ‘engaged researcher’ and (4) the use or dependence on ‘WEIRD’ theories (i.e. theories based on the study of Western, Educated, Industrialized, Rich, Democratic people) within testing and refinement. Realist evaluation’s enticing and straightforward slogan of finding ‘what works, for whom and why’ is in contrast to the complexity of the methodology used to generate these results (and often to the results themselves). Striking a balance between theory and pragmatism, while adhering to realist ontological underpinnings of generative causation and retroduction, is no easy task. This paper concludes by providing concrete recommendations for those who want to undertake a realist evaluation, with particular attention to cross-cultural settings, in light of the aforementioned challenges. In doing so, it aims to foster improved methodological rigour and help those engaging in this research methodology to work towards more appropriate and contextually relevant findings.

Keywords: child health, health systems evaluation, maternal health, other study design


Summary box.

  • Realist evaluations are increasingly being used in low- and middle-income countries and cross-cultural settings to study complex health interventions.

  • Methodological challenges of using realist evaluation in such circumstances have yet to be thoroughly explored.

  • Issues of power imbalance, using translators, limited contextual familiarity and utilisation of inappropriate theories may be prevalent in such cases.

  • Highlighting methodological challenges of using realist evaluation in low- and middle-income countries can ensure methodological rigour.

Introduction

Realist evaluation, a methodology for understanding complex health interventions, is a form of theory-driven evaluation which acknowledges that interventions and their outcomes are subject to contextual influences. As such, a realist evaluator’s duty is to understand ‘how, why, for whom and under which conditions’ interventions work.1 To do this, realist evaluators identify Context-Mechanism-Outcome Configurations (CMOCs).1 These configurations describe how specific contextual factors (C) work to trigger particular mechanisms (M), and how this combination generates outcomes (O), thus introducing the concept of generative causality (the idea that mechanisms operate in specific contexts to generate outcomes). By exploring these configurations of change, realist evaluations aim to understand how a programme is expected to work within specific contexts and what conditions may hinder successful outcomes,1 2 in order to produce policy relevant findings that can be transferred across settings and contexts.3 4

The process of conducting a realist evaluation follows a cycle, as shown in figure 1. The evaluation will usually start and end with a theory or hypothesis about how the programme works. The study design, data collection, data analysis and data synthesis all contribute to testing and refining that theory. A multimethod evidence base and inclusion of various stakeholder groups as research participants is usually recommended,5 6 though not always necessary. Regardless of the data source, it is important that data collection and analysis work to refine the programme theory or theories by identifying generative causality. Data analysis should be retroductive, which refers to ‘the identification of hidden causal forces that lie behind identified patterns or changes in those patterns’.7 Retroduction includes the researcher’s insights and can use both deductive and inductive reasoning to identify generative causation.8 By applying principles of generative causation and retroduction, ‘engaged realist’ researchers6 can elicit CMOCs and refine contextually relevant programme theories that explain how, why and for whom, interventions work (or do not work).9

Figure 1.

Figure 1

Realist evaluation cycle. Source: Mukumbang et al (2016).48

The Realist And Meta-narrative Evidence Synthesis: Evolving Standards (RAMESES) II Project has produced important guidance to support researchers in conducting and reporting realist evaluations.10 11 Their development was informed by published realist evaluation studies, the large majority of which occurred in high-income countries, and a Delphi study with 33 participants, all of whom were affiliated with institutions in high-income countries.10 Considerations for the use of this methodology within a wide variety of contexts is an important endeavour moving forward.

While realist evaluation offers an alternative approach to study complex health interventions, they are not without challenges, some of which can be attributed to a lack of procedures and precedent regarding its practice.12 A concern most frequently highlighted by realist researchers is the difficulty of defining ‘mechanisms’, and distinguishing them from ‘context’, both of which are tightly intertwined.3 5 12 13 It is also possible to find multiple combinations of mechanism and context which can bring about a variety of outcomes.14 There is limited guidance on how to code for and identify CMOCs and theories,8 which requires a substantial amount of time,15 researcher reflection and creativity.16 Striking a balance between theory and pragmatism, while adhering to realist ontological underpinnings of generative causation and retroduction, is no easy task and often requires researchers to fall back on their own experience.

Realist evaluation within low- and middle-income countries by foreign researchers

Realist evaluations are increasingly being employed within low- and middle-income countries by academics, non-governmental organisations (NGOs) and other programme implementers as evidenced by the rapid growth of the number of published realist evaluations from these countries.17–25 There remains, however, little precedent for its use within these countries. While the principles of realist evaluation should not change, challenges and implications for the use of this methodology in various settings must be acknowledged and carefully considered. This may be especially pertinent when the researchers are not from the settings where the work is taking place, and questions need to be asked on what influences an unfamiliar context or stakeholder groups has on a realist researcher’s ability to be ‘engaged’.

The main methodological challenges experienced by the author include the influence of power imbalances within data collection, use of translators for realist interviews, limited contextual understanding or engagement and identifying appropriate programme theories reflective of the context. Sharing lessons from practice and ensuring these important methodological discussions occur is an important step to move realist evaluation forward within these contexts. Several authors have shared reflections or recommendations for realist evaluation from projects that were conducted within LMICs;26 27 but their focus was not on the cross-cultural challenges. This paper therefore contributes towards the further development of this traction-gaining methodology and aims to inform future work as well as contribute new knowledge on how to conduct realist evaluations within cross-cultural contexts through experiences of a non-local researcher in an LMIC.

Findings and discussions presented in this paper arose from two mixed methods realist evaluations of community health interventions implemented by international NGOs. The first is of a Maternal, Newborn and Child Health (MNCH) intervention within Tanzania and Uganda that consisted of 6 case studies, with a total of 213 participants (116 in Uganda and 97 in Tanzania). This involved doing realist interviews and/or focus group discussions (FGDs) using realist techniques with research participants, majority of whom were rural community members. In total, 18 FGDs were held (three within each case study) and 38 individual interviews, 21 in Uganda and 17 in Tanzania. Interviews were conducted in English, Kiswahili (Tanzania) or Rukigian (Uganda).

The second study, currently ongoing, aims to explore community engagement for health behaviour change. It uses four mixed-method case studies across four different subcounties within Marsabit, Kenya. Data collection using a variety of methods including quantitative surveys, FGDs, semistructured interviews, key informant interviews, document reviews and observations, are being conducted at three different iterative timepoints throughout a 12-month intervention (baseline, midterm and endline). Non-English interviews are conducted in either Borana, Gabra, Rendille or Samburu. To date, 398 surveys, 10 FGDs, 35 interviews and 9 months of programme observations have been completed. The discussions that follow are based on the experience of conducting these two evaluations.

Both studies had coinvestigators from the country in which the research occurs. Coinvestigators specifically supported the design of the study and data collection tools and the analysis. Data collection was performed in each context by myself and a research assistant who provided translation. For the first study (Uganda and Tanzania), research assistants took part in a 3-day training. The training for research assistants for the second study in Kenya took place over 2 days with a 1-day refresher training before each additional round of data collection. Training in both studies included: the intervention of interest, the research study, realist methodology, realist data collection and interviewing and the interview guides. For any interview not conducted in English, the research assistant conducted the interview or FGD. I was present for all qualitative data collection within both studies.

The research studies this paper draws from are not without limitations. Notably, in the first study, there was no theory gleaning phase and initial programme theories (IPTs) were quite specific. Both studies had in-country researchers as coinvestigators who understood the local context; however, none of these were realist researchers. Additionally, all research assistants had experience in qualitative data collection, but none had previous experience in realist evaluation or realist interviews. Furthermore, while I had exposure to realist methodologies prior to the first study, this was my first complete realist evaluation and as such, learnt important lessons along the way—many of which I share below.

Challenge 1: power imbalances

The topic of power imbalances between researchers and research participants is extensively discussed in qualitative research methodologies.28 29 Due to the nature of the researcher-participant relationship, power imbalances may frequently and easily come into play, which can result in controlling or constraining views and respondent bias through enforcing opinions.29 Power is noted to be influenced by a number of factors, including educational or professional backgrounds, socioeconomic status, gender and ethnic identities.30 This imbalance of power may be exacerbated in less resourced contexts or cross-cultural research situations, is relevant in many global health research projects,31 and requires particular attention when working with marginalised populations.32 Power discrepancies between researcher and research participant that influence participant responses were also apparent while conducting realist interviews. However, this issue has yet to be thoroughly explored within realist evaluation methodology.

The realist interview is a method for data collection which uses semi-structured, theory-driven interviews to ‘confirm, refute or refine’ theories about how the programme works.6 9 33 Using the ‘teacher-learner approach’, the realist interview entails having the interviewer teach the research participant the theory being explored, and after learning the theory, the participant then teaches the researcher their ideas on that theory.6 A detailed discussion on the realist interview can be found in Pawson’s work ‘Theorizing the interview’.34 More recently, Manzano’s33 publication reviews interview techniques within realist evaluation noting that ‘surprisingly, qualitative interviewing is treated as unproblematic with little attention given to fieldwork processes or the act of the interview itself’.

The interview process conducted as part of the studies outlined here was found to be problematic for some participant groups (exception being key informants mostly outside of the community setting, like programme managers). Likely exacerbated by the rural, community-based setting in which this research was conducted, assumed power imbalances were believed to influence collaborative theory refinement. The introduction of theories or ideas was not often met with contradictions or substantial changes or elaborations from research participants. That is, the research participants’ answer was often the same as the researchers proposed idea, resulting in little theory refinement. Challenges in terms of constrained views and respondent bias were noted, highlighting that power imbalances may have influenced theory refinement.

Challenge 1: recommendations

Addressing power imbalances within research, especially community-based research, involves many considerations throughout the research cycle.35 36 Within the interview phase specifically, the research team tried different techniques to mitigate effects arising from these imbalances. All interviews included a short discussion that did not focus on theory testing or refinement, conducted prior to theory testing components, but rather asked the participants about their experience with the programme or other related topics.37 During this, the research team asked more open-ended questions that were then used to help facilitate the more specific theory testing portion of the data collection. Direct questions related to the programme theories being explored were often posed before the more theory testing questions, as demonstrated by this excerpt from an interview guide from the second study (note, wording of questions and further insight into realist interviewing was informed by the ‘Realist Qualitative Interviewing’ lecture, facilitated by Dr Justin Jagosh within the Advanced Study in Realist Methodology course38):

Q: Do you think that the trainings have made you more confident in caring for a sick child?

Q: There is this idea that, the more confident you are your ability, the more likely you are to seek care from a health worker. What do you think about that?

Q: What is it about being confident that leads you to consult a health worker?

Notably, and as a consequence of the research context and translation (discussed below), the questions/theories put forth were presented as straightforward as possible. In some sites, visual diagrams depicting the theories were used but were met with mixed results. This was less successful when diagrams were not simplified enough. The challenges and the strategies we deployed to overcome these led to the development of key recommendations for managing power imbalances within realist evaluation interviews. Please see box 1 for a full list of recommendations.

Box 1. Recommendations for reducing power imbalances.

  • Follow traditional qualitative methodology interview techniques of building rapport, active listening, body language and so on.

  • Use non-intimidating language and be mindful of how you introduce the conceptual structure of the interview.

  • Incorporate an open interview phase before working towards theory testing/refinement with the realist interview.

  • Consider using focus group discussions with groups that may be particularly influenced by power imbalances (if appropriate).

  • Keep visuals or other tools (ie, examples, vignettes) simple and to the point.

  • Be reflexive and document your interviews well, including if/how power may have influenced your interview. Ensure that you bring this into your analysis phase.

Challenge 2: working through translators and across language

It is likely that conducting a realist evaluation in a cross-cultural context will require translation while interviewing some stakeholder groups. To date, very little methodological writing has explored realist data collection through translation. As RAMESES II Reporting Standards requires researchers to ‘describe the steps taken to enhance the trustworthiness of data collection and documentation’,10 this topic is timely. Conducting research outside of your language and/or culture without a properly managed translation process can result in missing important pieces of data or difficulty in discerning the CMOC elements. Appropriately managing the translation process requires both careful practical and ontological considerations. Ontologically, I considered translation in realist research, like the paradigm itself, to sit in the middle of constructionist (decoding meaning) and more positivist (direct translation). While it was important to have a record of the direct translation, these had to be reviewed with subjectivity in mind.

Managing translation was met with several challenges. As an ‘engaged realist researcher’ working to identify CMOCs, it was difficult to have the interviewer use the semistructured guide with little input from myself throughout the course of interview. After reflecting on some early transcripts, this process was amended. Moving forward, the interviewer would work through the guide, but then translate participant’s contributions to theory (whether they be contexts, mechanism, outcomes or CMOCs) when they arose so that together, we, as engaged realists, could further explore these important aspects with research participants. As such, the English parts of the recordings of each interview compared with the translated transcripts of that interview provided almost as much important explanatory information.

Challenge 2: recommendations

Some of these challenges could be addressed through frequent communication and a constant feedback loop between researcher and translator/interviewer. In-depth methodological and tool training was required. Both translation and transcription needed to be followed by further unpacking meaning behind research participant’s responses and iteration became a necessity to ensure our understanding and interpretation of the data was accurate. Translators themselves constituted a separate research participant group; in-depth discussions were had on their understanding of theory refinement and CMOCs generated after interviews were conducted. Together, the teams reviewed the preliminary analysis and frequently discussed their programme theories. Being diligent on documenting reflexivity and decision-making processes also proved helpful, especially when reviewing transcripts at a later date.

It became clear that main considerations arising from translation within this work were less to do with eliminating bias (though not power, as discussed above) and keeping flow and more to do with ensuring that intended and unintended outcomes, and their generative causality, are captured and are contributing to theory refinement.10 Future realist work under similar conditions would benefit from a more intentional exploration of how best to conduct realist data collection through translation including consideration on how to incorporate translator’s understanding into the analysis of CMOCs and programme theories more robustly. Please see box 2 for a full list of recommendations.

Box 2. Recommendations for translation.

  • Plan for robust training on realist methods and realist interview techniques and allow for additional translation time.

  • When translating from the local language to that preferred for analysis, be cautious of ‘direct’ translation alone. Theory refinement relies less on the exact wording, and more on the meaning and explanations. Make sure you can balance both of these.

  • Consider translators as research subjects—after data collection is complete, interview them on their understanding of the intervention and theory refinement based on that interview.

  • Team data analysis or at least feedback findings to translators or other research assistants.

  • As with all realist evaluations, iteration becomes extremely important in when conducted through other language.

  • Ensure diligent documentation of decision-making processes and reflexivity.

Challenge 3: being ‘engaged’ with limited contextual familiarity

Realist evaluators are ‘engaged’ researchers—that is, we have an understanding of what is happening and why and use this to identify CMOCs and refine theories. It is advised, when possible, to have an understanding of the natural setting prior to interviews.37 It is, of course, impossible to know and understand any context in its entirety, but these challenges are exacerbated if the researcher is new to the research setting or a foreign researcher. This is often the case for global health research in LMICs and may be a particular challenge for realist evaluations. I believed, nevertheless, that during the studies presented here, I was a somewhat well-engaged realist researcher prior to commencement, having previously done work in Kabale (Uganda), having lived in Tanzania and having been seconded to the NGO in Kenya for 2 years. Yet, there were still many challenges in identifying context and mechanisms. Social context, including culture and tradition, power dynamics and relationships, require additional time and investment to begin to understand how they influence people’s reasoning and reactions when one is less familiar with such an environment. Without unpacking context and its influence on people in your setting, one can easily think that incomplete, or untrue, mechanisms are being triggered.

Challenge 3: recommendations

Several authors have provided recommendations that can support increasing contextual understanding and/or relevance of a realist evaluation. Ebenso et al.’s recent work highlights how embedding a realist evaluation within existing Ministry of Health policies and practice facilitated collaboration between researchers and in-country policymakers and implementers.24 Involving such collaborators across all stages of the realist evaluation ensured study relevance and contextual significance,24 which could also contribute to the uptake of findings.39

The studies discussed herein took a number of steps to mitigate contextual misunderstanding at the onset. Following the suggestion by Adams and colleagues,26 the study designs included multidisciplinary partnerships to enable insights that go beyond technical recommendations. Clear communication channels were established to support the contribution by team members, and procedures were taken to ensure the engagement of programme decision-makers throughout the research process. Reflexivity was also built into the process which enabled methodological challenges to be identified and addressed throughout the study.8 26

While the above likely contributed to greater contextual significance, based on experiences in the first study, additional steps were taken in the second study, including a detailed gender analysis informed by the Research in Gender Ethics group RinGs.40 This analysis was used to support the theory gleaning phase, and together interviews for theory gleaning and the gender analysis worked to identify IPTs that are then tested through two subsequent rounds of data collection. Frequent referral back to the gender analysis during CMOC identification and theory refinement continues to highlight its importance. Using such a tool may also contribute to issues described in Challenge 1, as applying a feminist lens within research can be an important resource to understand and address power.41 The incorporation of one of the many tools used within research and development sectors to understand context (ie, contextual analysis, gender analysis, political analysis and so on) warrants further consideration and exploration within realist evaluation methodology.

Communication was also found to be an important tool in this regard. It is important once an initial understanding has been reached through reading and engaging with a particular context to discuss (both formally and informally) with others involved and not involved in the research process. Your ability to do this might be predicated on your relationships and length of time spent within the environment: do not underestimate the importance of both of these. Please see box 3 for a full list of recommendations.

Box 3. Recommendations for contextual familiarity.

  • Embed yourself in the research setting as much, and as long, as possible prior to and during study.

  • Read outside of the specific discipline. While important for any realist research, might be especially important while working in an unfamiliar context.

  • Have a diverse research support team who have insights into the context of study and ensure they are meaningfully incorporated across the realist evaluation cycle (including translators).

  • Incorporate additional tools early on in theory elicitation, like gender analysis or contextual analysis.

  • Read, watch, listen and then ask questions to friends and colleagues—it is amazing what you can learn with a little insight and some open dialogue.

Challenge 4: programme theories—the detailed and the WEIRD

Many may fall victim to the desire to elicit IPTs with great detail and specificity; honestly, I did. However, when doing research in new or cross-cultural contexts, this could do your study a disservice. As per Challenge 3 above, if you are less familiar with the context of your research, IPTs (tentative theories proposed during early stages of the evaluation) may be misplaced. These early theories serve as the backbone for the rest of your study; poor IPTs will affect your subsequent activities which may mean missing out on some key elements and underlying theories. As summarised by Leeuw42 in relation to theories and interviewing, ‘to establish a dialectic atmosphere on the basis of empirically incorrect assumptions is inefficient and ineffective’.

This may be particularity important if formal theories are included, and specifically, WEIRD theories. WEIRD theories are those which have been developed through the study of ‘Western Educated Industrialized Rich Democratic’ individuals.43 Frequently belonging within the study of human behaviour and psychology, these more formal theories are often used as ‘universal’. Many question the ability of WEIRD theories to provide explanatory power to different contexts as they are developed through only studying ‘WEIRD’ people (whom are actually among the least representative populations).44 These broad claims on human psychology assume that there is little variation between individuals and populations; thus, a theory developed through the study of WEIRD people (often undergraduate psychology students) is applied to communities in rural Kenya without considerations for the people and context.

Challenge 4: recommendations

Incorporating a theory gleaning phase into research design is recommended,33 and particularly encouraged when conducting work in a cross-cultural setting or less familiar context. After this, iteration should still be included which ideally results in having minimum of three rounds of primary data collection for gleaning, refining and testing,33 though such processes can occur within the same round of collection.37 If not including a theory gleaning phase, keep early theories to more broad concepts that provide space for meaningful refinement and bring WEIRD theories in during later rounds of refinement (if at all) or keep their use for examining your specific findings within a wider body of knowledge.

Within the first study, the socioecological model was used as an abstract conceptual framework to organise early candidate/initial theories, which proved very helpful. Using abstract conceptual frameworks has also been encouraged when evaluating ‘large complex and messy programmes’.45 46 This can allow for structure to be introduced and to draw important relationships between concepts, without requiring substantive theories early on. In evaluations with NGO partners, an intervention’s Theory of Change (if available) could provide a starting framework that is already designed around notions of how change will occur. The realist field could benefit from an exploration of efforts to use an intervention’s Theory of Change as a candidate or initial programme theory. Comparing, challenging and questioning WEIRD theories in relation to your findings can help to elevate our dependence on them within potentially inappropriate contexts, and these findings should be shared widely. Please see box 4 for a full list of recommendations.

Box 4. Recommendations for theory.

  • Include a theory gleaning phase, with minimum two subsequent rounds of iteration to refine theories.

  • If no theory gleaning phase, keep early theories simple and allow for several iterative rounds of data collection.

  • Avoid using WEIRD theories during early stages of your research.

  • Bring WEIRD theories in to help explain your findings and situate them with wider the wider body of knowledge. If the findings allow, challenge or even refine WEIRD theories to suit your context.

Conclusion

Realist evaluation methodology is being used within LMICs and cross-cultural contexts, yet little work explores this research process. Critical reflection of practices and knowledge generation within global health research is needed to enhance processes and work towards more relevant research that can have a positive impact on health.32 This paper presents some reflections and recommendations based on experience, but recognises the need for more detailed discussions on these topics, and for others to reflect on challenges and solutions they encountered while undertaking realist evaluations in general and within cross cultural contexts. Figure 2 summarises recommendations across the realist evaluation cycle noted within this paper.

Figure 2.

Figure 2

Realist evaluation cycle and considerations for cross context research. Realist cycle adapted from Mukumbang et al (2016).48

Depending on the context and stakeholders, realist methods and tools might need careful consideration to address power imbalances. Exploring translators/translations within realist evaluations and how to conduct a realist interview to reduce power imbalances and promote theory refinement are especially encouraged. In less familiar contexts, considerations on how to ensure researchers can be properly engaged should be incorporated into the overall study design following examples set by Ebenso and colleagues27 and Adams and colleagues.26 This also aligns with continued calls for increased meaningful global health research collaboration.39 47

Particular attention should also be paid to early phases of the realist evaluation, leaving space for either theory gleaning or keeping theories simple. WEIRD theories should be used with caution, and researchers should disseminate their middle range theories widely so that there are more contextually specific theories others can draw from, to work to break the WEIRD conundrum. For realist evaluations within NGO interventions, exploring synergies between the intervention’s Theory of Change and realist candidate programme theories is recommended. Capacity strengthening for realist evaluation within such contexts is encouraged. Given that many realist evaluations in LMICs are being done for NGO programming to improve design and implementation, how best to ensure findings contain well refined theories that also have concrete and practical recommendations for action warrants further consideration.

These discussions need to continue, given the increasing interest from both academics and practitioners to use realist evaluation methodology within global health research and evaluations. Ensuring rigorous research that is responsive to the conditions in which it is being conducted will improve methodological rigour and make for more useful, reflective and accurate programme findings and recommendations.

Acknowledgments

The author would like to acknowledge World Vision Ireland, World Vision Uganda, World Vision Tanzania and Irish Aid for the support of the first study. They would also like to acknowledge Concern Worldwide and Concern Worldwide Kenya, and the Irish Research Council, Marie Skłodowska-Curie Actions and the European Union, for the support of the second study. They are also very grateful to all the research participants, programme staff and the research team and translators who supported this work. They would like to thank their advisors on both studies who were instrumental throughout the research process. They would also like to thank Jessica Power for their insights and readings of this manuscript, as well as the two reviewers who provided very helpful feedback.

Footnotes

Handling editor: Seye Abimbola

Funding: BG is currently a postdoctoral fellow, conducting the study in Marsabit, Kenya. This project has received funding from the European Union’s Horizon 2020 research and innovation programme under the Marie Skłodowska-Curie grant agreement No. 713279.

Competing interests: None declared.

Patient consent for publication: Not required.

Provenance and peer review: Not commissioned; externally peer reviewed.

Data availability statement: There are no data in this work.

References

  • 1. Pawson R. Evidence-Based policy. A realist perspective. London: Sage, 2006. [Google Scholar]
  • 2. Jagosh J, Pluye P, Macaulay AC, et al. . Assessing the outcomes of participatory research: protocol for identifying, selecting, appraising and synthesizing the literature for realist review. Implementation Science 2011;6 10.1186/1748-5908-6-24 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3. Salter KL, Kothari A. Using realist evaluation to open the black box of knowledge translation: a state-of-the-art review. Implementation Sci 2014;9 10.1186/s13012-014-0115-y [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4. Fletcher A, Jamal F, Moore G, et al. . Realist complex intervention science: applying realist principles across all phases of the Medical Research Council framework for developing and evaluating complex interventions. Evaluation 2016;22:286–303. 10.1177/1356389016652743 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5. Westhorp G. Development of realist evaluation models and methods for use in small-scale community based settings. Nottingham, UK: Nottingham Trent University, 2008. [Google Scholar]
  • 6. Pawson R, Tilley N, Evaluation R. Realistic evaluation. London: Sage, 1997. [Google Scholar]
  • 7. The RAMESES II Project Retroduction in realist evaluation, 2017. Available: http://ramesesproject.org/Standards_and_Training_materials.php#RAMESES_II
  • 8. Gilmore B, McAuliffe E, Power J, et al. . Data analysis and synthesis within a realist evaluation: toward more transparent methodological approaches. International Journal of Qualitative Methods 2019;18 10.1177/1609406919859754 [DOI] [Google Scholar]
  • 9. Pawson R, Sridharan S. Theory-driven evaluation of public health programmes : Killoran A, Kelly MP, Evidence-Based public health: effectiveness and efficiency. Oxford, UK: Oxford University Press, 2010. [Google Scholar]
  • 10. Wong G, Westhorp G, Manzano A, et al. . RAMESES II reporting standards for realist evaluations. BMC Med 2016;14:96 10.1186/s12916-016-0643-1 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11. Wong G, Westhorp G, Greenhalgh J, et al. . Quality and reporting Standards, resources, training materials and information for realist evaluation: the RAMESES II project. Health Services and Delivery Research 2017;5:1–108. 10.3310/hsdr05280 [DOI] [PubMed] [Google Scholar]
  • 12. Rycroft-Malone J, Fontenla M, Bick D, et al. . A realistic evaluation: the case of protocol-based care. Implementation Sci 2010;5 10.1186/1748-5908-5-38 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13. Marchal B, van Belle S, van Olmen J, et al. . Is realist evaluation keeping its promise? A review of published empirical studies in the field of health systems research. Evaluation 2012;18:192–212. 10.1177/1356389012442444 [DOI] [Google Scholar]
  • 14. Byng R, Norman I, Redfern S. Using realistic evaluation to evaluate a Practice-level intervention to improve primary healthcare for patients with long-term mental illness. Evaluation 2005;11:69–93. 10.1177/1356389005053198 [DOI] [Google Scholar]
  • 15. Tolson D, McIntosh J, Loftus L, et al. . Developing a managed clinical network in palliative care: a realistic evaluation. Int J Nurs Stud 2007;44:183–95. 10.1016/j.ijnurstu.2005.11.027 [DOI] [PubMed] [Google Scholar]
  • 16. Dalkin SM, Greenhalgh J, Jones D, et al. . What’s in a mechanism? Development of a key concept in realist evaluation. Implementation Sci 2015;10 10.1186/s13012-015-0237-x [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17. Maluka S, Kamuzora P, SanSebastián M, et al. . Implementing accountability for Reasonableness framework at district level in Tanzania: a realist evaluation. Implement Sci 2011;6 10.1186/1748-5908-6-11 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18. Marchal B, Dedzo M, Kegels G. Turning around an ailing district Hospital: a realist evaluation of strategic changes at HO municipal Hospital (Ghana). BMC Public Health 2010;10:787 10.1186/1471-2458-10-787 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19. Goicolea I, Coe A-B, Hurtig A-K, et al. . Mechanisms for achieving adolescent-friendly services in Ecuador: a realist evaluation approach. Glob Health Action 2012;5:18748 10.3402/gha.v5i0.18748 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 20. Van Belle SB, Marchal B, Dubourg D, et al. . How to develop a theory-driven evaluation design? lessons learned from an adolescent sexual and reproductive health programme in West Africa. BMC Public Health 2010;10:741 10.1186/1471-2458-10-741 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21. Prashanth NS, Marchal B, Devadasan N, et al. . Advancing the application of systems thinking in health: a realist evaluation of a capacity building programme for district managers in Tumkur, India. Health Res Policy Syst 2014;12 10.1186/1478-4505-12-42 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22. Gilmore B, McAuliffe E, Larkan F, et al. . How do community health committees contribute to capacity building for maternal and child health? A realist evaluation protocol. BMJ Open 2016;6:e011885 10.1136/bmjopen-2016-011885 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 23. Graham HR, Bakare AA, Gray A, et al. . Adoption of paediatric and neonatal pulse oximetry by 12 hospitals in Nigeria: a mixed-methods realist evaluation. BMJ Glob Health 2018;3:e000812 10.1136/bmjgh-2018-000812 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 24. Limbani F, Thorogood M, Gómez-Olivé FX, et al. . Task shifting to improve the provision of integrated chronic care: realist evaluation of a lay health worker intervention in rural South Africa. BMJ Glob Health 2019;4 10.1136/bmjgh-2018-001084 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25. Mukumbang FC, Marchal B, Van Belle S, et al. . Unearthing how, why, for whom and under what health system conditions the antiretroviral treatment adherence Club intervention in South Africa works: a realist theory refining approach. BMC Health Serv Res 2018;18:343 10.1186/s12913-018-3150-6 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 26. Adams A, Sedalia S, McNab S, et al. . Lessons learned in using realist evaluation to assess maternal and newborn health programming in rural Bangladesh. Health Policy Plan 2016;31:267–75. 10.1093/heapol/czv053 [DOI] [PubMed] [Google Scholar]
  • 27. Ebenso B, Manzano A, Uzochukwu B, et al. . Dealing with context in logic model development: reflections from a realist evaluation of a community health worker programme in Nigeria. Eval Program Plann 2019;73:97–110. 10.1016/j.evalprogplan.2018.12.002 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 28. Kvale S. Dominance through interviews and Dialogues. Qualitative Inquiry 2006;12:480–500. 10.1177/1077800406286235 [DOI] [Google Scholar]
  • 29. Wang J. Questions and the exercise of power. Discourse & Society 2006;17:529–48. 10.1177/0957926506063127 [DOI] [Google Scholar]
  • 30. Anyan F. The influence of power shifts in data collection and analysis stages: a focus on qualitative research interview. Qualitative Report 2013;18. [Google Scholar]
  • 31. Benatar SR, Fleischer TE. Ethical issues in research in low-income countries [Leading Article]. The International Journal of Tuberculosis and Lung Disease 2007;11:617–23. [PubMed] [Google Scholar]
  • 32. Keikelame MJ, Swartz L. Decolonising research methodologies: lessons from a qualitative research project, Cape town, South Africa. Glob Health Action 2019;12:1561175 10.1080/16549716.2018.1561175 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 33. Manzano A. The craft of interviewing in realist evaluation. Evaluation 2016;22:342–60. 10.1177/1356389016638615 [DOI] [Google Scholar]
  • 34. Pawson R. Theorizing the interview. Br J Sociol 1996;47:295–314. 10.2307/591728 [DOI] [Google Scholar]
  • 35. Israel BA, Schulz AJ, Parker EA, et al. . Review of community-based research: assessing partnership approaches to improve public health. Annu Rev Public Health 1998;19:173–202. 10.1146/annurev.publhealth.19.1.173 [DOI] [PubMed] [Google Scholar]
  • 36. Sriram V, Topp SM, Schaaf M, et al. . 10 best resources on power in health policy and systems in low- and middle-income countries. Health Policy Plan 2018;33:611–21. 10.1093/heapol/czy008 [DOI] [PubMed] [Google Scholar]
  • 37. Greenhalgh T, Pawson R, Wong G, et al. . The Realist Interview In: The RAMESES II project, 2017. [Google Scholar]
  • 38. Jagosh J. Realist qualitative interviewing, in advanced study of realist methodology course. Liverpool University, 2018. [Google Scholar]
  • 39. Kok MO, Gyapong JO, Wolffers I, et al. . Towards fair and effective North–South collaboration: realising a programme for demand-driven and locally led research. Health Research Policy and Systems 2017;15 10.1186/s12961-017-0251-3 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 40. RinGs Research in gender and ethics (rings): building stronger health systems. How to do gender analysis in health systems research: a guide. Liverpool, 2016. [Google Scholar]
  • 41. Reid C. Advancing women's social justice agendas: a feminist action research framework. International Journal of Qualitative Methods 2004;3:1–15. 10.1177/160940690400300301 [DOI] [Google Scholar]
  • 42. Leeuw FL. Reconstructing program theories: methods available and problems to be solved. Am J Eval 2003;24:5–20. 10.1177/109821400302400102 [DOI] [Google Scholar]
  • 43. Henrich J, Heine SJ, Norenzayan A. Most people are not WEIRD. Nature 2010;466:29 10.1038/466029a [DOI] [PubMed] [Google Scholar]
  • 44. Henrich J, Heine SJ, Norenzayan A. The weirdest people in the world? Behav Brain Sci 2010;33:61–83. 10.1017/S0140525X0999152X [DOI] [PubMed] [Google Scholar]
  • 45. Shearn K, Allmark P, Piercy H, et al. . Building realist program theory for large complex and messy interventions. Int J Qual Methods 2017;16 10.1177/1609406917741796 [DOI] [Google Scholar]
  • 46. Westhorp G. Using complexity-consistent theory for evaluating complex systems. Evaluation 2012;18:405–20. 10.1177/1356389012460963 [DOI] [Google Scholar]
  • 47. Boum II Y, Burns BF, Siedner M, et al. . Advancing equitable global health research partnerships in Africa. BMJ Global Health 2018;3:e000868 10.1136/bmjgh-2018-000868 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 48. Mukumbang FC, Van Belle S, Marchal B, et al. . Realist evaluation of the antiretroviral treatment adherence club programme in selected primary healthcare facilities in the metropolitan area of Western Cape Province, South Africa: a study protocol. BMJ Open 2016;6:e009977 10.1136/bmjopen-2015-009977 [DOI] [PMC free article] [PubMed] [Google Scholar]

Articles from BMJ Global Health are provided here courtesy of BMJ Publishing Group

RESOURCES