Skip to main content
Gates Open Research logoLink to Gates Open Research
. 2022 Sep 27;6:13. Originally published 2022 Mar 16. [Version 2] doi: 10.12688/gatesopenres.13417.2

Monitor to innovate with feedback loops: process evaluation protocol for an anemia prevention intervention

Ichhya Pant 1,a, Lipika Patro 2, Erica Sedlander 1, Shikha Chandrana 1, Rajiv Rimal 3
PMCID: PMC9525569  PMID: 36226259

Version Changes

Revised. Amendments from Version 1

Version 2 has been uploaded accounting for reviewer feedback. We have added clarifying text to ensure reviewers and readers are aware that this is a process monitoring protocol and not a randomized control trial protocol. Specifically, this protocol focuses on identifying and incorporating near-to-real-time feedback loops and such an approach does not lend to establishing a theory of change as is typical within field trial protocols. We have also added clarifying text to detail how our approach was a mixed-methods approach. Additionally, we've revised our manuscript to provide further details regarding the algorithm incorporated in the back end of RPM4K and its functionalities. Finally, we've also added a brief conclusion section to summarize our overall protocol.

Abstract

Background: Digital process monitoring and evaluation tools designed to capture near-to-real-time intervention data paired with feedback loops have the potential to innovate intervention delivery.

Objective: To describe how a multilevel social norms field trial (RANI) is using feedback loops to enhance intervention delivery.

Methods: We use a mixed-methods process evaluation design to monitor the Reduction of Anemia through Normative Innovations (RANI) project; a three-year randomized control trial which aims to lower rates of anemia among women in Odisha, India. Surveys and structured observation monitor fidelity to implementation and acceptability of implementation activities among study participants. Quantitative data evaluates implementation dose, coverage, exposure, and reach of intervention activities, and qualitative data will delve more deeply into reasons for high or low functioning. Iron folic acid supplement supply and demand are also monitored for stock-outs. Data collected from 130 intervention villages is processed, visualized, and triangulated in near to real-time via Real-time Monitoring for Knowledge Generation (RPM4K), a locally developed software application. Data visualization products facilitate the examination of monitoring data to mitigate bottlenecks and identify and implement tweaks to our intervention delivery strategy on an ongoing basis.

Discussion: Feedback loops facilitate timely course corrections. Feedback loops can also engender a shared understanding of ground realities for a geographically dispersed and culturally diverse team. Leveraging feedback loops, we identify opportunities to provide on-going supportive supervision for our community facilitators promoting joint problem-solving, and communication. Monthly media and hemoglobin level demonstration strategies are informed by participant engagement and acceptability. Stock-outs of iron folic acid tablets activate contingency plans to mobilize local stakeholders and advocate for timely resolutions. Unintended effects are monitored based on ongoing feedback from community facilitators.

Conclusions: Documenting our processes can inform the future implementation or scale up of similar projects embracing feedback loops to iterate and innovate their intervention delivery.

Keywords: neat to real time, monitoring, evaluation, feedback loops, social norms, rural, India, resource constrained

Introduction

Anemia endemic among women of reproductive age

Anemia affects roughly a third of the world's population, making it a major global public health problem that impacts maternal and child mortality, cognitive and physical performance 1 , and work capacity and productivity 2, 3 . In India, more than 50% of women of reproductive age have anemia 3 . The largest contributor to anemia worldwide is iron deficiency 4 . Despite concentrated efforts and the presence of several government programs (e.g., National Iron Plus Initiative and Anemia Mukt Bharat) 5 , which offer women of reproductive age free iron supplements 69 , anemia prevalence remains high 3 . India’s Demographic Health Data (2015–16) shows that only 30% of pregnant women consumed iron folic acid for more than 100 days during pregnancy 3 .

Several studies have examined supply side issues to IFA consumption 1013 , but studies on demand side challenges including knowledge and social barriers are limited 14 . Recently, researchers have begun to ask questions about the role that community-level factors, such as social and gender norms, can play in women’s IFA consumption 1518 . The Reduction in Anemia Through Normative Innovations (RANI) project seeks to bridge this gap with a multilevel intervention focused on increasing demand for IFA by influencing anemia-related social norms 19 .

The RANI (Reduction in Anemia through Normative Innovations) project

The RANI project is a cluster randomized trial launched in 2019 in two blocks, Athamalik and Kishorenagar, within the Angul district of Odisha on the eastern coast of India. These blocks are spread over 1278 sq. kilometers and they constitute a total of 558 villages with a total of 218, 373 people in greater than 50,000 households per the 2011 census 20 . Here, the majority of residents (83%) live in rural areas, a third are literate, and primarily identify as Hindu 20 . Like the rest of the country and the district of Odisha, India, almost half of all women of reproductive age are anemic 14 . In the study, there are 89 selected clusters of villages, which we randomized into treatment and control on a 1:1 basis. The usual-care control arm receives no intervention, whereas the treatment arm receives the RANI project components. In total, 15 clusters (40–41 villages) were selected and 4000 women (2000 in each arm) living in the selected clusters were randomly selected to take part in data collection.

The RANI Project’s overall implementation approach, based on its goal to improve IFA consumption through a social norms-based intervention, includes a number of intervention components (see Figure 1 for a visual description of each intervention component), which require monitoring and iterating the intervention delivery strategy at multiple ecological levels throughout the implementation period. Ethical approvals to conduct the RANI project have been acquired from appropriate institutions within the United States and in India (see Yilma et al. 2020 for the full intervention protocol 19 ).

Figure 1. Visualization and description of RANI Project intervention components.

Figure 1.

Process evaluation framework

In public health interventions, process evaluation allows researchers and practitioners to understand what works, what does not work, for whom, and under what conditions. It also serves the important function of providing ongoing feedback so that interventions can be responsive to the changing conditions in the field 2123 . Process evaluations also help determine whether fidelity to the implementation protocol was maintained during the course of the intervention period, whether intended participants were reached, and if any unintended consequences resulted from the project 24, 25 .

Health care delivery and outcomes can be improved by using innovations (i.e., new ideas, technologies, and practices) supported by scientific evidence. Over the last couple of decades, process evaluation methodologies have been evolving 21 . They were originally centered on qualitative research alongside trials and were conducted to provide a deeper understanding of the disease condition, implementation issues and mechanisms of the intervention 22 . Now, there is growing consensus that qualitative and quantitative data (mixed methods) can help facilitate trial implementation, identification of the effectiveness of “active ingredients”, and research translation. As methods have advanced, integration of feedback loops is emerging as the next methodological frontier for process evaluations.

Use of feedback loops in Process Monitoring and Evaluation

Feedback loops are mechanisms that allow intervention information to flow back to the program implementers on an ongoing basis so that appropriate changes in intervention delivery strategy can be adopted throughout the implementation period. Incorporating feedback loops into the process evaluation design ensures interventions are meaningfully and strategically responsive to emerging requirements. In modern times, these loops increasingly rely on mobile and other forms of digital technology, which improves accuracy and reduces the time lag between events on the ground and program adaptations. As such, feedback loops can improve decision-making throughout the intervention.

Historically, institutionalizing feedback loops as an integral part of process evaluations has proven to be challenging 2629 . Many scholars 29 have characterized the need for feedback loops “as a bold challenge to current orthodoxy aided by developments in theory, methods, and practice” (p. 3). They call for an approach that promotes interaction between project designers, implementers, researchers, and decision-makers to encourage adaptation through learning. Such an approach would center on agility, responsiveness, a culture of experimentation, responsiveness, and relevant, timely, actionable data as its key characteristics. Furthermore, scholars advocate moving away from current dominant models of static intervention design and implementation, which tend to be modeled after trials that do not permit projects to be responsive to the complexities and unpredictability of implementation challenges that arise in social and behavioral intervention trials 29 .

The use of feedback loops is, however, mainstream in the commercial sector for corporations such as Uber, eBay, and Airbnb 3033 . Similarly, there are several examples of improved implementation delivery and outcomes by leveraging feedback loops. They range from engaging millions of youths in informing policy, improving use of evidence in decision-making, improving flood response time, and efficiency in HIV treatment targeting in Zimbabwe 3440 . Citizen reports of drug stock-outs and improvement in community health worker performance 34, 4143 are other examples of the effective application of feedback loops 34, 4446 .

Motivation for incorporating feedback loops

The synergy between feedback loops integrated within process evaluations and enhancement of implementation delivery is still an under-explored area, especially in the context of randomized control trials (RCTs) 41 . One of the monitoring goals of the RANI Project was to reduce this knowledge and implementation gap. Specifically, we describe the process evaluation methods we are adopting in a multilevel intervention to increase iron folic acid supplement use for anemia prevention in Odisha, India.

In sum, the synergy between feedback loops integrated within process evaluations and enhancement of implementation delivery is still an under-explored area, especially in the context of randomized control trials (RCTs) 47 . One of the monitoring goals of the RANI project was to reduce this knowledge and implementation gap. Specifically, we describe the process evaluation methods we are adopting in a multilevel intervention to increase iron folic acid supplement use for anemia prevention in Odisha, India.

Methods

We use a mixed methods approach which has largely benefited from key pieces of literature 35, 47, 48 that detail systematic approaches to designing and incorporating feedback loops to facilitate adaptive management for health improvement interventions, and we have adapted them on an ongoing basis to meet the RANI project’s specific needs. Quantitative data is collected via surveys and structured observation is used to monitor implementation fidelity (i.e., dose, coverage, exposure, reach of intervention activities, village-level supply of and demand for iron folic acid supplements), and acceptability among program participants. In parallel, monthly qualitative reports provide insights into barriers and facilitators to intervention delivery. Overall, our globally dispersed team leverages near to real-time data to mitigate bottlenecks and pain-points as well as leverage opportunities to enhance implementation delivery. We collectively identify, agree upon, and incorporate tweaks to our intervention delivery strategies on an ongoing basis.

Ethics approval and consent to participate

Approval to conduct the study was gained from The Institutional Review Board at The George Washington University (FWA00005945) and the Sigma Science and Research, an IRB located in New Delhi, India (10031/IRB/D/18-19). This trial was registered with Clinical Trial Registry- India (CTRI) (CTRI/2018/10/016186) on 29 October 2018. All participants went through a verbal and written informed consent process before data collection.

Process evaluation aims and questions

The RANI project process evaluation has three aims.

  • 1.

    To assess fidelity to implementation while guiding intervention planning and delivery for the larger trial via feedback loops.

  • 2.

    To monitor the quality of intervention delivery and acceptability of the intervention among its intended audiences.

  • 3.

    To monitor the supply and demand of iron folic acid supplement in intervention delivery sites to mitigate possible supply chain disruption.

Our process evaluation questions are anchored in the RANI project’s process evaluation framework and aims. They were finalized in collaboration with our stakeholders ( Table 1), who serve as key members of the feedback loop decision-making sub-team (i.e., principal investigator, project director, M&E (monitoring and evaluation) technical lead, intervention implementation and evaluation manager) within our project.

Table 1. Process evaluation questions.

  1.  To what extent was fidelity to implementation maintained during the implementation of the T4 sessions, community
engagement meetings, media demonstrations, and hemoglobin testing sessions and demonstrations?
         a. What were the total number of implementation sessions conducted per village?
         b. Were community facilitators adequately supported during implementation delivery?
         c. Were T4 sessions informative and interactive for our program participants?
         d. What was the level of exposure to our media products at the village-level?
         e. What were our program participants’ reactions and acceptability of our implementation activities?
         f. What were the barriers and facilitators faced by our community facilitators and our target population?
         g. Were there any potential unintended consequences (neutral, positive or negative) as a result of our program?
         h. Was there adequate supply of Iron Folic Acid supplement at the village level?
         i. Did demand for Iron Folic Acid supplement increase or decrease at the village-level during the implementation period?
  2.  Did we reach RANI’s target audience with our implementation activities?
         a. What was the total reach of our implementation activities?
         b. Who did we reach with our implementation activities?

Following the finalization of our evaluation aims and questions, we developed a logic model and M&E planning framework (See Online Repository). This framework details key inputs, outputs, outcomes (short, intermediate, and long-term) and links them with data collection instruments. It outlines the frequency of data entry, responsible parties, reporting frequencies, data visualization tools, and a plan for disseminating our findings. Furthermore, the framework served as a guiding document when developing the wireframe and final design for our process evaluation software application, called the Real-time Performance Monitoring for Knowledge (RPM4K).

Real-Time Performance Monitoring for Knowledge (RPM4K)

RPM4K is a software application customized and developed locally for the RANI Project process evaluation. It facilitates intervention planning, delivery, performance monitoring, and automated production of data visualization and reports using near to real-time data to identify and inform feedback loops ( Figure 2). RPM4K application offers a dual mobile application and a web-based interface.

Figure 2. RANI Project’s Process Evaluation Conceptual Framework.

Figure 2.

The process evaluation data lifecycle centers and revolves around our community facilitators and their cluster supervisors. RANI community facilitators, who serve as frontline workers to deliver intervention components, use the system to enter field level data with the help of a hand-held (mobile) device. Their cluster supervisors use the mobile as well as the web interface for entering supervision data and also monitoring intervention delivery and quality by reviewing data entered by RANI community facilitators. Key stakeholders use the web interface to monitor the pace and delivery of intervention activities of both cluster supervisors and facilitators and utilize aggregated data to identify feedback loops and implement tweaks to the intervention. Access to RPM4K is tiered at multiple levels.

Community facilitators are only able to access their own performance data as well as intervention data for the villages they work in. Similarly, their cluster supervisors can only access performance data for the community facilitators they manage and the villages they work in. Key members of the process evaluation team (e.g., process evaluation lead, state implementation manager, project managers etc.) are able to access and view intervention data for all villages and performance data for all personnel. However, the system administrator is the only person with access to all structural functionalities of RPM4K. They are responsible for mapping intervention villages to cluster supervisors or community facilitators and defining data access for all users.

At the beginning of each month, the system administrator is responsible for uploading intervention plans onto RPM4K. They direct the RANI project’s process evaluation data lifecycle, quality control, and technical assistance requests from RPM4K end users. At month’s end, data on key performance indicators is auto aggregated into dashboards and reports in the form of easy to interpret heat maps and visual reports. Raw data for specific time periods or the entirety of the intervention delivery period can also be downloaded from RPM4K by users with defined data privileges.

Data collection

We collect, process, visualize, and triangulate quantitative, qualitative, and administrative data from 130 intervention villages in near to real-time using RPM4K. Data collected ranges from the individual level (e.g., community facilitator performance and hemoglobin levels) to the village-level (e.g., IFA supplement supply and demand). Key performance indicators for each ecological level are summarized in Table 2. A data quality assurance protocol is in place to review various levels and layers of data in a collaborative manner (See Supplementary Material). Data entry, management, review occurred in monthly cycles. Frontline community facilitators enter data in near-to-real time. Their supervisors conduct the first round of data review followed by a final check of data quality by the Data Manager prior to approving submitted data. Data quality review focused on conducting range checks for data values, monitoring and removing duplicate, blank or outlier values, timeliness with encoded deadlines for data entry submissions, completeness check against monthly implementation plans. All data was housed in a secure cloud server with full privileges limited to the Data Manager only.

Table 2. Key performance indicators summarized at various ecological levels.

MEASURE INDICATORS Individual Interpersonal Village
1. Fidelity – Dose 1.1 Number of participatory learning acitivity sessions X
1.2 Number of media demonstrations X
1.3 Number of women participating in Hemoglobin
testing
X X X
1.4 Number of community engagement meetings X X
1.5 Number of hemoglobin testing demonstrations X X X
1.6 Number of women tested as anemic (mild, moderate, or
severe)
X X
1.7 Number of referrals to community clinical linkages for
anemic women
X X
1.8 % of self-reported iron folic acid supplementation intake
status (currently taking, previously taken, or
never taken)
X X
2. Quality – Reach 2.1 Number of women of reproductive age reached X
2.2 Number of adolescent girls reached X
2.3 Number of mothers-in-law reached X
2.4 Number of pregnant women reached X
2.5 Number of frontline workers reached X
2.6 Number of policy workers reached X
2.7 Number of men reached X
3.1 CF performance X X
3.2 Qualitative barriers, facilitators, challenges, or
opportunities
X X X
3.3 Unintended consequences (neutral, positive or negative) X X X
4. Acceptability
– Reactions
4.1 Number of thumbs up for comm products X
4.2 Number of thumbs down for products X
4.3 Number of thumbs up for participatory learning
acitivitysessions
X
4.4 Number of thumbs down for participatory learning acitivity
sessions
X
4.5 Number of attendees willing to return to participatory
learning acitivity sessions and community engagement
meetings
X
4.6 Number of returning attendees for participatory learning
acitivity sessions and community engagement
meetings
X
4.7 Qualitative barriers, facilitators, challenges, or
opportunities
X X X
4.8 Unintended consequences (neutral, positive or negative) X X X
5. Iron Folic Acid
– Supply and
Demand
5.1 Number of iron folic acid supply points with readily
available stock
X
5.2 Number of iron folic acid supply points with IFA in short
supply
X
5.3 Number of iron folic acid supply points with IFA stock-outs X
5.4 Number of iron folic acid supply points reporting
an increase in demand
X
5.5 Qualitative barriers, facilitators, challenges, or
opportunities
X X X
5.6 Unintended consequences (neutral, positive or
negative)
X X X

Monthly qualitative summary report

RANI project community facilitators submit a monthly qualitative report summarizing their reflections, perceptions, and experiences regarding the quality of their intervention sessions, self-assessment of their performance, participant and community reactions and acceptability to intervention activities, barriers and facilitators and unintended consequences, if any (See Online Repository for the full interview guide). All open-ended responses get captured in RPM4K in Odiya, the local language, and then get translated and deductively summarized by the Data Manager using the interview guide to formulate themes in a monthly report written in English using Microsoft Word.

Quantitative measures and instruments

The majority of quantitative data is collected and entered into RPM4K through the Community Facilitators Input Form (submitted by community facilitators), the Community Facilitators Evaluation Form (submitted by their cluster supervisors). These forms have dedicated sections dealing with key performance indicators for each intervention component (i.e., T4, Hemoglobin testing, RANI Comms, IFA supply and demand status, etc.). The forms also have built-in global positioning system (GPS) and activity image capture features to associate location and visual data with intervention activities. All forms are submitted within four days of completing an intervention activity. See Online Repository to review our data collection instruments and screenshots of RPM4K interface, dashboards, and data visualization products.

Dose

Dose is operationalized as the total number of intervention activities (e.g., T4 sessions, community engagement meetings, RANI Comm video demonstration sessions, hemoglobin testing sessions and results demonstrations) reported on RPM4K and summarized at the village-level by algorithms embedded within RPM4K.

Reach

Reach is measured as the total number of women of reproductive age, their families (i.e., husbands and mothers-in-law), frontline workers and policy makers in the village exposed to intervention activities. Reach is summarized at the village-level by algorithms embedded within RPM4K.

Acceptability

Acceptibility of intervention components is operationalized as the total number of likes versus dislikes shared by program participants following the delivery of an intervention activity by RANI community facilitators. Acceptability is aggregated at the village-level by algorithms embedded within RPM4K as the total number of likes and dislikes allocated to all delivered intervention components. Acceptibility is also assessed via qualitative feedback noted by RANI community facilitators made on a monthly basis ( see supplementary materials to review how we captured acceptability of RANI intervention components via the monthly generated qualitative reports).

Community facilitator performance

Community facilitator performance is assessed and monitored using three distinct tools implemented at different timepoints. The first tool measures the baseline professional abilities of RANI community facilitators when they are first onboarded as part of the RANI project implementation team. Each RANI community facilitator is classified on a scale of 1 (exceptional) to 3 (needs improvement) points on six measures – 1) their prior work experience in participatory learning activities or community mobilization, 2) communication skills, 3) professional network, 4) familiarity with the intervention area, 5) references from prior colleagues or supervisors, 6) logistical preparedness related to job requirements.

The second tool for assessing and monitoring community facilitator performance is integrated within RPM4K. It consists of a self-assessment scale reported by community facilitators following every intervention session followed by monthly cluster supervisor assessment of community facilitators. The community facilitator self-assessment scale consists of three measures which ask 1) how easy or difficult it was to deliver the intervention session they’re reporting on, 2) how much the community facilitator thinks participants understood each session, and 3) how they would rate their overall performance for the session they just delivered. The response options consist of five-point Likert scales ranging from very difficult to very easy, did not understand at all to understood all of the content, and very poor to very good, respectively. Similarly, cluster supervisors conduct random visits covering at least 10% of all intervention sessions delivered by a community facilitator during any given month. They perform intervention session evaluations and report community facilitator performance by observing and assessing their pace, accuracy, confidence, and ability to hold the audience’s attention when delivering intervention components. The response options consist of five-point Likert scales ranging from too slow to too fast and inaccurate to accurate; three-point Likert scales ranging from not at all confident to very confident and not at all focused to very focused. Data from these two tools are summarized at the village-level by algorithms embedded within RPM4K.

The third and final tool is utilized and reported with varying periodicity ranging from monthly to quarterly assessments while community facilitators work in the field. Each RANI community facilitator is classified on a scale of 1 (exceptional) to 3 (needs improvement) points on several measures, which assess programmatic and human resource related performance such as a community facilitators’ ability to deliver intervention content and respond to intervention feedback loops in a timely manner, need for supportive supervision, and community mobilization skills. This data is collected by the district implementation management team and analyzed using Microsoft Excel.

IFA supply and demand

In 2019, the RANI project’s implementation leadership team based in Bhubaneshwar, Odisha conducted an initial assessment to map the IFA supply chain, supply status and monitoring techniques in intervention villages before intervention delivery began. Using this information, they operationalized how to monitor IFA supply status, developed tentative ideas for streamlining IFA supply in the intervention villages, and designed an IFA supply and demand monitoring module for RPM4K. IFA supply status is measured as the total number of supply points reporting adequate supply, shortage of supply or stock-outs. Similarly, IFA demand is measured as the total number of supply points reporting an increase in demand. IFA supply and demand is summarized at the village-level by algorithms embedded within RPM4K.

Hemoglobin levels

Hemoglobin testing gets conducted monthly among 15 women, using a HemoC ue photometer (model HB 301), which provides instant results in terms of hemoglobin concentration in gm/dL (grams per deciliter). This group consists of five women who volunteered as repeat testers every month and ten women who have not been tested before through the RANI project. Test results are shared as color coded (green – anemia free; yellow – mild anemia; orange – moderate anemia; red – severe anemia) blood shaped cards visualizing severity of anemia or lack thereof. Post testing, recommendations for behavioral actions and nudges are tailored according to participant’s test results. Measurement and monitoring of anemia levels at the village-level are analyzed and visualized using embedded algorithms and heat maps on RPM4K subsequently to monitor trends at the village-level.

Identifying, documenting and implementing feedback loops

Generally, process evaluation data are collected and analyzed at the end of the intervention delivery period, which limits the understanding and insights stakeholders have on barriers and facilitators on the ground 8 . The merger of feedback loops with applications such as RPM4K allow complex, multi-component interventions like RANI to assess ground-realities to innovate intervention delivery strategies in near-to-real-time by triangulating quantitative intervention process monitoring data, qualitative monthly reports, and administrative data related to community faciliator performance. To do this systematically and efficiently, for every scenario that emerges, the M&E team first triangulates quantitative and qualitative monitoring data collected on a near-to-real-time basis and then identifies relevant tailoring variables 48 . They are defined as variables that serve as indicators of when an intervention delivery strategy requires optimization or calibration by the intervention delivery team. For the RANI project, the optimization or calibration takes the form of feedback loops. Table 3 provides several examples of hypothetical scenarios, measures, tailoring variables, and potential feedback loops identified to explicate this process in practice. All feedback loops are currently documented in an offline tool and will be built on as a formal module of RPM4K during future upgrades to the application (see Online Repository to review the RANI Project’s Feedback Loops Documentation Tool).

Table 3. Hypothetical Examples of Tailoring Variables and Feedback Loops.

SCENARIO MEASURES TAILORING
VARIABLE
FEEDBACK LOOPS
There are waitlists to get Hemoglobin
testing by women in X % of villages
or X village the week the intervention
kicked off
Fidelity – Dose Hemoglobin testing
resource allocation
○ Continue to monitor demand (i.e.,
number of villages with waitlists and
number of women on waitlists)
○ Refer women to the near clinical
services where they can get tested
for anemia
○ Pivot intervention resources
towards allocating more resources
towards testing
Community facilitator performance is
low for X % of Community facilitators
or Community facilitators from X % of
supervisory teams
Quality – Community
facilitator performance
Supportive
supervision
○ Identify right type and time of
support needed for community
facilitator
○ Identify right type and time of
support needed for community
facilitator
Iron folic acid supply is low X% of
villages
Iron folic acid supply Policy advocacy ○ Continue to monitor IFA stock-outs
○ Engage local officials and frontline
workers to advocate for iron folic
acid supply in impacted villages
○ Refer women to private supply
points where iron folic acid supply is
readily available although may not
be free of cost
○ Refer women to nearby villages
where iron folic acid is readily
available
Rumors regarding Activpal utilized
during baseline data collection by
another firm
Unintended
Consequences
Rumor
management
○ Designate as urgent or non-urgent
matter
○ Continue to monitor prevalence of
rumors
○ Shed light and awareness on the
instrument

Tailoring variables for the RANI Project primarily focus on:

      1] Supportive supervision to bolster community facilitator performance. In general, a group of 7-8 RANI community facilitators are supported by a cluster supervisor. Overall, the supervisor helps the community facilitator gain mastery over the session content by organizing field level demonstrations. The supervisor provides on-site support in mobilizing participants and delivering intervention sessions. They provide constructive feedback to the community facilitator for improving session quality and participation levels based on both direct observation of session delivery and interaction with community members. Any discussion with policy makers and sections of the village which are unwilling to participate in the intervention or coordination with local health care officials are also facilitated with support from cluster supervisors.

Beyond cluster supervisors, the intervention management team also conducts monitoring visits and provides feedback to the field teams for improving intervention delivery. All broader issues identified in the field are shared during monthly meetings. Protocols and audio recordings for improving intervention delivery in local vernacular are developed for community facilitators and cluster supervisors on a regular basis. The entire intervention delivery team is also part of WhatsApp groups where updates, plans, requests for technical assistance for day-to-day activities are shared by all team members with instant feedback by the intervention management team.

      2] Refining RANI comm strategies. We assess how each video resonates with our program participants informed by acceptability indicators (both qualitative and quantitative) for RANI comm videos data on RPM4K. For example, if we find that adolescents are more receptive to the videos than older women, we pivot to show videos at events where adolescents are more likely to attend (e.g., a festival or school-related event) and increase other intervention activities among older women.

      3] Altering hemoglobin testing strategies. We alter our implementation plans based on hemoglobin testing data. For example, when we find that there is a high demand for hemoglobin testing in our intervention villages, we review quantitative and qualitative reports to examine how test results are being perceived and if they are influencing IFA use. If the reports indicate that hemoglobin testing is having the desired effect, we consider increasing the number of women that we test while remaining mindful of resource constraints.

      4] Policy advocacy to accommodate IFA supply or demand disruption. In the Angul district of Odisha, the government provides daily IFA supplementation for pregnant women and weekly IFA for adolescents for free. While adolescents and pregnant women can acquire IFA free of cost, non-pregnant women of reproductive age remain uncovered. Keeping this coverage gap in mind, the RANI project intervention team assesses and shares requirements for providing IFA supplements to non-pregnant women with local officials and frontline workers on a monthly basis. Every month after village-level hemoglobin testing is completed, aggregated test results are shared with local health care workers and higher-level administration at the intervention sites. This information is supplemented with a demand for IFA as per the aggregated anemia levels of pregnant and non-pregnant women of reproductive age in the intervention villages. IFA stock is usually replenished and delivered within a fortnight. In case of any delay or shortfall, the issue is escalated first at the block level and then to the district level, if required.

      5] Engaging in participatory consensus building with all stakeholders. The decision-making process we’ve adopted is participatory and based on 360-degree feedback from all key stakeholders. The intervention plan for a month is developed during monthly meetings of the intervention team with feedback incorporated from the process evaluation sub-team prior to seeking feedback from our intended program participants and local stakeholders. Next, the village-specific intervention plans are shared with women of reproductive age in our intervention villages, members of local influential groups, policy makers (sarpanch, ward members etc.) and frontline workers before finalization. Once the feedback loops are implemented, RANI community facilitators solicit comments from the participants on the session content and facilitation quality. Comments are also sought by supervisors and the intervention management team during monitoring visits via informal opinion polls. The quantitative ratings are recorded in RPM4K’s mobile-based forms, whereas qualitative feedback is compiled in monthly qualitative reports shared with all stakeholders. Field-level findings, intervention achievements and qualitative findings are shared with the process evaluation team on a weekly basis. These discussions, along with interaction with the field teams, is used to tweak RANI’s intervention delivery strategy on a monthly basis.

Monthly coordination and advocacy meetings are also conducted with block and district level government counterparts from the Departments of Health, Women and Child and Odisha Livelihood Mission. Village-wise hemoglobin status along with demand for IFA tablets is discussed during these interactions. The district collector (local government leadership) holds quarterly review meetings with all officials from the local departments involved with RANI. Through these interactions, local government officials also provide feedback and suggestions for improving the coordination and convergence of the RANI project with other ongoing anemia and health improvement programs in the district.

Survey data triangulation and analysis

Beyond using feedback loops to inform RANI intervention tweaks, we plan to triangulate RANI’s endline survey data with our monitoring data to assess whether intervention delivery factors (such as village-level reach, community facilitator performance over time, testing coverage, implementation coverage, IFA supply and demand) influence intervention exposure levels at the village-level. We will also explore whether and how 1) these intervention delivery factors influence psychosocial (e.g., self-efficacy, knowledge, awareness, risk perception), normative or behavioral (e.g., IFA use and adherence) outcomes at the village-level controlling for village-level demographics (i.e., age, income, education, caste) and structural factors (e.g., proximity to supply points). Finally, we will assess whether intervention exposure at the village-level mediates the relationship between village-level intervention delivery factors and psychosocial, normative, or behavioral outcomes.

Discussion

The purpose of this paper is to shed light on the RPM4K, a process evaluation and monitoring system being implemented by the RANI Project (which has now completed its implementation delivery with final impact evaluation underway). It is built on the idea that even the best designed interventions are not able to foresee all changes that occur as project activities are rolled out. The overall environment is likely to change (as was the case in this study with the advent of coronavirus disease 2019 (COVID-19) that required significant delays and adaptations) as are the social and political realities on the ground. Furthermore, changes brought about by the intervention can themselves precipitate other disturbances in the system – as was the case when our campaign to generate demand for IFA resulted in shortages in supply.

Given these realities, and many others that cannot be anticipated at the beginning of the intervention, it is critical to build feedback loops into the overall implementation plan so that changes can be incorporated on an ongoing basis. This requires at least three criteria to be met. First, the different sequential steps in data processing (collecting, analyzing, and reporting) have to occur with minimal delays. In the RANI Project, this meant data collection in real (or near to real) time, automated algorithms built into the analysis step, and a designated person for analysis and reporting on a predetermined timeline.

Second, adaptations made to the intervention rollout plans because of monitoring feedback have to be visible to the team. When data collectors and others involved in data processing come to view their efforts as being inconsequential (because adaptations are not being made to the intervention based on feedback received), few incentives remain for them to continue their work with full commitment. Conversely, the ability to see changes in the intervention as a result of one’s data collection, processing, or reporting activity can serve as powerful motivators to continue and even improve one’s performance. The RANI project held monthly meetings with facilitators, supervisors, and RANI’s state implementation manager during which time changes made to the implementation schedule and activities were shared with the team. More so, we’ve engaged in community and policy-level dissemination events on a monthly basis informing our stakeholders and intervention recipients of changes in social norms, anemia as well as iron folic acid supply, demand, and uptake within their localities. Disseminating the knowledge we’ve gathered from our monitoring and survey data has been an integral part of feedback loops built into the RANI intervention delivery strategy. Beyond our local stakeholders and intervention recipients, we’ve engaged in dissemination via regional, national and international conferences, webinars, academic journals, social media platforms, and media outlets.

Finally, it is also important that the overall team buy into a culture of innovation. This is likely the most difficult task, particularly for projects perceived as being conceptualized and implemented externally, with minimal local stakeholder input, and ones thought to be driven solely by outcomes, without regard for underlying processes. The idea that a project is open to innovations means that failures are anticipated, allowed to happen without retribution, and framed in terms of learning opportunities for approaching the task in a way different from how it was originally done. The process evaluation protocol for the RANI project, detailed in this paper, provides a blueprint for other programs or trials aiming to harmonize responsive feedback loops, (near to) real-time data, and adaptive tailoring variables to enhance their intervention delivery and overall impact.

Conclusion

The process of incorporating of feedback loops paired with digital intervention tools which offer near-to-real-time intervention data remains a gap in literature despite calls from leading experts in the field. This protocol bridges this gap and contributes methods, tools, templates, instruments, processes, frameworks, and a publicly available software that facilitate the incorporation of feedback loops to innovate the intervention delivery of the RANI project.

Study status

The RANI study was implemented between September 2018 and March 2021.

List of abbreviations

IFA- Iron and Folic Acid

RANI- Reduction in Anemia Through Normative Innovations

USAID- The United States Agency for International Development

RCTs- Randomized Control Trials

M&E- Monitoring and Evaluation

RPM4K- Real-time Performance Monitoring for Knowledge

Acknowledgements

An earlier version of this article can be found on Research Square (DOI: 10.21203/rs.3.rs-389608/v1)

Funding Statement

The author(s) declared that no grants were involved in supporting this work.

[version 2; peer review: 2 approved]

Data availability

Underlying data

Figshare. The RANI Project Process Monitoring and Evaluation Dataset and Codebook. DOI: https://doi.org/10.6084/m9.figshare.16709452.v1 49 .

This project contains the following underlying data:

  • RANI Project Process Monitoring and Evaluation Dataset

  • RANI Project Process Monitoring and Evaluation Data Dictionary

Extended data

Figshare. RPM4K Supplementary Files. DOI: 10.6084/m9.figshare.20186837

This project contains the following extended data:

  • Appendix 1: Sample Template for RANI’s Monitoring and Evaluation Framework

  • Appendix 2: Qualitative Interview Guide

  • Appendix 3: RANI M&E Sample Data Collection Form

  • Appendix 4: RPM4K Dashboard and Data Visualization Tools

  • Appendix 5: Feedback Loops Documentation Tool

  • Appendix 6: RPM4K Data Quality Protocol

Data are available under the terms of the Creative Commons Zero "No rights reserved" data waiver (CC BY 4.0 Public domain dedication).

Software availability

Zenodo. ipant/RANI: RANI DOI. DOI: https://doi.org/10.5281/zenodo.5062358 50

Data are available under the terms of the Creative Commons Zero "No rights reserved" data waiver (CC BY 4.0 Public domain dedication).

References

  • 1. Lopez A, Cacoub P, Macdougall IC, et al. : Iron deficiency anaemia. Lancet. 2016;387(10021):907–916. 10.1016/S0140-6736(15)60865-0 [DOI] [PubMed] [Google Scholar]
  • 2. Horton S, Ross J: The economics of iron deficiency. Food Policy. 2003;28(1):51–75. 10.1016/S0306-9192(02)00070-2 [DOI] [Google Scholar]
  • 3. International Institute for Population Sciences (IIPS) and ICF: National Family Health Survey (NFHS-4), India, 2015-16: Odisha.Mumbai: IIPS,2017. Reference Source [Google Scholar]
  • 4. Stevens GA, Finucane MM, De-Regil LM, et al. : Global, regional, and national trends in haemoglobin concentration and prevalence of total and severe anaemia in children and pregnant and non-pregnant women for 1995-2011: a systematic analysis of population-representative data. Lancet Glob Health. 2013;1(1):e16–25. 10.1016/S2214-109X(13)70001-9 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5. Ministry of Health and Family Welfare Government of India – Adolescent Division: Guidelines for Control of Iron Deficiency Anaemia. National Iron+ Initiative.2013. Reference Source [Google Scholar]
  • 6. Aguayo VM, Paintal K, Singh G: The Adolescent Girls' Anaemia Control Programme: a decade of programming experience to break the inter-generational cycle of malnutrition in India. Public Health Nutr. 2013;16(9):1667–76. 10.1017/S1368980012005587 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7. Sreedevi A: An overview of the development and status of national nutritional programs in India. J Med Nutr Nutraceut. 2015;4(1):5–13. 10.4103/2278-019X.146146 [DOI] [Google Scholar]
  • 8. Kumar A: National nutritional anaemia control programme in India. Indian J Public Health. 1999;43(1):3–5, 16. [PubMed] [Google Scholar]
  • 9. Sedlander E, Long MW, Mohanty S, et al. : Moving beyond individual barriers and identifying multi-level strategies to reduce anemia in Odisha India. BMC Public Health. 2020;20(1):457. 10.1186/s12889-020-08574-z [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10. Kadiyala S, Morgan EH, Cyriac S, et al. : Adapting Agriculture Platforms for Nutrition: A Case Study of a Participatory, Video-Based Agricultural Extension Platform in India. PLoS One. 2016;11(10):e0164002. 10.1371/journal.pone.0164002 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11. Noronha JA, Bhaduri A, Bhat HV, et al. : Interventional study to strengthen the health promoting behaviours of pregnant women to prevent anaemia in southern India. Midwifery. 2013;29(7):e35–e41. 10.1016/j.midw.2012.07.014 [DOI] [PubMed] [Google Scholar]
  • 12. Shet AS, Zwarenstein M, Mascarenhas M, et al. : The Karnataka Anemia Project 2--design and evaluation of a community-based parental intervention to improve childhood anemia cure rates: study protocol for a cluster randomized controlled trial. Trials. 2015;16(1):599. 10.1186/s13063-015-1135-x [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13. Bentley P, Parekh A: Perceptions of Anemia and Health Seeking Behavior among Women in Four Indian States.Technical Working Paper #9. Mother Care/ John Snow Inc.1998. Reference Source [Google Scholar]
  • 14. de Benoist B, McLean E, Egli I, et al. : Worldwide prevalence of anemia 1993-2005.WHO Global Database on Anemia Geneva, World Health Organization,2008. Reference Source [Google Scholar]
  • 15. Nguyen PH, Sanghvi T, Kim SS, et al. : Factors influencing maternal nutrition practices in a large scale maternal, newborn and child health program in Bangladesh. PLoS One. 2017;12(7):e0179873. 10.1371/journal.pone.0179873 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16. Dongre AR, Deshmukh PR, Garg BS: Community-led initiative for control of anemia among children 6 to 35 months of age and unmarried adolescent girls in rural Wardha, India. Food Nutr Bull. 2011;32(4):315–323. 10.1177/156482651103200402 [DOI] [PubMed] [Google Scholar]
  • 17. Sedlander E, Talegawkar S, Ganjoo R, et al. : How Gender Norms affect Anemia in select villages in Rural Odisha, India: a Qualitative Study. Nutrition. 2021;86:111159. 10.1016/j.nut.2021.111159 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18. Yilma H, Sedlander E, Rimal RN, et al. : Is fatigue a cue to obtain iron supplements in Odisha, India? A mixed methods investigation. BMJ Open. 2020;10(10):e037471. 10.1136/bmjopen-2020-037471 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19. Yilma H, Sedlander E, Rimal RN, et al. : The Reduction in Anemia through Normative Innovations (RANI) Project: Study Protocol for a Cluster Randomized Controlled Trial in Odisha, India. BMC Public Health. 2020;20(1):203. 10.1186/s12889-020-8271-2 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 20. Patton MQ: Developmental Evaluation. Applying Complexity Concepts to Enhance Innovation and Use.Guilford Press, New York,2010. Reference Source [Google Scholar]
  • 21. Urquhart R, Kendell C, Geldenhuys L, et al. : The role of scientific evidence in decisions to adopt complex innovations in cancer care settings: a multiple case study in Nova Scotia, Canada. Implement Sci. 2019;14(1):14. 10.1186/s13012-019-0859-5 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22. Liu H, Muhunthan J, Hayek A, et al. : Examining the use of process evaluations of randomised controlled trials of complex interventions addressing chronic disease in primary health care-a systematic review protocol. Syst Rev. 2016;5(1):138. 10.1186/s13643-016-0314-5 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 23. Oakley A, Strange V, Bonell C, et al. : Process evaluation in randomised controlled trials of complex interventions. BMJ. 2006;332(7538):413–416. 10.1136/bmj.332.7538.413 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 24. Scheirer MA, Dearing JW: An agenda for research on the sustainability of public health programs. Am J Public Health. 2011;101(11):2059–2067. 10.2105/AJPH.2011.300193 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25. Craig P, Dieppe P, Macintyre S, et al. : Developing and evaluating complex interventions: the new Medical Research Council guidance. BMJ. 2008;337:a1655. 10.1136/bmj.a1655 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 26. Acevedo GL, Rivera K, Lima L, et al. : Challenges in Monitoring and Evaluation: An Opportunity to Institutionalize M&E Systems. Fifth Conference of The Latin American and the Caribbean Monitoring and Evaluation (M&E) Network. Reference Source [Google Scholar]
  • 27. Burdescu R, del Villar A, Mackay K, et al. : Institutionalizing Monitoring and Evaluation Systems: Five Experiences from Latin America. en breve; No. 78. World Bank, Washington, DC. © World Bank.2005. Reference Source [Google Scholar]
  • 28. USAID: Scientific Research Policy.2014;5. Reference Source [Google Scholar]
  • 29. Viswanath K, Synowiec C, Agha S: Responsive feedback: Towards a new paradigm to enhance intervention effectiveness [version 2; peer review: 4 approved]. Gates Open Res. 2019;3:781. 10.12688/gatesopenres.12937.2 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 30. Glöss M, McGregor M, Brown B: Designing for labour: uber and the on-demand mobile workforce. In: Proceedings of the 2016 CHI conference on human factors in computing systems. ACM.2016;1632–1643. 10.1145/2858036.2858476 [DOI] [Google Scholar]
  • 31. Kollock P: The production of trust in online markets.1999;16(1):99–123. Reference Source [Google Scholar]
  • 32. Ma X, Hancock JT, Mingjie KL, et al. : Self-Disclosure and Perceived Trustworthiness of Airbnb Host Profiles. In: CSCW.2017;2397–2409. 10.1145/2998181.2998269 [DOI] [Google Scholar]
  • 33. Melnik MI, Alm J: Does a seller’s ecommerce reputation matter? Evidence from eBay auctions. J Ind Econ. 2002;50(3):337–349. Reference Source [Google Scholar]
  • 34. Whidden C, Kayentao K, Liu JX, et al. : Improving Community Health Worker Performance by Using a Personalised Feedback Dashboard for Supervision: A Randomised Controlled Trial. J Glob Health. 2018;8(2):020418. 10.7189/jogh.08.020418 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 35. Whittle D, Campbell M: A Guide to Digital Feedback Loops: An Approach to Strengthening Program Outcomes through Data for Decision Making. Reference Source 2019. [Google Scholar]
  • 36. Schaefer M: USAID. Interview.2018. Reference Source [Google Scholar]
  • 37. UNICEF: U-Report Nigeria.(n.d.). Retrieved October 12,2018. Reference Source [Google Scholar]
  • 38. USAID: Collaborating, Learning, and Adapting: An Analysis of what CLA Looks Like in Development Programming.2017. Reference Source [Google Scholar]
  • 39. McManus L: IntraHealth International.2018. [Google Scholar]
  • 40. McDonough M: Souktel.2018. [Google Scholar]
  • 41. DeRenzi B, Dell N, Wacksman J, et al. : Supporting community health workers in India through voice- and web-based feedback.In: Proceedings of the 2017 CHI Conference on Human Factors in Computing Systems. ACM,2017;2770–2781. 10.1145/3025453.3025514 [DOI] [Google Scholar]
  • 42. Guijt I: Participatory Approaches, Methodological Briefs: Impact Evaluation 5.UNICEF Office of Research, Florence,2014. Reference Source [Google Scholar]
  • 43. Lennie J, Tacchi J, Koirala B, et al. : Equal Access Participatory Monitoring and Evaluation toolkit.2011. Reference Source [Google Scholar]
  • 44. Okeke F, Nene L, Muthee A, et al. : Opportunities and Challenges in Connecting Care Recipients to the Community Health Feedback Loop.In: Proceedings of The Tenth International Conference on Information and Communication Technologies and Development. 2019;1–11. 10.1145/3287098.3287111 [DOI] [Google Scholar]
  • 45. Barron P, Pillay Y, Fernandes A, et al. : The MomConnect mHealth initiative in South Africa: Early impact on the supply side of MCH services. J Public Health Policy. 2016;37(Suppl 2):201–212. 10.1057/s41271-016-0015-2 [DOI] [PubMed] [Google Scholar]
  • 46. Holeman I, Cookson TP, Pagliari C: Digital technology for health sector governance in low and middle income countries: A scoping review. J Glob Health. 2016;6(2):020408. 10.7189/jogh.06.020408 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 47. Ramalingam B, Barnett I, Levy A, et al. : Bridging the Gap: How Real-Time Data Can Contribute to Adaptive Management in International Development.USAID,2018. [Google Scholar]
  • 48. Lei H, Nahum-Shani I, Lynch K, et al. : A "SMART" design for building individualized treatment sequences. Annu Rev Clin Psychol. 2012;8:21–48. 10.1146/annurev-clinpsy-032511-143152 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 49. Yilma H, Pant I: The RANI Project Process Monitoring and Evaluation Dataset and Codebook. figshare. Dataset.2021. 10.6084/m9.figshare.16709452.v1 [DOI]
  • 50. ipant: ipant/RANI: RANI DOI (Open). Zenodo. 2021. 10.5281/zenodo.5062358 [DOI] [Google Scholar]
Gates Open Res. 2022 Sep 30. doi: 10.21956/gatesopenres.15310.r32614

Reviewer response for version 2

Ankit Anand 1

The problem here is that the all of the intervention is already published and the trial have also been approved. This paper is an extension on the monitoring part of the intervention. It gives a very little scope to review this as an original contribution. Additionally, it will be helpful to read with the actual intervention.

In that case, I would recommend indexing.

Is the study design appropriate for the research question?

Yes

Is the rationale for, and objectives of, the study clearly described?

Partly

Are sufficient details of the methods provided to allow replication by others?

No

Are the datasets clearly presented in a useable and accessible format?

Yes

Reviewer Expertise:

Maternal and Child Health, Adolescent Health, mHealth, Health System, Evaluation Framework

I confirm that I have read this submission and believe that I have an appropriate level of expertise to confirm that it is of an acceptable scientific standard.

Gates Open Res. 2022 May 27. doi: 10.21956/gatesopenres.14672.r32055

Reviewer response for version 1

Ankit Anand 1

The study describes the RANI project extensively. The protocol provides good detail and indicators for monitoring and evaluating the project. However, I feel it needs a lot of improvement in its current form

  • The authors need to include details about the geography and the motivation for this project. The process of arriving at this intervention is somewhat mentioned in different sections but needs to be more clear in the introduction part.

  • The protocol states using an innovative digital tool as an intervention. I would think the videos are innovative tools. However, the program focuses on multiple actors/aspects of the communities and even intervenes on the demand side of the issues. I think it is a quite comprehensive intervention but could not understand the innovation part of it.

  • How will the innovation be linked to improvement in outcome, for example in measuring dosage? Almost all of the activities (sessions, video demonstration sessions, haemoglobin testing sessions, and results demonstrations) are combined and considered as dose. It will be impossible to point out which activities are more effective, or which need more improvements.

  • The same thing for receptivity, I could not understand how likes versus dislikes will measure receptivity? More nuances of receptivity need to be there.

  • The intervention, especially the videos seem to be open-ended intervention which could be called innovation? Feedback loops are usually part of most digital intervention strategies.

  • However, I really appreciate the comprehensiveness of the approaches which may lead to a better outcome but evaluating activities and attributing effectiveness to all these activities will be a challenge. It might be related to more efficiency in their fieldwork/community engagements activity rather than the innovation. Which one (algorithms or procedures) leads to changes?

  • The replicability of the intervention will also be difficult.

  • More clarity needed on sources of data. Who will be interviewed? What information will be collected from each set of respondents and how it will correspond to the aims of the project? It is somewhat mentioned in table 2 but not able to know how it will be measured.

  • Too many evaluation questions? For example: "level of exposure to our media products" is an intermediate or process outcome, it can not be a final outcome.

  • A diagrammatic version of the theory of change might help in understanding.

  • The protocol mentioned has three very different aims. What are the areas the authors see as a problem? Whether it is a supply-chain, human resources, community engagements or demand-side issue? That is why I feel a paragraph stating the geography and their challenges is needed. It also needs to be linked with the motivation of the intervention. The first part of the introduction is very generic and does not link to the second part. The authors needs to provide specific details not generic statements.

  • A clear timeline for the intervention and process of feedback loops must also be clearly mentioned. If not then the intervention becomes non-replicable.

Is the study design appropriate for the research question?

Yes

Is the rationale for, and objectives of, the study clearly described?

Partly

Are sufficient details of the methods provided to allow replication by others?

No

Are the datasets clearly presented in a useable and accessible format?

Yes

Reviewer Expertise:

Maternal and Child Health, Adolescent Health, mHealth, Health System, Evaluation Framework

I confirm that I have read this submission and believe that I have an appropriate level of expertise to confirm that it is of an acceptable scientific standard, however I have significant reservations, as outlined above.

Gates Open Res. 2022 Sep 18.
Ichhya Pant 1

Thank you for reviewing our manuscript and offering your feedback. We offer the following point by point response to address the concerns raised by you in your review.

It may be that the reviewer is asking about the motivation for the intervention, given the geographical area where the work is done. We note, however, that this paper is about the monitoring of the intervention and not the intervention itself (for which we have already published quite a bit, including provide a link to Yilma et al. and Sedlander et al.)  But we have taken the reviewer’s comment seriously and described the motivations for adopting a monitoring approach.

Our motivations for adopting a process monitoring an evaluation approach is now highlighted under the Motivation for incorporating feedback loops section within the revised manuscript.

This comment, together with the comment above, suggests to us that we did not do a good job in making the distinction between two things: a protocol paper about the intervention (which this paper is NOT about, and one which the reviewer seems to be focusing on) versus a protocol paper about MONITORING the intervention. We hope that the new language we have adopted in this round reflects the difference more accurately.

To clarify, we do not state that we offer RPM4K (our digital monitoring tool) as an intervention component, rather, we use the tool to monitor the delivery of our intervention components. The innovation we offer is the incorporation of feedback loops with near-to-real-time data and we highlight and detail the contributions of our approach in the recently added Conclusion section among other sections throughout the manuscript.

This comment, too, focuses on the intervention, rather than on its monitoring. We propose that the innovation here is in how real-time monitoring was conceptualized, implemented, and used to provide feedback to the intervention. Indeed, to the reviewer’s point, we do monitor each intervention component separately and the monitoring system is explicitly designed to be able to report back on each individual component. The reviewer makes an important point about how we know which component had what kind of effect; indeed, we have just published such a paper:

Sedlander, E., Pant, I., Bingenheimer, J., Paltro, L., Yilma, H., Ganjoo, R., Mohanty, S., Rimal, R. (2022). How does a social norms-based intervention affect behavior change? Interim findings from a cluster randomized controlled trial in Odisha, India. BMJ Open.

The reviewer is correct that these measures are superficial. What they offer, however, are real time quick measures. They are NOT the whole story, of course, but they minimize participant burden and provide the project a snapshot about what people like and what they dislike. Our receptivity (now renamed as acceptability) measure was limited to likes versus dislikes to reduce participation burden for our intervention participants and frontline workers alike (both groups also reported having limited digital and educational literacy). However, we also monitored receptivity (which has now been renamed acceptability) qualitatively on a regular basis. As for replicability of the intervention, our overall study protocol is pre-registered and offers clarity on the motivations, goals, aims, measures, and summative outcomes such that replicability of the intervention will be entirely possible using the details provided in the RCT protocol trial.

Thank you for this observation and suggestion. Details regarding who collects and submits data and how these insights will inform identification of feedback loops for our tailoring variables directly related to achieving the aims of the RANI project are now detailed under the data collection section with specific sub-headings for each process evaluation measure and the identification and incorporation of feedback loops. Involvement of key actors is also outlined within these section and sub-sections. Instruments used to collect the data as well as other clarifying tools and processes are also included in the Extended data section. Finally, identifying relevant research questions, aims, and the entire evaluation framework was a collaborative effort involving local stakeholders, project team members from Odisha, and from George Washington University therefore we stand by our research questions and aims in their current form. This process is detailed in the section Process evaluations aims and questions section. The M&E framework included within the Extended Data or Supplementary Materials section brings together all of these details under one table for readers and reviewers who are curious to delve deeper into these details unavailable in the main manuscript.

Since this is a process evaluation and monitoring protocol that aims to innovate intervention delivery in near-to-real-time with the identification and integration of feedback loops, there isn’t a theory of change we’re able to include within this protocol. The entire premise of this process evaluation approach is to “build the ship as it was sailing .” We provide our rationale and existing calls to action to adopt such a method by leading experts in the field in the Use of Feedback Loops section.

The three distinct aims listed in our process evaluation protocol are centered around the intervention components designed for the RANI trial following a rigorous mixed-methods formative evaluation which included in-depth interviews, focus group interviews, stakeholder input sessions, and a quantitative survey. RANI team members and stakeholders identified all three areas addressed with the process evaluation aims as integral to ensuring our trial intervention delivery can proceed seamlessly. Our protocol includes specific details regarding how the feedback loops will be identified and delivered within the Use of Feedback Loops section. Table 3 provides hypothetical examples that visualize the process described under the Identifying, documenting and integrating feedback loops section. Furthermore, the feedback loops documentation template is included in the extended data section as well. All of these details in their entirety provide a detailed picture on how to apply, identifying and integrate feedback loops into process monitoring and evaluation efforts for randomized control field trials similar to the RANI project. This process evaluation protocol is centered on the use of near-to-real-time data to foster adaptive intervention delivery by integrating feedback loops during the intervention delivery process. We should also note that, in the interest of making this manuscript readable and concise, we did not repeat details provided in-depth in the RANI RANI RCT trial protocol that is described elsewhere and cited within this manuscript. We have added brief details to the RANI project description section to address all requested detailed intervention information and we’ve encouraged our readers to reference the RANI trial protocol for any and all details typically included in the trial protocol.

Gates Open Res. 2022 Apr 26. doi: 10.21956/gatesopenres.14672.r31875

Reviewer response for version 1

Claudia Abreu Lopes 1

This article is a protocol for a process evaluation with feedback loops using an innovative digital tool for an intervention (RANI project) aimed at lower rates of anemia among women of reproductive age in selected villages in India. The article is relatively well structured and the research denotes thoughtfulness and rigour. The research questions are clearly stated and the methods are adequate to answer the research questions. However, the article would benefit from improvements in four main aspects:

  1. The study uses a multi-method approach that consisted of collecting information from surveys, qualitative reports and structured observations. The description of the method as mixed methods is not entirely correct - multi-method would be preferable - since the methods or data sources were not triangulated or used in sequence. As it reads, different methods/data sources were used to measure different indicators/KPIs.

  2. Research aims refer to implementation research outcomes, such as fidelity. The second aim is about the acceptability of the intervention, referred to in this study as 'receptivity'. The KPI refers to satisfaction which is a related but different concept. The way the data is gathered with likes and dislikes and the way the indicator is calculated are a 'black box'. For example, it says: 'Receptivity is summarized at the village-level by algorithms embedded within RPM4K'. The authors should review the concept of receptivity (acceptability would be preferable) and explain how the data is gathered and the KPIs calculated (providing/explaining the algorithm).

  3. Table 1 presents 2 process evaluation questions but there are 3 process evaluation aims. The correspondence between aims and questions is not clear. Table 2 is also confusing. There may be a third measure missing as it jumps from 2 to 4 (please check other typos in Table 2 also). Table 1 can be improved by linking aims, research questions and measures. Table 2 would complement Table 1 by linking the KPIs and ecological levels to the measures. There is also not enough information to understand how the KPIs are measured at different ecological levels. For example 'why the number of women participating in hemoglobin testing is an interpersonal indicator?' Please add more information to the description of KPIs - method, data sources and calculation/formula for indicators.

  4. Please add a short conclusion.

Finally, the background sentence in the abstract needs to be improved (proliferation of digital age or proliferation of digital technologies?)

Is the study design appropriate for the research question?

Yes

Is the rationale for, and objectives of, the study clearly described?

Yes

Are sufficient details of the methods provided to allow replication by others?

Yes

Are the datasets clearly presented in a useable and accessible format?

Yes

Reviewer Expertise:

Implementation research; Digital health; Mental health; Research methods; Gender and health

I confirm that I have read this submission and believe that I have an appropriate level of expertise to confirm that it is of an acceptable scientific standard.

Gates Open Res. 2022 Sep 18.
Ichhya Pant 1

We would like to start by thank you for reviewing this manuscript. To respond to your feedback, we offer the following point by point response to address the concerns raised by Reviewer #1.

Thank you for making that distinction, which is valid. Based on that distinction, however, we do indeed triangulate information collected via qualitative and quantitative surveys on a monthly basis and well as with our mid-line survey. Thus, our description of our methodology as mixed methods is intentional. This process of information triangulation in order to identify and integrate feedback loops are detailed under the sections – 1) identifying, documenting, and implementing feedback loops 2) survey data triangulation.

We’ve added brief text to make this more explicit for our reviewers and readers.

The KPI formerly listed as receptivity has been renamed as acceptability to address your concerns.

We’ve also incorporated revisions to clarify that the algorithm simply summarizes the total count of likes versus dislikes for intervention components by village. Additional File 1 includes details for how each measure is operationalized and aggregated by RPM4K. Furthermore, we operationalize and define all measures at an aggregate level within the Data Collection section as well. Last but not the least, the software code for RPM4K is also publicly available in a Github repository which we’ve shared in the Extended data section. Auditing and examining how the algorithms are set up within RPM4K is possible by examining the source code as well. Our choice to make the software available publicly is supported by our funders and speaks to our intentions to veer away from black box calculations of KPIs.

We direct the reviewer to the two evaluation questions and sub-questions listed underneath them that address all three aims. Our evaluation framework included in the Extended Data section connects Table 1 and 2 per your suggestion. In the interest of space, word count, and formatting, we aren’t able to integrate Table 1 and 2 per your suggestion here. However, the framework included in the Supplementary Material ties together the aims, questions, indicators, measures, data sources, formulas, responsible parties, operational definitions and ecological levels.

We have added a short conclusion section as well as reframed the background sentence in the abstract per your recommendation.

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Data Citations

    1. Yilma H, Pant I: The RANI Project Process Monitoring and Evaluation Dataset and Codebook. figshare. Dataset.2021. 10.6084/m9.figshare.16709452.v1 [DOI]

    Data Availability Statement

    Underlying data

    Figshare. The RANI Project Process Monitoring and Evaluation Dataset and Codebook. DOI: https://doi.org/10.6084/m9.figshare.16709452.v1 49 .

    This project contains the following underlying data:

    • RANI Project Process Monitoring and Evaluation Dataset

    • RANI Project Process Monitoring and Evaluation Data Dictionary

    Extended data

    Figshare. RPM4K Supplementary Files. DOI: 10.6084/m9.figshare.20186837

    This project contains the following extended data:

    • Appendix 1: Sample Template for RANI’s Monitoring and Evaluation Framework

    • Appendix 2: Qualitative Interview Guide

    • Appendix 3: RANI M&E Sample Data Collection Form

    • Appendix 4: RPM4K Dashboard and Data Visualization Tools

    • Appendix 5: Feedback Loops Documentation Tool

    • Appendix 6: RPM4K Data Quality Protocol

    Data are available under the terms of the Creative Commons Zero "No rights reserved" data waiver (CC BY 4.0 Public domain dedication).


    Articles from Gates Open Research are provided here courtesy of Bill & Melinda Gates Foundation

    RESOURCES