Skip to main content
PLOS ONE logoLink to PLOS ONE
. 2020 Aug 14;15(8):e0237703. doi: 10.1371/journal.pone.0237703

Effect of data quality improvement intervention on health management information system data accuracy: An interrupted time series analysis

Zewdie Mulissa 1,*, Naod Wendrad 2, Befikadu Bitewulign 1, Abera Biadgo 1, Mehiret Abate 1, Haregeweyni Alemu 1, Biruk Abate 3, Abiyou Kiflie 1, Hema Magge 1,4,5, Gareth Parry 6,7
Editor: Russell Kabir8
PMCID: PMC7428163  PMID: 32797091

Abstract

Background

As part of a partnership between the Institute for Healthcare Improvement and the Ethiopian Federal Ministry of Health, woreda-based quality improvement collaboratives took place between November 2016 and December 2017 aiming to accelerate reduction of maternal and neonatal mortality in Lemu Bilbilu, Tanqua Abergele and Duguna Fango woredas. Before starting the collaboratives, assessments found inaccuracies in core measures obtained from Health Management Information System reports.

Methods and results

Building on the quality improvement collaborative design, data quality improvement activities were added and we used the World Health Organization review methodology to drive a verification factor for the core measures of number of pregnant women that received their first antenatal care visit, number of pregnant women that received antenatal care on at least four visits, number of pregnant women tested for syphilis and number of births attended by skilled health personnel. Impact of the data quality improvement was assessed using interrupted time series analysis. We found accurate data across all time periods for Tanqua Abergele. In Lemu Bilbilu and Duguna Fango, data quality improved for all core metrics over time. In Duguna Fango, the verification factor for number of pregnant women that received their first antenatal care visit improved from 0.794 (95%CI 0.753, 0.836; p<0.001) pre-intervention by 0.173 (95%CI 0.128, 0.219; p<0.001) during the collaborative; and the verification factor for number of pregnant women tested for syphilis improved from 0.472 (95%CI 0.390, 0.554; p<0.001) pre-intervention by 0.460 (95%CI 0.369, 0.552; p<0.001) during the collaborative. In Lemu Bilbilu, the verification factor for number of pregnant women receiving a fourth antenatal visit rose from 0.589 (95%CI 0.513, 0.664; p<0.001) at baseline by 0.358 (95%CI 0.258, 0.458; p<0.001) post-intervention; and skilled birth attendance rose from 0.917 (95%CI 0.869, 0.965) at baseline by 0.083 (95%CI 0.030, 0.136; p<0.001) during the collaborative.

Conclusions

A Data quality improvement initiative embedded within woreda clinical improvement collaborative improved accuracy of data used to monitor maternal and newborn health services in Ethiopia.

Introduction

Since October 2013, the Institute for Healthcare Improvement (IHI) has worked in partnership with the Ethiopian Federal Ministry of Health (FMoH), with the support of the Bill and Melinda Gates Foundation and Margaret A. Cargill Philanthropies to explore how quality improvement (QI) methodologies can accelerate progress of the FMoH to improve maternal and neonatal health in Ethiopia. As part of this work, woreda (woreda) based focused improvement collaboratives, based on the IHI Breakthrough Series collaborative [1] commenced in November 2016 with the aim of accelerating the progress in reducing maternal and neonatal mortality in three woredas. A key principle of the collaborative methodology is that participating teams submit and share data with each other, and, in particular, use a common core set of measures to understand changes in processes and outcomes over time. In the improvement collaboratives, a number of core process and outcome measures related to maternal and neonatal outcomes were taken from the national health management information system (HMIS) in Ethiopia. Additionally, the HMIS data was intended to be used as the basis for an overall summative impact evaluation of the improvement collaborative approach. However, numerous studies have reported problems with the quality of HMIS data, and many users do not trust these data [24].

Recent studies in Nigeria [5] and South Africa [6] reported improvement in data quality following quality improvement intervention as measured using a World Health Organization (WHO) data quality review methodology [3]. This WHO methodology can be employed from a facility level to a national level and provides an indication of the completeness of reporting and a verification factor to indicate data accuracy for specific reporting periods.

Prior to the start of the improvement collaboratives, HMIS data accuracy was assessed in the period May to October 2016, revealing inaccurate data in the core process and outcome metrics. We sought to avoid data falsification/intentional manipulation of data by focusing on improving data quality through training participants on the importance of high-quality data, monthly data review and feedback and trust building with health care leaders and workers. Consequently, the improvement collaboratives embedded a data quality improvement initiative aimed at improving the accuracy of maternal and neonatal health data from the HMIS system in participating sites. This paper describes the extent to which the accuracy of the HMIS data was improved during the improvement collaborative, as measured using the WHO review methodology.

Materials and methods

Improvement collaboratives

The Federal Ministry of Health (FMOH) of Ethiopia, regional health bureaus (RHBs) and IHI Ethiopia selected one woreda in each of the four most populous agrarian regions of Ethiopia to introduce the improvement collaborative approach. The woredas were purposefully selected in consultation with the Federal Ministry of Health (FMoH) of Ethiopia and regional health bureaus (RHBs) based on pre-set criteria, including high maternal and perinatal deaths, high level of leadership commitment to improve the service and the absence of other partner organizations working on quality improvement project. All facilities in each woreda were included into a collaborative. This study includes results from the first three collaboratives introduced simultaneously at twenty health facilities in: Lemu Bilbilu (8), Tanqua Abergele (6) and Duguna Fango (6) woredas of Oromia, Tigray and Southern Nations and Nationalities People’s region respectively.

The overall structure of the improvement collaboratives is summarized in Fig 1. In brief, the improvement collaboratives brought together teams from participating sites for 18 months to pursue a collective aim of improving maternal and child health care and outcomes across Ethiopia. Participating teams attended four 2 to 3-day learning sessions, where they came together to learn about the topic and to plan tests of change. During three Action Periods (time between learning sessions), teams were expected to use the Model for Improvement [7] to test changes using Plan-Do-Study-Act cycles in their local settings.

Fig 1. Summary of the core improvement activities and embedded data quality improvement activities.

Fig 1

Each month the improvement collaboratives measured implementation progress using a number of core measures, including four sourced from HMIS data described in Table 1. We embedded additional activities focusing participating sites on improving the quality of selected HMIS data associated with these specific measures. These activities were timed to coincide with the wider activities of the improvement collaboratives in Fig 1. We sought to avoid data falsification/intentional manipulation of data by focusing on improving data quality through training participants on the importance of high-quality data, monthly data review and feedback and trust building with health care leaders and workers. Training involved health care workers, health care leaders, the health care facility and woreda information officers who were responsible for data collection and was intended to help them understand that high-quality data are essential for improving service quality. Monthly data reviews and feedback were done by IHI Ethiopia senior project officers. Quarterly learning sessions and progress review meetings were used to reflect on progress of data quality among collaboratives and build trust with health care leaders and workers.

Table 1. HMIS-derived core measures used to track progress in the improvement collaboratives.

Core Measures Name Definition Data Source
Antenatal Care 1 Number of pregnant women who had at least one antenatal care visit during their pregnancy. National Health Management Information System (HMIS)
Antenatal Care 4 Number of pregnant women who had four or more antenatal care visits during their pregnancy. National Health Management Information System (HMIS)
Syphilis Screening Number of pregnant women tested for syphilis National Health Management Information System (HMIS)
Skilled Birth Number of births attended by skilled health personnel National Health Management Information System (HMIS)
Post-natal care 48 hours Number of women who attended post-natal care at least once within 48 hours after delivery. National Health Management Information System (HMIS)

Data accuracy measures

To understand how data accuracy changed from the pre-intervention period to intervention period to the post-Intervention period, we followed the WHO data review methodology [3] to create a measure of data accuracy for selected core measures as follows:

Original HMIS report value

Monthly, from May 2016 to December 2018, for each selected core measure, experienced IHI senior project officers collected data from the archive of HMIS reports at each facility.

Recounted (audited) value

Experienced IHI senior project officers undertook an audit by repeating the data collection, from May 2016 to December 2018, from standard antenatal care and delivery registers, developed by the Federal Ministry of Health (FMOH) of Ethiopia. This approach resulted in a monthly audited value for the selected core measures.

For each selected core measure, a monthly verification factor was calculated by dividing the recounted value by the original HMIS report value. A verification factor of 0.9–1.1 is considered “accurate”; <0.9 is considered over reported and a value > 1.1 as under reported [3].

Data management and analysis

We used Microsoft Excel 2016 for data entry and STATA V13 for analysis. Mean and standard deviation were used for descriptive analysis. Interrupted time series was used to assess change in data accuracy, as measured using the verification factor, from the pre-intervention or baseline (May to October 2016) to during (November 2016 to December 2017) and post-intervention (January to December 2018) phases. Specifically, the time-series model for each measure, assessed whether a change in verification factor had occurred across each phase by adding a shift and slope-change term for each phase. Statistical significance was set at P < 0.05.

Sample size

There are no fixed limits regarding the number of data points for interrupted time series study, as the power depends on various other factors including distribution of data points before and after the intervention, variability within the data, strength of effect and the presence of confounding effects such as seasonality [8]. For each facility, for each month, we used all data available from May 2016 to December 2018. We aggregated the data across facilities for each woreda, resulting in 32 data points for each woreda, 6 before and 26 after the start of the intervention.

Ethical considerations

This research is part of a broader evaluation study that was reviewed and approved by Ethiopian Public Health Association (EPHA) Scientific and Ethical Review Committee. A letter of support was obtained from IHI Ethiopia project office.

Results

General Health facilities information

All health facilities in the three improvement collaboratives (20); Lemu Bilbilu, Tanqua Abergele and Duguna Fango provide antenatal care services including the first to fourth visits, Syphilis screening and skilled birth attendance (institutional delivery). Monthly HMIS reports and registers were also available and complete at all the health facilities in the improvement collaboratives; 3 primary hospitals and 17 health centers. All health facilities also reported their performances monthly to their respective woreda health offices.

Data personnel, HMIS officers, were available at all hospitals and 13 (76.5%) health centers. Training on data quality was given to all HMIS officers at hospitals and 3 (17.6%) health center staffs including HMIS officers. Data quality was checked monthly by lot quality assurance sampling (LQAS) method at all the hospitals and 15 (88.2%) of the health centers. As HMIS officers and relevant staff participated in the improvement collaboratives, they all received training on data quality.

Verification factor

The mean & standard deviation of verification factors for core measure for each woreda over time period are summarized in Table 2. Post-natal care was frequently provided in other facilities, and within the scope of this study, it was not practical to undertake the data audit and the Post-natal care 48hours measure was dropped.

Table 2. Verification factor by woreda for the baseline, intervention and post-intervention periods.

Mean (standard deviation)
Baseline May 2016 to Oct 2016 Intervention Nov 2016 to Dec 2017 Post-Intervention Jan 2018 to Dec 2018
Antenatal Care 1
Lemu Bilbilu 0.842 (0.123) 1.012 (0.043) 0.897 (0.133)
Duguna Fango 0.794 (0.082) 0.972 (0.039) 0.963 (0.048)
Tanqua Abergele 1 (0) 0.983 (0.062) 0.971 (0.099)
Overall 0.817 (0.077) 0.991 (0.018) 0.923 (0.078)
Antenatal Care 4
Lemu Bilbilu 0.589 (0.131) 0.866 (0.171) 0.946 (0.147)
Duguna Fango 0.486 (0.173) 0.842 (0.148) 0.982 (0.052)
Tanqua Abergele 1 (0) 0.990 (0.063) 1.029 (0.069)
Overall 0.549 (0.116) 0.851 (0.142) 0.964 (0.081)
Syphilis Screening
Lemu Bilbilu 0.664 (0.442) 0.875 (0.235) 1.058 (0.114)
Duguna Fango 0.472 (0.141) 0.933 (0.089) 0.932 (0.101)
Tanqua Abergele 1.045 (0.070) 0.993 (0.014) 0.982 (0.059)
Overall 0.531 (0.195) 0.899 (0.154) 0.993 (0.067)
Skilled Birth Attendance
Lemu Bilbilu
  • 0.917 (0.031)

0.979 (0.049) 1.024 (0.072)
Duguna Fango 1.001 (0.040) 0.961 (0.047) 0.989 (0.019)
Tanqua Abergele 1 (0) 0.984 (0.078) 1.008 (0.039)
Overall 0.917 (0.031) 0.975 (0.032) 1.007 (0.032)

Interrupted Time Series (ITS) analysis

For each time-series analysis, applying the Durbin-Watson test, suggested there was no autocorrelation in for all the verification factor measures. Additionally, we found no seasonality in the data. The results of the time series analysis are shown in Table 3 and illustrated in Fig 2. For each measure and Woreda, we show the coefficients from the best fitting model.

Table 3. Results of the time series analysis for the verification factor for the core measures across woreda over the baseline, intervention and post-intervention phases.

Baseline May 2016 to Oct 2016 Intervention Nov 2016 to Dec 2017 Post-Intervention Jan 2018 to Dec 2018
Constant Slope Constant Slope Constant Slope
Coefficient (95% CI) P Coefficient (95% CI) P Coefficient (95% CI) P Coefficient (95% CI) P Coefficient (95% CI) P Coefficient (95% CI) P
Antenatal Care 1
Lemu Bilbilu 0.842 <0.001 - - 0.170 0.002 - - -0.116 0.01 - -
(0.761, 0.923) (0.073, 0.267) (-0.194, -0.038)
Duguna Fango 0.794 <0.001 - - 0.173 <0.001 - - - - - -
(0.753, 0.836) (0.128, 0.219)
Antenatal Care 4
Lemu Bilbilu 0.589 <0.001 - - - - 0.037 <0.001 0.358 <0.001 - -
(0.513, 0.664) (0.026, 0.048) (0.258, 0.458)
Duguna Fango 0.486 <0.001 - - 0.149 0.044 0.028 <0.001 0.349 <0.001 - -
(0.405, 0.567) (0.01, 0.287) (0.015, 0.041) (0.223, 0.474)
Skilled birth attendance
Lemu Bilbilu 0.917 <0.001 - - 0.083 0.004 - - - - - -
(0.869, 0.965) (0.03, 0.136)
Duguna Fango 1.008 <0.001 - - -0.034 0.064 - - - - - -
(0.977, 1.039) (-0.069, 0.001)
Syphilis Screening
Lemu Bilbilu 0.664 <0.001 - - 0.296 0.018 - - - - - -
(0.454, 0.873) (0.064, 0.528)
Duguna Fango 0.472 <0.001 - - 0.460 <0.001 - - - - - -
(0.39, 0.554) (0.369, 0.552)

CI = Confidence Interval, P = P-value.

Fig 2. Interrupted time series graph of change in verification factor from baseline to post-intervention.

Fig 2

The verification factor for all four measures averaged 1 for all three time periods at Tanqua Abergele. Thus, we conducted the time series analysis only in Lemu Bilbilu and Duguna Fango.

For Antenatal Care 1 (Table 2 and Fig 2A), the interrupted time series for the verification factor for Lemu Bilbilu increased from 0.842 during the baseline period by 0.170 (95% CI 0.073, 0.267; p = 0.002) during the collaborative period and then fell by 0.116 (95% CI 0.194, 0.038; p = 0.007) during the follow-up period. For Duguna Fango, the verification factor rose from 0.794 during the baseline period by 0.173 (95% CI 0.128,0.219; p<0.001) during the collaborative period and no change in the follow-up period.

For Antenatal Care 4 (Table 2 and Fig 2B), the verification factor for Lemu Bilbilu increased from 0.589 during the baseline period by 0.358 (95%CI 0.258, 0.458; p<0.001) during the follow-up period. For Duguna Fango, the verification factor rose from 0.486 at baseline by 0.149 (95%CI 0.01, 0.287; p = 0.04) during the collaborative period and then by 0.349 (95%CI 0.223, 0.474; p<0.001) during the follow up period.

For Syphilis Screening (Table 2 and Fig 2C), in Lemu Bilbilu, the verification factor rose from 0.664 at the baseline by 0.296 (95%CI 0.064, 0.528; p = 0.018) during the collaborative period and no change in the follow-up period. In Duguna Fango the verification factor rose 0.472 at the baseline by 0.460 (0.369, 0.552; p<0.001) and no change in the follow-up period.

For Skilled Birth Attendance (Table 3 and Fig 2D), in Lemu Bilbilu, the verification factor rose from 0.917 at the baseline by 0.083 (95%CI 0.030, 0.136; p = 0.004) during the collaborative period and no change in the follow-up period. The relatively higher baseline verification factor for Duguna Fango (1.008) hasn’t changed during collaborative and follow up periods. The overall finding is as shown in Fig 2 and Table 3.

Discussion

This data quality improvement initiative, embedded within a wider set of improvement collaboratives, significantly improved accuracy of antenatal care visits and syphilis screening data within these improvement collaboratives in Ethiopia. The finding is comparable to a similar embedded data quality improvement initiative in South Africa [6] where data accuracy of public health facilities participating in an improvement initiative improved from 37% to 65%.

Data quality improvement activities focusing on training, regular audits, monthly review and feedback have been successfully applied in other settings, without being embedded in a wider care improvement initiative. For example, in the United Republic of Tanzania [9] improved health information systems were associated with data use workshops actively engaging data users. In Kenya, on-site assessment and feedback was used to improve the accuracy of routine HIV health information [10]. In Peru, phone reminders to epidemiological surveillance teams and clinic visits were used to improve the timeliness and accuracy of data reported in an electronic surveillance system of infectious disease outbreaks. Phone reminders but not clinic visits improved timeliness, whereas in some settings data accuracy was improved by visits but not phone reminders [11]. Sequential data quality audits in six countries improved the quality and accuracy of data on immunization [12]. Elsewhere, improvement in data accuracy has been associated with increased use of real-time data entry using touch screen computers in Malawi [13], suggesting additional technology-driven approaches to improve data accuracy. Although a direct comparison is not possible, in all of these examples, the overall improvement in data quality was less than in the embedded approach used in the current study in Ethiopia. This suggests that data quality improvement may be more successful when participants experience directly how the data is being used to provide a positive improvement in the care of their patients. Despite applying a similar approach to data accuracy, variation in improvement in data accuracy across collaboratives was reported. This may be due to variation in the issues faced within each collaborative, and shared learning within each collaborative not being similarly shared across collaboratives.

At baseline, a significant proportion of health facilities over reported syphilis screening (50%) and pregnant women that received antenatal care four visit (60%) as compared to pregnant women that received antenatal care first visit (35%) and Skilled Birth Attendance (15%). The finding is consistent with recent studies in south west [14] and southern [15] Ethiopia. This could be due to error in counting non-serial data elements from registers or intentional over reporting.

A limitation of this study is the use of data in a limited number of health facilities (20) and a lack of data on factors associated with data accuracy at baseline. The lack of 8 time points before intervention, limited the statistical power of the study [16]. Despite the improvement in data accuracy observed during this study, the absence of a comparison study arm limits our ability to conclude that there was a strong cause-and-effect relationship with the intervention. Part of the improvement may have resulted from the readiness to improve service quality. However, there have been no specific efforts to improve data quality in the study facilities or woredas at the time of the intervention.

Despite these limitations, the improvement in data quality observed in this study is encouraging, it suggests a similar approach of embedding data quality improvement efforts within a wider initiative where participants experience how the data can be used to improve care more broadly, could improve the quality of the data needed for decision-making and resource allocation in other public health programs.

Conclusion

This study reports a simple, practical approach to improving the quality of public health information, both locally at health facility and in a woreda health information system. This data quality intervention improved both data accuracy on antenatal care1, antenatal care4 and syphilis screening in this study. Further research is needed to assess the effectiveness of similar data quality improvement approaches prospectively on a large scale.

Acknowledgments

We would like to thank Institute for HealthCare Improvement (IHI), Ethiopia project office, Addis Ababa, Ethiopia for unreserved cooperation in providing data for this study and time for data analysis & manuscript writing.

Data Availability

The data underlying the results presented in the study are available from Institute for HealthCare Improvement(IHI)http://www.ihi.org/.

Funding Statement

The author(s) received no specific funding for this work.

References

  • 1.The Breakthrough Series: IHI’s Collaborative Model for Achieving Breakthrough Improvement. [cited 18 Jun 2019]. Available: http://www.ihi.org/resources/Pages/IHIWhitePapers/TheBreakthroughSeriesIHIsCollaborativeModelforAchievingBreakthroughImprovement.aspx
  • 2.Wagenaar BH, Sherr K, Fernandes Q, Wagenaar AC. Using routine health information systems for well-designed health evaluations in low- and middle-income countries. Health Policy Plan. 2016;31: 129–135. 10.1093/heapol/czv029 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3.WHO | Data Quality Review (DQR) Toolkit. In: WHO [Internet]. [cited 16 Dec 2018]. Available: http://www.who.int/healthinfo/tools_data_analysis/dqr_modules/en/
  • 4.Mutale W, Chintu N, Amoroso C, Awoonor-Williams K, Phillips J, Baynes C, et al. Improving health information systems for decision making across five sub-Saharan African countries: Implementation strategies from the African Health Initiative. BMC Health Serv Res. 2013;13 Suppl 2: S9 10.1186/1472-6963-13-S2-S9 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5.Nwankwo B, Sambo MN. Can training of health care workers improve data management practice in health management information systems: a case study of primary health care facilities in Kaduna State, Nigeria. Pan Afr Med J. 2018;30 10.11604/pamj.2018.30.289.15802 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6.Mphatswe W, Mate KS, Bennett B, Ngidi H, Reddy J, Barker PM, et al. Improving public health information: a data quality intervention in KwaZulu-Natal, South Africa. Bull World Health Organ. 2012;90: 176–182. 10.2471/BLT.11.092759 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7.How to Improve. [cited 18 Jun 2019]. Available: http://www.ihi.org/resources/Pages/HowtoImprove/default.aspx
  • 8.Bernal JL, Cummins S, Gasparrini A. Interrupted time series regression for the evaluation of public health interventions: a tutorial. Int J Epidemiol. 2017;46: 348–355. 10.1093/ije/dyw098 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9.Braa J, Heywood A, Sahay S. Improving quality and use of data through data-use workshops: Zanzibar, United Republic of Tanzania. Bull World Health Organ. 2012;90: 379–384. 10.2471/BLT.11.099580 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10.Otwombe KN, Wanyungu J, Nduku K, Taegtmeyer M. Improving national data collection systems from voluntary counselling and testing centres in Kenya. Bull World Health Organ. 2007;85: 315–318. 10.2471/blt.06.033712 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11.Huaman MA, Araujo-Castillo RV, Soto G, Neyra JM, Quispe JA, Fernandez MF, et al. Impact of two interventions on timeliness and data quality of an electronic disease surveillance system in a resource limited setting (Peru): a prospective evaluation. BMC Med Inform Decis Mak. 2009;9: 16 10.1186/1472-6947-9-16 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12.Accuracy and quality of immunization information systems in forty‐one low income countries—Bosch‐Capblanch—2009—Tropical Medicine & International Health—Wiley Online Library. [cited 9 Apr 2019]. Available: https://onlinelibrary.wiley.com/doi/full/10.1111/j.1365-3156.2008.02181.x [DOI] [PubMed]
  • 13.Douglas G, Killam W, Hochgesang M, Deula R, Limbe W, Davis M. Improving Completeness, Accuracy & Timeliness of HIV Voluntary Counseling & Testing Client Data in Malawi Using Touchscreen Computers. AMIA Annu Symp Proc. 2005;2005: 942. [PMC free article] [PubMed] [Google Scholar]
  • 14.Ouedraogo M, Kurji J, Abebe L, Labonté R, Morankar S, Bedru KH, et al. A quality assessment of Health Management Information System (HMIS) data for maternal and child health in Jimma Zone, Ethiopia. PLOS ONE. 2019;14: e0213600 10.1371/journal.pone.0213600 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15.Endriyas M, Alano A, Mekonnen E, Ayele S, Kelaye T, Shiferaw M, et al. Understanding performance data: health management information system data accuracy in Southern Nations Nationalities and People’s Region, Ethiopia. BMC Health Serv Res. 2019;19: 175 10.1186/s12913-019-3991-7 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16.Penfold RB, Zhang F. Use of Interrupted Time Series Analysis in Evaluating Health Care Quality Improvements. Acad Pediatr. 2013;13: S38–S44. 10.1016/j.acap.2013.08.002 [DOI] [PubMed] [Google Scholar]

Decision Letter 0

Russell Kabir

9 Jan 2020

PONE-D-19-32598

Effect of data quality improvement intervention on health management information system data accuracy: an interrupted time series analysis

PLOS ONE

Dear Mulissa,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

We would appreciate receiving your revised manuscript by 10 February 2020. When you are ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter.

To enhance the reproducibility of your results, we recommend that if applicable you deposit your laboratory protocols in protocols.io, where a protocol can be assigned its own identifier (DOI) such that it can be cited independently in the future. For instructions see: http://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). This letter should be uploaded as separate file and labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. This file should be uploaded as separate file and labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. This file should be uploaded as separate file and labeled 'Manuscript'.

Please note while forming your response, if your article is accepted, you may have the opportunity to make the peer review history publicly available. The record will include editor decision letters (with reviews) and your responses to reviewer comments. If eligible, we will contact you to opt in or out.

We look forward to receiving your revised manuscript.

Kind regards,

Russell Kabir, PhD

Academic Editor

PLOS ONE

Journal Requirements:

When submitting your revision, we need you to address these additional requirements.

1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at

http://www.journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and http://www.journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Partly

Reviewer #2: Partly

**********

2. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: No

Reviewer #2: I Don't Know

**********

3. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: No

Reviewer #2: No

**********

4. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

**********

5. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: Thank you very much for giving an opportunity to review the present manuscript. This study sets out to describe the extent to which the accuracy of the HMIS data was improved during the improvement collaborative embedded data quality improvement initiative, as measured using the World Health Organization (WHO) data quality review methodology in Ethiopia. They found accurate data across all time periods in one region (Tanqua Abergele) while in two of the remaining three regions (Lem Bilbilu and Duguna Fango), data quality improved for all core metrics over time.

The manuscript needs some improvement in mainly the methods and results sections as a number of pages and lines should were not well described in the manuscript.

[Materials and methods]

1. Line 85 - The authors indicated that “one woreda in each of the four most populous agrarian regions of Ethiopia to introduce the improvement collaborative approach” but how the woreda were selected was not described.

2. Measures

Table 1: The authors definition for Antenatal Care 1 and Antenatal Care 4 were not that clear. I suggest the authors defined Antenatal Care 1 as “Number of pregnant women who had at least one antenatal care visit during their pregnancy”.

131 Data management and analysis

3. The authors did not mention any of the descriptive statistics used in this section.

4. What also informed the authors’ choice of median (inter-quartile range) instead of mean and standard deviation as the measure of central tendency and dispersion in the results section. If a test of normality was done, indicate the specific test statistic used.

5. 143-144: The authors described the selection of data used as consecutive sampling but go further to indicate that all data available during the study period was used. This is an indication of a census and not a sample as described by the authors.

6. What informed the authors choice of the slope change impact model used?

7. Did the authors assessed serial autocorrelation, non-stationarity and seasonality?

8. How was the fitness of the final selected model assessed? The authors should provide information on that.

9. Scatter plots of the various core measures over time should be added to help visualize the distribution of the data.

10. What was the level of significance used?

Results

11. 157-158 The authors mentioned nothing about what happened to the health centres without data personels and the 82.4% of health center staffs who were not trained. If nothing was done for them, if nothing was done for them won’t it affect the expected results?

12. An overall verification factor for the whole 3 regions will be important to be added in Table 2.

13. Verification factor for the intervention and post intervention period combined will be very informative since the impact of the intervention is not expected to take longer to be released.

14. Overall time series analysis for the verification factor for the core measures for the whole 2 regions combined will be important to be added in Table 3.

15. Time series analysis for the verification factor for the core measures for the intervention and post intervention periods combined will be very informative since the impact of the intervention is not expected to take longer to be released

16. Foot notes should be added to Tables to explain abbreviations used (CI, P).

17. The interpretations for the result should preside the tables.

18. The interpretation of Table 3 results was poorly done . it was explained in only line 178

19. Nothing was mentioned on the significance of the changes observed.

Conclusion

20. The conclusion should be rewritten to be based on the findings of the study.

Reviewer #2: This paper is may be interesting but very difficult to understanding for reader. Actually I can not understand the objectives of this paper. even i think the findings of this paper may be not sound. i think advanced statistical analysis may be upgrade the quality of this paper.

**********

6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files to be viewed.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email us at figures@plos.org. Please note that Supporting Information files do not need this step.

PLoS One. 2020 Aug 14;15(8):e0237703. doi: 10.1371/journal.pone.0237703.r002

Author response to Decision Letter 0


13 Jul 2020

Reviewer#1

[Materials & Methods]

1. Line 85 – The authors indicated that “one woreda in each of the four most populous agrarian regions of Ethiopia to introduce the improvement collaborative approach” but how the woreda were selected was not described.

We have added details of the Woreda selection (line 88).

2. Measures

Table1: The authors definition for antenatal care1 and antenatal care4 were not that clear. I suggest that authors defined antenatal care1 as “Number of pregnant women who had at least one antenatal care visit during their pregnancy”.

Thank you, we have redefined Antenatal care1 and antenatal care4 as per the suggestion.

131: Data management and analysis

3. The authors didn’t mention any of the descriptive statistics used in this section

We have added details of the descriptive statistics used in the data management and analysis section (line 138).

4. What also informed the authors’ choice of median (inter-quartile range) instead of mean and standard deviation as the measure of central tendency and dispersion in the result section. If a test of normality was done, indicate the specific test statistic used.

As we used mean for the time series analysis, we replaced all median (inter-quartile range) throughout the paper with mean (standard deviation).

5. 143-144: The authors described the selection of data used as consecutive sampling but go further to indicate that all data available during the study period was used. This is an indication of a census and not a sample as described by the authors.

We have clarified this description in the Sample Size section.

6. What informed the authors choice of the slope change impact model used?

The study referred to in the introduction, set in South Africa by Mphatswe W et al (Bull World Health Organ. 2012) indicated data accuracy changed shortly after the data quality intervention. This led us to hypothesize that a similar finding may occur in the current study, and thus we applied a model that would allow us to detect changes in the slope and overall average (step change) after the introduction of the data quality improvement efforts in the current study.

7. Did the authors assessed serial autocorrelation, non-stationarity and seasonality?

Yes, we checked autocorrelation using Durbin-Watson d-statistic and it is zero (0) suggesting the outcomes are independent. We found no non-stationarity and seasonality. We have added details of this into the results section (line 179).

8. How was the fitness of the final selected model assessed? The authors should provide information on that.

We have added details to the results section describing how we used standard approaches to assess fitness, using residual plots and examination of autocorrelation and seasonality, as described in #7 above.

9. Scatter plots of the various core measures over time should be added to help visualize the distribution of the data.

We have updated the Figure 2, to show both the individual data points and the fitted models.

10. What was the level of significance used?

We used a significance level of 0.05 and have added this to the methods section.

Results

11. 157-158 The authors mentioned nothing about what happened to the health centers without data personnels and the 82.4% of health center staffs who were not trained. If nothing was done for them, won’t it affect the expected results?

We have clarified this section in the text. The data quality improvement activities were embedded within a wider improvement collaborative. Part of the underlying theory of an improvement collaborative is that attendees will take their learning back to their home facilities and apply it with local staff.

12. An overall verification factor for the whole 3regions will be important to be added in Table2.

We have added the overall verification factor to Table 2

13. Verification factor for the intervention and post intervention period combined will be very informative since the impact of the intervention is not expected to take longer to be released.

The data quality improvement initiative was embedded within a wider quality improvement collaborative. As such the data quality improvement activities were introduced and built upon over time and ended when the wider improvement collaborative was completed. The post intervention period is thus a different period to the intervention period allowing us to assess sustainability. Consequently, we believe it is important to keep the intervention and post-intervention periods separate.

14. Overall time series analysis for the verification factor for the core measures for the whole 2regions combined will be important to be added in Table3.

Quality improvement initiatives are reliant on local engagement with activities, and on local context. This leads to the timing of improvement occurring, or not, to often differ from one setting to another. Consequently, we choose to focus of the time series analysis on the two Woredas separately, rather than combine them, which can risk to important variation being lost. As such, we argue strongly not to combine the results across the two Woredas, and to focus on learning from the individual woredas.

15. Time series analysis for the verification factor for the core measures for the intervention and post intervention periods combined will be very informative since the impact of the intervention is not expected to take longer to be released.

As described in our response to #13 above, we believe it is important to keep the intervention and post-intervention phases separate.

16. Foot notes should be added to tables to explain abbreviations used (CI, P)

We have added the foot notes as suggested.

17. The interpretations for the result should preside the tables.

We have re-positioned the tables so the description of the results comes before them.

18. The interpretation of table3 results was poorly done. It was explained in only line 178.

We have updated the wording to this section to better describe the analysis.

19. Nothing was mentioned on the significance of the changes observed.

We have included reference to the relevant p-values in the results section.

Conclusion

20. The conclusion should be rewritten to be based on the findings of the study.

We have re-written the conclusion to be more closely based on the findings of the study (line 258).

Reviewer#2: This paper may be interesting but very difficult to understand for the reader. Actually, I can’t understand the objectives of this paper, even I think the findings of this paper may be not sound. I think advanced statistical analysis may upgrade the quality of this paper.

Based on the responses to Reviewer #1, we believe this paper is now clearer, and is statistically sound.

Attachment

Submitted filename: Response to reviewers.docx

Decision Letter 1

Russell Kabir

3 Aug 2020

Effect of data quality improvement intervention on health management information system data accuracy: An interrupted time series analysis

PONE-D-19-32598R1

Dear Dr. Mulissa,

We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements.

Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication.

An invoice for payment will follow shortly after the formal acceptance. To ensure an efficient process, please log into Editorial Manager at http://www.editorialmanager.com/pone/, click the 'Update My Information' link at the top of the page, and double check that your user information is up-to-date. If you have any billing related questions, please contact our Author Billing department directly at authorbilling@plos.org.

If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

Kind regards,

Russell Kabir, PhD

Academic Editor

PLOS ONE

Additional Editor Comments (optional):

Reviewers' comments:

Acceptance letter

Russell Kabir

5 Aug 2020

PONE-D-19-32598R1

Effect of data quality improvement intervention on health management information system data accuracy: An interrupted time series analysis.

Dear Dr. Mulissa:

I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now with our production department.

If your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information please contact onepress@plos.org.

If we can help with anything else, please email us at plosone@plos.org.

Thank you for submitting your work to PLOS ONE and supporting open access.

Kind regards,

PLOS ONE Editorial Office Staff

on behalf of

Dr. Russell Kabir

Academic Editor

PLOS ONE

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    Attachment

    Submitted filename: Response to reviewers.docx

    Data Availability Statement

    The data underlying the results presented in the study are available from Institute for HealthCare Improvement(IHI)http://www.ihi.org/.


    Articles from PLoS ONE are provided here courtesy of PLOS

    RESOURCES