Abstract
Using electronic health records (EHR) to automate publicly reported quality measures is receiving increasing attention and is one of the promises of EHR implementation. Kaiser Permanente has fully or partly automated six of 13 the joint commission measure sets. We describe our experience with automation and the resulting time savings: a reduction by approximately 50% of abstractor time required for one measure set alone (surgical care improvement project). However, our experience illustrates the gap between the current and desired states of automated public quality reporting, which has important implications for measure developers, accrediting entities, EHR vendors, public/private payers, and government.
Keywords: Electronic Health Records, Quality Assurance/Health Care, Automatic Data Processing
Introduction
Quality measurement, a key lever to improve healthcare, has traditionally relied on administrative claims data and time-consuming manual chart abstraction.1 Health information technology (IT) promises to generate quality measurement and public reporting through automated data collection more readily.2
Many believe that electronic health records (EHR) offer new potential for quality measurement.3 The US Health Information Technology for Economic and Clinical Health Act of 2009 invested US$20 billion for health IT infrastructure and Medicare and Medicaid ‘meaningful use’ (MU) incentives. The proportion of acute care hospitals adopting at least a basic EHR more than doubled between 2009 and 2011, and in 2011 85% of hospitals planned to attest to MU of certified EHR technology by 2015, which includes submitting clinical quality measures.4 Stage 1 measures include 15 metrics focusing on emergency department (ED) throughput, stroke care, venous thromboembolism (VTE) prevention, and anticoagulation.5
The taxonomy for measuring and reporting performance from EHR is evolving; we use ‘e-measures’ and ‘automated measures’ interchangeably to refer to all partly or fully automated processes for generating performance information from EHR-contained data. Multiple stakeholders are vested in their development: the consumers and communities e-measures are intended to benefit, governmental entities, and commercial payers.1 At the request of the US Department of Health and Human Services, the National Quality Forum convened diverse stakeholders to provide recommendations for retooling 113 paper-based measures to an electronic format.6
e-Measure implementation occurs at the interface between measure developers and providers using EHR. Understanding the work of retooling paper-based quality measures for automated reporting illuminates the gap between the current and desired states of e-measures; we report here Kaiser Permanente's experience with automating quality reporting.
Automated quality reporting at Kaiser Permanente
Kaiser Permanente's EHR, KPHealthConnect, enables clinicians and employees to manage the healthcare and administrative needs of nine million members across eight geographic regions in a seamless and integrated way, with resulting quality and efficiency benefits.2 7 8 Implemented beginning in 2003, KP HealthConnect is deployed program-wide; individual regions customize EHR builds to local conditions and needs.
In 2010, Kaiser Permanente care reporting staff began to re-tool selected the joint commission (TJC) core measures for automated quality reporting. The Northern California, Southern California and Northwest regions currently use or are developing e-measures. Each measure comprises numerous data elements identifying population inclusion and exclusion and outcomes criteria.
Kaiser Permanente selected measures for automation on the basis of their clinical significance, importance to regulatory reporting, and reliance on discrete data elements: unambiguous individual numeric or coded values, such as cardiac rate and rhythm. Discrete data greatly reduce (but do not eliminate) the possibility of inaccurate reporting. They enable semantic understanding of the nature of the underlying information. In contrast, non-discrete data fields contain strings of indistinguishable characters; an example is the VTE-5 measure: VTE warfarin therapy discharge instructions. Documentation takes the form of embedded text, and text mining or current natural language processing cannot meet the current 100% accuracy requirement for publicly reported measures.
To date, Kaiser Permanente has focused on six core measure sets: acute myocardial infarction, ED patient flow, immunizations, the surgical care improvement project, pneumonia, and VTE prophylaxis. Kaiser Permanente has partly or fully automated 21 of 29 measures in these sets. The number of data elements per measure ranges from eight to 93, and the proportion that is discrete and mapped ranges from 43% to 100% (table 1). We derive the balance of TJC core measure reports from administrative claims data and manual chart abstraction.
Table 1.
TJC core measure | Required data fields | Mapped discrete data fields | Mapped data elements (%) | Time saved per case in minutes |
---|---|---|---|---|
SCIP | 93 | 43 | 46 | 11 |
AMI | 41 | 20 | 49 | 10 |
PN | 41 | 28 | 68 | 5 |
IMM | 15 | 15 | 100 | 5 |
VTE | 67 | 23 | 43 | 14 |
ED | 8 | 8 | 100 | 5 |
AMI, acute myocardial infarction; ED, emergency department; IMM, immunizations; PN, pneumonia; SCIP; surgical care improvement project; TJC, the joint commission; VTE, venous thromboembolism.
Figure 1 represents the iterative five-step process of developing and maintaining selected measures for automated reporting. A more extended discussion is available online (see supplementary web-only appendix 1, available online only).
Core measure interpretation
Measures first require interpretation in light of current specifications.9 Domain experts, such as abstractors and analysts, informaticists, legal and compliance staff, and clinical experts, provide input. Measures interpretation must be consistent throughout Kaiser Permanente.
Mapping
Mapping links specifications of the interpreted measure to EHR data tables. EHR database design is highly complex and its relationship to clinical documentation is often obscure. In addition, regional configurations and local documentation workflows vary.
Coding
Coding extracts required data from the mapped data tables. Convergent medical terminology ‘groupers’—lists of similar medications maintained over time—are intended to eliminate the need to ‘hard-code’ the constantly evolving universe of individual medications.10
Quality assurance/validation
Validation of automated quality reports ensures accuracy and is conducted by comparing automated results to the official submission and re-examining interpretation, mapping, and coding to rectify discrepancies.
Maintaining code over time
TJC measure definitions change to reflect evolving evidence and clinical practice, and EHR updates and new releases may change database design. Diagnostic nomenclature, procedural codes, and medication identifiers may change, as can internal factors impacting automated quality reporting.
Results of automated quality reporting
With validated accuracy of 100%, potential gains to the organization result from increased efficiency. To assess these, we compared pre and post-automation extraction time.
We measured the time required to abstract each of 20 randomly selected cases using manual abstraction methods. After partial or full automation was complete, we again measured the time required to abstract each case. Our calculations omit the time required for automation; for example, an 8-week collaborative effort between national reporting and regional abstraction staff fully automated five individual immunizations measures.
Table 1 contains the average time savings per case from using partly or fully automated reporting, compared to manual chart abstraction. In addition, we estimated the costs of development and ongoing maintenance for these measures, which included time for programmers, registered health information technicians (chart abstractors), and management and oversight, as well as a portion of IT infrastructure (eg, hardware and software). We found a breakeven in savings over cost by the fourth year, by then achieving an ongoing savings stream of approximately US$1 million annually for one Kaiser region.
Discussion
Our experience illustrates opportunities and challenges inherent in e-measures and gaps between the current and potential states of automated quality reporting. The five-step process we describe can serve as a template for the development of automated measures elsewhere, but it is unlikely to expedite the local development process. The time savings we observed highlight substantial opportunities for increased efficiency to augment the intrinsic benefits of performance reporting. Overhead costs of public quality reporting are significant. For example, Kaiser Permanente has reported on 50 well-established metrics for 15 years; the annual cost is approximately US$6.75 million, excluding expenses related to IT systems, storage, and oversight (unpublished data, Kaiser Permanente, 2010). We observed an approximate 50% reduction in abstraction time just for the partly automated surgical care improvement project measures; this time savings is likely to be broadly achievable.
Partial abstraction saves time over a completely manual process, expanding the capacity of existing abstraction staff and allowing us to forego hiring additional abstractors despite an expanding number of quality measures. Each automated element has been rigorously tested and validated by subject matter experts and needs no manual review for confirmation. In addition, even in some instances when we cannot fully automate a field, we supply a ‘trigger location’ within the EHR for confirmation by manual review, changing abstraction from an intuitive search to a focused verification.
The number of contemplated and required public reporting initiatives is growing exponentially. Automated quality reporting offers an opportunity to obtain the transparency and accountability benefits of increased reporting without adding to high overhead costs or depleting organizational quality improvement budgets.11 In addition, when quality measures are automated, some data are available as real-time organizational intelligence, expediting care improvement cycles.
Automated reporting can also be extended to quality measures that are not uniformly publicly reported; for example, surgical site infections (SSI). The Centers for Medicare and Medicaid Services requires hospitals to report infections related to colon surgeries and abdominal hysterectomies to the National Healthcare Safety Network for eventual publication on the Centers for Medicare and Medicaid Services Hospital Compare website.12 13 Although state policies vary, the California Department of Public Health requires SSI reporting for 29 National Healthcare Safety Network-defined procedures, with data published online.14–16 An automated SSI reporting process reduced Kaiser Permanente's manual surveillance full time equivalent by 80%, reflecting an aggregate savings of US$2 million (see supplementary appendix 2, available online only).
However, our experience also highlights the challenges of automating quality reporting. A recent report questions the accuracy of automated reporting across conditions.17 Achieving 100% accuracy required additional staff time to rectify discrepancies between methods; this time would probably be required in all settings.
Although little has been reported on automated quality reporting across conditions, our experience is supported by existing evidence. A primary challenge is that EHR were not initially designed to calculate, compile, and report on quality measures. Their core function is to capture, store, and track clinical data to support transactions.18 19
Consequently, much of the quality reporting supported by EHR is not fully automated.20 EHR meeting explicit stage 1 MU data capture requirements are estimated to provide approximately 35% of needed data.21 EHR also including electronic physician notes and medication administration records may provide up to 65% of needed data.21 Similarly, across the measures reported here, an average of 61% of needed data was available as discrete elements. Our experience is likely to be generalizable to a large extent, as we rely on data that are available to providers during the course of an encounter. All certified EHRs are likely to have these data. Some types of data lend themselves to discrete representation: vital signs, medications, and coded diagnoses and procedures, for instance. Other types of data are intrinsically more nuanced, such as discharge teaching. Our priority is to capture data that are a natural ‘exhaust’ from provider workflows, rather than asking physicians and nurses to interrupt natural care processes to complete a template or form.
Solutions to bridge the gap between the current and desired states of automated quality reporting are complex. One option, full automation, requires that all data elements are represented by standardized terminologies and codes within an EHR system and that the same standards are used locally and nationally.22 This is impractical. Many data elements that are difficult or impossible to automate are also essential for measure meaningfulness. For instance, an acute myocardial infarction measure relates to smoking cessation counseling, which is recorded narratively in progress notes or a teaching summary, precluding automation. If measures lack sufficient meaningfulness, physicians and other clinicians will have less incentive to drive operational and practice changes to improve performance on them.
Strong federal involvement and guidance would be required to achieve a highly coordinated approach to addressing gaps in automated quality measurement standards and processes;23 this may risk generating reporting requirements that inappropriately drive cumbersome clinical workflows. A collaborative approach is evidenced by numerous stakeholders at the national level, including the Agency for Healthcare Quality and Research and, as noted earlier, the National Quality Forum, working together and independently to advance e-measures.24 However, the continuing development of quality measures should not embrace automation at the cost of meaningful clinical detail or over-burdening clinician workflow.
Many questions remain. Even with improved standardization of terminologies and codes, EHR content, structure, and data format vary, as do local data capture and extraction procedures.25 Within a single institution, significant differences in denominators, numerators, and rates arise from different electronic data sources, and documentation habits of providers vary.26 Data entered into the EHR may not be interpreted or recognized, resulting in substantial numerator loss and underestimates of the delivery of clinical preventive services.27
EHR vendors can potentially support e-measures. However, organizations typically customize vendor-provided builds and workflows are local; data extraction will require local customization, too. Structured clinical data are often captured electronically through clinical reminders that are relatively insensitive to context and interfere with workflows, and adapting workflows to support documentation is short sighted.28 29 For instance, data supporting the medication reconciliation measure for stage 2 MU can be found in medication actions, such as new, changed, discontinued, or adjusted medications or a refill request, from which review and reconciliation can be inferred. A vendor-generated solution might be a check box for ‘medication reconciliation complete.’ This both adds an inefficient step to the workflow and creates the possibility of providers indicating that reconciliation occurred without conducting it.
A final comment pertains to the difference between retooling quality measures that were designed for manual abstraction and developing quality measures for internal use de novo, with which Kaiser Permanente has robust experience. Given the level of exactitude required to retool manual measures for automation, de-novo development of quality measures can potentially be more straightforward. However, our experience is that the latter also incurs substantial development costs to capture variability in clinical workflows.
Conclusions
Kaiser Permanente has fully or partly automated six TJC measures. Time savings from automation were substantial, but our experience illustrates the complex nature of this undertaking and the gap between the current and desired states of automated quality reporting. The goal of fully automating quality measurement may challenge the goals of supporting provider-driven efficient workflows and retaining the meaningfulness of quality measures.
Supplementary Material
Footnotes
Correction notice: This paper has been made unlocked since it was published Online First.
Acknowledgements: The authors acknowledge the support and leadership of the following members of the Kaiser Permanente quality and informatics teams: Greg Huffman, current director of care reporting, Randy Watson, quality data manager; Lorelle Poropat, practice leader for regional data consulting; and Brian Hoberman, MD, physician leader for Northern California KP HealthConnect. In addition, dozens of experts and providers help support the ongoing effort to automate publicly reported quality measures at regional and national levels, including Andy Amster, director of the Center for Health Care Analytics.
Contributors: TG, JW, AW and BC sponsored the work. TG and AW initiated the work. TG, SK, AW and BC designed the structure of the work and plan. SK, ML, DK and JL did the programing, automation, and quality assurance. TG, AW and BC removed barriers as the work progressed. TG and SK designed the data collection. SK, ML, DK and JL conducted the data collection. TG, SK, ML, DK and JL documented the automation process. TG and JW drafted the initial paper. TG, JW, AW, BC, SK and DK reviewed and provided comments. TG is the guarantor.
Competing interests: None.
Provenance and peer review: Not commissioned; externally peer reviewed.
References
- 1.Anderson KM, Marsh CA, Flemming AC, et al. Quality measurement enabled by health IT: overview, possibilities, and challenges. Rockville, MD: Agency for Healthcare Research and Quality, 2012 [Google Scholar]
- 2.Weissberg J. Electronic health records in a large, integrated health system: It's Automatic...NOT! At Least, Not Yet. Rockville, MD: Agency for Healthcare Research and Quality; 3 August 2009. http://www.qualitymeasures.ahrq.gov/expert/expert-commentary.aspx?id=16458 (accessed 15 Oct) [Google Scholar]
- 3.Roski J, McClellan M. Measuring health care performance now, not tomorrow: essential steps to support effective health reform. Health Aff (Millwood) 2011;30:682–9 [DOI] [PubMed] [Google Scholar]
- 4.Charles D, Furukawa M, Hufstader M. Electronic health record systems and intent to attest to meaningful use among non-federal acute care hospitals in the United States: 2008–2011. ONC Data Brief 2012. http://www.healthit.gov/media/pdf/ONC_Data_Brief_AHA_2011.pdf (accessed 5 Mar 2013)
- 5.Centers for Medicare and Medicaid Services Medicare & Medicaid EHR incentive program: meaningful use stage 1 requirements overview. Centers for Medicare & Medicaid Services, 2010. https://www.cms.gov/Regulations-and-Guidance/Legislation/EHRIncentivePrograms/downloads/MU_Stage1_ReqOverview.pdf (accessed 5 Mar 2013) [Google Scholar]
- 6.National Quality Forum Measuring performance: electronic quality measures (eMeasures). Washington, DC: National Quality Forum, 2012. http://www.qualityforum.org/Projects/e-g/eMeasures/Electronic_Quality_Measures.aspx?section=2011UpdatedeMeasures2011-02-012011-04-01 (accessed 24 Nov 2012) [Google Scholar]
- 7.Silvestre AL, Sue VM, Allen JY. If you build it, will they come? The Kaiser Permanente model of online health care. Health Aff (Millwood) 2009;28:334–44 [DOI] [PubMed] [Google Scholar]
- 8.Chen C, Garrido T, Chock D, et al. The Kaiser Permanente Electronic Health Record: transforming and streamlining modalities of care. Health Aff (Millwood) 2009;28:323–33 [DOI] [PubMed] [Google Scholar]
- 9.The Joint Commission Specifications manual for national hospital inpatient quality measures. Oakbrook Terrace, IL, 2012. http://www.jointcommission.org/specifications_manual_for_national_hospital_inpatient_quality_measures.aspx (accessed 15 Oct 2012) [Google Scholar]
- 10.Dolin RH, Mattison JE, Cohn S, et al. Kaiser permanente's convergent medical terminology. Stud Health Technol Inform 2004;107:346–50 [PubMed] [Google Scholar]
- 11.Meyer GS, Nelson EC, Pryor DB, et al. More quality measures versus measuring what matters: a call for balance and parsimony. BMJ Qual Saf 2012;21:964–8 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 12.Centers for Medicare and Medicaid Services. 2011. Operational guidance for reporting surgical site infection (SSI) data to CDC's NHSN for the purpose of fulfilling CMS's hospital inpatient quality reporting (IQR) program requirements. http://www.cdc.gov/nhsn/pdfs/final-ach-ssi-guidance.pdf (accessed 10 Dec 2012)
- 13.Hospital Compare. Washington, DC: Centers for Medicare and Medicaid Services, 2012. http://www.hospitalcompare.hhs.gov/ (accessed 10 Dec 2012) [Google Scholar]
- 14.Makary MA, Aswani MS, Ibrahim AM, et al. Variation in surgical site infection monitoring and reporting by state. J Healthc Qual 2012. http://onlinelibrary.wiley.com/doi/10.1111/j.1945-1474.2011.00176.x/abstract (accessed 2 Mar 2013) [DOI] [PubMed] [Google Scholar]
- 15.O'Reilly K. Only 14 states post hospital data on surgical site infections. American Medical News 2012. http://www.ama-assn.org/amednews/2012/04/02/prsb0402.htm (accessed 16 Nov 2012)
- 16.Surgical site infections (SSI) in California hospitals, 2011. Sacramento, CA: California Department of Public Health, 2012. http://www.cdph.ca.gov/programs/hai/Pages/SurgicalSiteInfections-Report.aspx (accessed 9 Dec 2012) [Google Scholar]
- 17.Kern LM, Malhotra S, Barron Y, et al. Accuracy of Electronically Reported ‘Meaningful Use’ Clinical Quality Measures: A Cross-sectional Study. Ann Intern Med 2013;158:77–83 [DOI] [PubMed] [Google Scholar]
- 18.Raths D. Report from eHI Forum: Moving EHRs from Transactions to Analytics. Healthc Inform 2012. http://www.healthcare-informatics.com/article/report-ehi-forum-moving-ehrs-transactions-analytics (accessed 5 Mar 2013) [Google Scholar]
- 19.Mandl KD, Kohane IS. Escaping the EHR trap—the future of health IT. N Engl J Med 2012;366:2240–2 [DOI] [PubMed] [Google Scholar]
- 20.Recommended common data types and prioritized performance measures for electronic healthcare information systems. Washington, DC: National Quality Forum, 2008 [Google Scholar]
- 21.Metzger J, Ames M, Rhoads J. Hospital quality reporting: the hidden requirements of meaningful use. Falls Church, VA: Computer Sciences Corporation, 2012 [Google Scholar]
- 22.Dykes PC, Caligtan C, Novack A, et al. Development of automated quality reporting: aligning local efforts with national standards. AMIA Annu Symp Proc 2010;2010:187–91 [PMC free article] [PubMed] [Google Scholar]
- 23.Fu PC, Jr, Rosenthal D, Pevnick JM, et al. The impact of emerging standards adoption on automated quality reporting. J Biomed Inform 2012;45:772–81 [DOI] [PubMed] [Google Scholar]
- 24.Agency for Healthcare Research and Quality Request for Information on quality measurement enabled by health IT. Federal Register 2012;77:42738–740 [Google Scholar]
- 25.Chan KS, Fowles JB, Weiner JP. Review: electronic health records and the reliability and validity of quality measures: a review of the literature. Med Care Res Rev 2010;67:503–27 [DOI] [PubMed] [Google Scholar]
- 26.Kahn MG, Ranade D. The impact of electronic medical records data sources on an adverse drug event quality measure. J Am Med Inform Assoc 2010;17:185–91 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 27.Parsons A, McCullough C, Wang J, et al. Validity of electronic health record-derived quality measurement for performance monitoring. J Am Med Inform Assoc 2012;19:604–9 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 28.Francis J. Quality measurement and the ubiquitous electronic health record. Health services research and development forum. Washington, DC: US Department of Veterans Affairs, 2012 [Google Scholar]
- 29.Burstin H. The journey to electronic performance measurement. Ann Intern Med 2013;158:131–2 [DOI] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.