Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2016 Oct 1.
Published in final edited form as: Med Care. 2015 Oct;53(10):901–907. doi: 10.1097/MLR.0000000000000418

Calculations of financial incentives for providers in a pay-for-performance program: Manual review versus data from structured fields in electronic health records

Tracy H Urech 1, LeChauncy D Woodard 1, Salim S Virani 1, R Adams Dudley 2,3, Meghan Z Lutschg 1, Laura A Petersen 1
PMCID: PMC4570867  NIHMSID: NIHMS708044  PMID: 26340661

Abstract

Background

Hospital report cards and financial incentives linked to performance require clinical data that are reliable, appropriate, timely, and cost-effective to process. Pay-for-performance plans are transitioning to automated electronic health record (EHR) data as an efficient method to generate data needed for these programs.

Objective

To determine how well data from automated processing of structured EHR fields (AP-EHR) reflect data from manual chart review and the impact of these data on performance rewards.

Research Design

Cross-sectional analysis of performance measures used in a cluster randomized trial assessing the impact of financial incentives on guideline-recommended care for hypertension.

Subjects

A total of 2,840 patients with hypertension assigned to participating physicians at 12 Veterans Affairs hospital-based outpatient clinics. Fifty-two physicians and 33 primary care personnel received incentive payments.

Measures

Overall, positive and negative agreement indices and Cohen's kappa were calculated for assessments of guideline-recommended antihypertensive medication use, blood pressure (BP) control, and appropriate response to uncontrolled BP. Pearson's correlation coefficient was used to assess how similar participants’ calculated earnings were between the data sources.

Results

By manual chart review data, 72.3% of patients were considered to have received guideline-recommended antihypertensive medications compared to 65.0% by AP-EHR review (k=0.51). Manual review indicated 69.5% of patients had controlled BP compared to 66.8% by AP-EHR review (k=0.87). Compared to 52.2% of patients per the manual review, 39.8% received an appropriate response by AP-EHR review (k=0.28). Participants’ incentive payments calculated using the two methods were highly correlated (r≥0.98). Using the AP-EHR data to calculate earnings, participants’ payment changes ranged from a decrease of $91.00 (−30.3%) to an increase of $18.20 (+7.4%) for medication use (IQR, −14.4% to 0%) and a decrease of $100.10 (−31.4%) to an increase of $36.40 (+15.4%) for BP control or appropriate response to uncontrolled BP (IQR, −11.9% to −6.1%).

Conclusions

Pay-for-performance plans that use only EHR data should carefully consider the measures and the structure of the EHR before data collection and financial incentive disbursement. For this study, we feel that a 10% difference in the total amount of incentive earnings disbursed based on AP-EHR data compared to manual review is acceptable given the time and resources required to abstract data from medical records.

Keywords: payment system design, performance measurement, incentives

Introduction

Quality improvement initiatives, such as report cards and financial incentives linked to performance, must utilize clinical performance measures that are reliable, appropriate, timely, and cost-effective to process. Collecting these data via manual review is considered the gold standard but is very time- and resource-intensive. Electronic health records (EHRs) and structured data from electronic fields collected in data warehouses are changing how health care providers document care and are increasingly viewed as a potential tool for performance measurement.

There is a nascent but growing body of literature comparing data derived by automated processing of EHRs (AP-EHR) to measures collected by chart review. A study examining “meaningful use” clinical quality measures from electronic reporting and manual review found wide variation in the agreement for rates of recommended care for appropriate asthma medication, low-density lipoprotein cholesterol control for patients with diabetes mellitus, and pneumococcal vaccination.1 A comparison of EHR-queried and chart review data from the New York City Primary Care Information Project found that EHR-derived measures underreported patients meeting performance measures for six of 11 indicators.2 Recently, Gardner and colleagues reported low-to-moderate agreement for 12 measures of adolescent well-care across three sites when comparing data extracted from structured fields in the EHR to data collected via manual review.3

Understanding how well data from structured fields in EHRs reflect the totality of what is known about the patient is critical as national quality reporting programs transition from manual review to automated reports, since they are largely doing so without natural language processing of free text as a supplement to measurement algorithms. The CMS EHR Incentive Program4 has spurred the implementation of EHR systems. In 2013, 58.9% of hospitals had a basic or comprehensive EHR, an increase from approximately 15% in 2010, the initial year of the CMS EHR incentive program.5 Starting in 2014, hospitals with certified EHR systems can submit 16 electronically specified clinical quality measures for the CMS Hospital Inpatient Quality Reporting program.6 Reporting methods for the CMS Physician Quality Reporting System, one of the pay-for-performance programs established under the Affordable Care Act, include direct EHR using certified EHR technology.7 New York City's Health and Hospitals Corporation started a new pay-for-performance agreement linking physician bonus payments with quality and efficiency metrics, as well as implementation of an updated EMR system.8,9 Integrated Health Association's pay for performance program, the largest non-governmental incentive program in the United States with approximately 35,000 physicians in California, uses only electronic data sources to assess clinical quality measures.10 These policy decisions notwithstanding, to our knowledge there have been no studies that have assessed the impact of basing payment on AP-EHR instead of chart review.

We sought to determine how well data from AP-EHR of structured fields from the Department of Veterans Affairs’ (VA) EHR reflect data collected via manual chart review and to assess the potential impact of the data collection methodologies on incentive earnings disbursed to physicians and practice groups participating in a trial evaluating pay for performance for hypertension care.

Methods

Trial design and study period

Primary care physicians and practice team members (e.g., nurses) at 12 VA hospital-based outpatient clinics participated in a trial evaluating the impact of financial incentives on guideline-recommended care for hypertension. The study's design and main findings are described elsewhere.11,12 Briefly, hospitals were cluster randomized to one of four study groups – physician-level incentives, practice-level incentives, physician- plus practice-level (combined) incentives, and no incentives (control). The intervention included five fourth-month performance periods. For each period, we randomly sampled and reviewed the EHRs for 40 eligible hypertensive patients from the physician's panel. Participants in the intervention arms received incentive payments and all participants received audit and feedback performance reports.11,13 Participants earned payments based on the number of measures achieved per patient, with each measure rewarded the same amount.11,12 This study examines the trial's fifth performance period, April through July 2009.

Hypertension quality-of-care measures

We rewarded participants for prescribing guideline-recommended antihypertensive medications, achieving blood pressure (BP) thresholds, and appropriately responding to uncontrolled BP,11 as described in the “Seventh Report of the Joint National Committee on Prevention, Detection, Evaluation, and Treatment of High Blood Pressure” (JNC 7).14 Appropriate responses to uncontrolled BP included prescribing lifestyle modifications for Stage 1 hypertension or adjusting and/or starting guideline-recommended medications.

EHR manual record review

Trained research assistants, blinded to the study's objectives and study group assignments, reviewed assigned health records in the Veterans Health Information Systems and Technology Architecture (VistA) Computerized Patient Record System (CPRS).15,16 They collected demographics; comorbidities; vitals including BP readings, heart rate and pulse; names, dosages, and days supply for VA and non-VA prescribed medications; medication allergies, refusals, and contraindications; laboratory results; and lifestyle modifications.11 Abstraction included reviewing the structured sections of the EHR and the free-text provider notes from the clinical encounters where hypertension should be addressed. Chart reviewers examined care delivery in many different outpatient settings, including but not limited to the primary care, cardiology, diabetes, renal/nephrology, emergency department, and women's health. Some health information recorded during an encounter such as diagnoses and allergies/adverse reactions that are captured in structured sections can be linked to progress notes. Also, the abstraction tool contained questions to confirm the patient's eligibility for the study, including verifying the patient had hypertension, was on the physician's panel, and did not have a terminal illness.

AP-EHR data sources

We collected the same data elements listed in the chart abstraction tool from the AP-EHR of structured fields from the hospitals’ VistA/CPRS systems available in the VA Corporate Data Warehouse (CDW),17 the National Patient Care Database inpatient and outpatient SAS® datasets,18 and the VA Managerial Cost Accounting (MCA) Pharmacy and Laboratory National Data Extracts.19 These data sources capture all care provided in the VA system, not just at the twelve study sites. We also included data from the non-VA medical care files (care received outside the VA and paid by the VA)20 and the VA-CMS merged datasets.21 VA healthcare providers (e.g., physicians and nurses) enter their own data in the patient's medical record in CPRS.

Index visit

Chart reviewers determined if a patient had a qualifying outpatient visit during the performance period, defined as an encounter where hypertension should be addressed. We identified visits where BP readings were recorded using the CDW vital signs domain.

Use of guideline-recommended antihypertensive medications

To determine if patients were receiving guideline-recommended antihypertensive medications at index, we assessed their medications, accounting for coexisting conditions and allergies, contraindications, and refusals. Based on JNC 7 guidelines, we developed sophisticated medication scenario algorithms to determine medication appropriateness (Supplemental Appendix Table 1). We identified patients’ past and current medications using the MCA pharmacy data. Antihypertensive medications filled within 120 days prior to and including the index visit date were assessed. Although the standard prescription fill for these medications is 90 days, we included extra time for patients to refill medications or finish any stockpiled medications.

Using inpatient and outpatient diagnosis and procedure codes, we searched the AP-EHR data to identify patients’ JNC 7 compelling indications and other conditions that impact pharmacologic treatment, applying the same time intervals used in the abstraction tool (Supplemental Appendix Table 2). An internist from the trial study team and a cardiologist independently reviewed the codes.

To account for contraindications, allergies, refusals, and use of non-VA antihypertensive medications, we incorporated data from the allergy, vital signs, and health factors domains in the VA CDW. Providers are able to document refusals and use of non-VA antihypertensive medications in the free-text area of the Notes section in CPRS or by resolving a clinical reminder which creates text. Clinical reminder software in CPRS includes health factors, computerized components capturing patient information that does not have existing standard coding (e.g., outside medication – ACE inhibitor).22 The clinical reminder software, when activated by the provider, uses dialogue boxes to prompt the provider to check a box from a list of options that best reflects the patient encounter. Also, providers are able to document non-VA medications in structured fields in the Orders section in CPRS.

BP control and appropriate response to uncontrolled BP

We used outpatient BP readings from the VA CDW vital signs domain to assess each patient's BP control at index. We included BP readings from many types of clinics, including but not limited to primary care, cardiology, diabetes, renal/nephrology, emergency department, and women's health. Unlike the EHR manual review, we were not able to examine home or outside BP readings.

We used the MCA pharmacy data to examine antihypertensive medication adjustments in response to uncontrolled BP. Implementing the same approach as the EHR-abstracted data, we assessed appropriate responses at the index visit and during the follow-up period. We compared the patient's medications prior to the index visit, at index, and following the index visit. Using the prescription fill and daily dosage information, we identified dosage changes to existing medications. Medication adjustments were classified as either an increase, decrease, or no change in dosage.

We identified lifestyle modifications using outpatient visits and health factors. Using the list of modifications in the abstraction tool, we created the following categories to evaluate the AP-EHR data: nutrition/diet education and counseling, general patient education, exercise programs such as the Managing Obesity for Veterans Everywhere! (MOVE!®), BP education visits, and pharmacist-led education and counseling. We gave credit if the lifestyle modification health factor indicated the patient was referred to one of the above programs or if the patient declined the referral. Providers are able to document lifestyle modifications in the free-text area of the Notes section in CPRS or by resolving a clinical reminder (e.g., weight and nutrition screen).

Analysis

We calculated the overall, positive and negative agreement indices and Cohen's kappa for the JNC 7 compelling indications, other coexisting conditions, and the three performance measures.23-25 Overall agreement is the number of outcomes both data sources agree on divided by the total number of outcomes assessed. Positive/negative agreement is calculated by taking the number of positive/negative outcomes both data sources agree on and dividing by the number of positive/negative outcome identified by each data source. Cohen's kappa is a measure of agreement that corrects for agreement due to chance.26 We used the following guidelines to interpret the kappa values: 0.01-0.20 indicates slight agreement; 0.21-0.40 fair; 0.41-0.60 moderate; 0.61-0.80 substantial; and 0.81-0.99 reflects almost perfect agreement.27 Analyses were done using SAS version 9.1 (Cary, NC).

We calculated the financial incentive payments for the intervention participants using the same approach from the trial. For the guideline-recommended medication measure and the combined measure assessing controlled BP or an appropriate response to uncontrolled BP, we compared the overall amount of money that would have disbursed using the AP-EHR data to the amount that was disbursed based on manual chart abstraction. We assessed how similar the participants’ calculated earnings were between the two data collection methods using Pearson's correlation coefficient. We also examined the payment changes when AP-EHR data was used to determine earnings.

This study was approved by Baylor College of Medicine Institutional Review Board and the Michael E. DeBakey Research and Development Committee. The funding sources had no role in the study.

Results

We compared the performance measures for the randomly sampled 2,840 patients assigned to the 71 physicians (52 in the incentive groups and 19 in the control group) participating in the fifth (and final) intervention period of the trial. For the JNC 7 compelling indications, we found almost perfect agreement for diabetes mellitus and chronic kidney disease between the manual review and the AP-EHR data (Table 1). Agreement was lowest for unstable angina.

Table 1.

Agreement between manual review and AP-EHR data for comorbid conditions examined to evaluate guideline-recommended antihypertensive medication use

JNC 7 compelling indication or comorbid condition No. (%) pts, manual review data* No. (%) pts, AP-EHR data* Overall agreement Positive agreement Negative agreement Kappa
Diabetes mellitus 1090 (38.4) 1056 (37.2) 0.96 0.95 0.97 0.92
Chronic kidney disease 705 (24.8) 611 (21.5) 0.96 0.92 0.98 0.90
Nephropathy 212 (7.5) 174 (6.1) 0.98 0.87 0.99 0.86
Ischemic heart disease 686 (24.2) 682 (24.0) 0.92 0.84 0.95 0.79
Myocardial infarction 262 (9.2) 157 (5.5) 0.94 0.56 0.97 0.53
Unstable angina 45 (1.6) 42 (1.5) 0.98 0.39 0.99 0.38
Benign prostatic hyperplasia 900 (31.7) 473 (16.7) 0.83 0.66 0.89 0.56
Angioedema 21 (0.7) 19 (0.7) 0.99 0.55 0.99 0.55
Cor pulmonale 20 (0.7) 0 --- --- --- ---

AP-EHR indicates automated processing of structured fields in the electronic health record; JNC 7, Seventh Report of the Joint National Committee on Prevention, Detection, Evaluation, and Treatment of High Blood Pressure.

*

Examined 2,840 patient records.

In the manual review data, 2054 (72.3%) patients received guideline-recommended antihypertensive medications compared to 1847 (65.0%) patients in the AP-EHR data (Table 2). Overall agreement was 0.79 and kappa was 0.51 [95% confidence interval (CI), 0.48-0.55], indicating moderate agreement. Since medication appropriateness was dependent on the compelling indications, we analyzed the subset of those patients who had the same compelling indication classification in both data sources (n=2403). Agreement increased slightly (k=0.56, 95% CI, 0.53-0.60). When we included VA/CMS data to identify JNC 7 compelling indications not recorded in the AP-EHR data sources, agreement for medication use did not improve (k=0.50, 95% CI, 0.46-0.53).

Table 2.

Agreement for use of guideline-recommended antihypertensive medications, blood pressure control, and appropriate response to uncontrolled blood pressure

Rewarded clinical measure No. (%) pts, manual review data No. (%) pts, AP-EHR data Overall agreement Positive agreement Negative agreement Kappa
Use of guideline-recommended medications* 2054 (72.3) 1847 (65.0) 0.79 0.85 0.66 0.51
Controlled BP* 1974 (69.5) 1898 (66.8) 0.94 0.96 0.91 0.87
Appropriate response to uncontrolled BP 431 (52.2) 328 (39.8) 0.64 0.61 0.66 0.28

AP-EHR indicates automated processing of structured fields in the electronic health record; BP blood pressure.

*

Examined 2,840 patient records.

Among the 825 patients who had uncontrolled blood pressure in both data sources.

To assess the drivers of disagreement between the AP-EHR and manual review data, we re-examined 252 patient records that had VA antihypertensive medication use in both data sources and the same compelling indicator classification but did not agree on the performance measure. These represent 42.2% of the records that did not agree between the data sources. Manual review and AP-EHR data disagreed on only one element (e.g., beta blocker allergy) in the medication algorithms for 128 (50.8%) records and only two elements for 67 (26.6%) records. Among these 262 disagreements, 235 (89.7%) were due to the element not being captured in the AP-EHR data. Medication algorithm elements with the highest prevalence of disagreement were being prescribed a guideline-recommended thiazide medication (n=56, 21.4%), having a thiazide allergy/refusal (n=51, 19.5%), and having coexisting benign prostatic hyperplasia (n=49, 18.7%). Manual review documentation indicated that thiazide allergy/refusal and benign prostatic hyperplasia were primarily documented only in the Notes section of the record.

In the final performance period of the trial, 1974 (69.5%) patients had controlled BP based upon the manual review data. The proportion of patients meeting this measure was slightly lower (66.8%) using the AP-EHR data (Table 2). We report agreement indices separately for controlled BP and appropriate response to uncontrolled BP since appropriate response agreement could only be assessed for the subset of patients who had uncontrolled BP in both data sources (n=825). Among the subset, 52.2% received an appropriate response in the manual review data compared to 39.8% in the AP-EHR data. Of the three performance measures, agreement was highest for controlled BP and lowest for appropriate response, kappa values 0.87 (95% CI, 0.85-0.89) and 0.28 (95% CI, 0.22-0.35), respectively (Table 2). Agreement for controlled BP increased slightly after excluding the patients with controlled readings based on home/outside readings (k=0.90, 95% CI, 0.88, 0.92).

Among the 825 patients with uncontrolled BP in both data sources, 371 (45.0%) patients had documented lifestyle modifications according to the manual review data compared to 186 (22.6%) in the AP-EHR data. Overall agreement for lifestyle modifications was 0.60 and kappa was 0.16 (95% CI, 0.10-0.22). Restricting the analysis to the patients who had Stage 1 hypertension in both data sources (n=689) (and for whom lifestyle modifications were recommended by guidelines), overall agreement for the appropriate response measure was 0.61 and kappa was 0.24 (95% CI, 0.17-0.31).

Fifty-two physicians and 33 non-physician practice team members received incentive payments for the 4-month performance period. There were a total of 5 payment disbursements in the trial. Using manual review data, mean payments for the fifth performance period for the medication and the combined controlled BP or appropriate response to uncontrolled BP measures were $212.40 [standard deviation (SD)=$124.61] and $246.34 (SD=$126.60), respectively. The total amounts of incentives disbursed for guideline-recommended medications and controlled BP or appropriate response to uncontrolled BP were $18,054.40 and $20.939.10, respectively. Using the AP-EHR data, the participants’ mean payments would have been $191.21 (SD=$111.42) for the medication measure and $225.79 (SD=$114.55) for controlled BP or appropriate response. Total incentive earnings would have decreased by 10.0% and 8.3%, respectively. Individual participants’ incentive payments calculated using the two methods were highly correlated (r≥0.98). Using the AP-EHR data to calculate incentive earnings, payment changes ranged from a decrease of $91.00 (−30.3%) to an increase of $18.20 (+7.4%) for guideline–recommended medication use [interquartile range (IQR), −14.4% to 0%]. Payments would have decreased by 10% or more for 28 physicians and 11 practice team members (Figure). For BP control or appropriate response to uncontrolled BP, changes ranged from a decrease of $100.10 (−31.4%) to an increase of $36.40 (+15.4%) (IQR, −11.9% to −6.1%). Payments would have decreased by 10% or more for 18 physicians and 9 practice team members.

Figure.

Figure

Percent change in participants’ performance incentive payments using AP-EHR data sources.

Discussion

In this assessment of how well the AP-EHR reflect data from manual review and the impact of these data on performance rewards, we found almost perfect agreement for the BP control measure and low agreement for the appropriate response to uncontrolled BP measure. We also found moderate agreement between the data sources for use of guideline-recommended antihypertensive medication, a measure requiring information about the patients’ coexisting conditions, current medications, and medications allergies, contraindications, and refusals. We determined that the total amount of incentives disbursed to physicians and practice time members would have been lower using the AP-EHR data to reward performance. Using the trial's reward structure of individual-level, practice-level, and combined incentives, we found that payments calculated using the manual review and AP-EHR data were highly correlated.

Considering the manual review as the gold standard, the AP-EHR data underreported the number of patients receiving appropriate medications. This finding is similar to other studies comparing EHR-derived data and manual review.1,28 For BP agreement, we used the vital signs domain from the CDW. Borzecki and colleagues found that these data from VA CPRS are sufficient for assessing the quality of hypertension care and reviewing clinicians’ notes in the EHR for BP readings adds little information.29

This study's novel contributions to the literature are twofold. First, we are not aware of any studies that have compared manual review data to data from automated sources to evaluate providers’ responses to uncontrolled blood pressure. Secondly, and more importantly, we are not aware of any studies that have assessed how the differences in performance measures from manually reviewed data and EHR automated data impact providers financially in pay-for-performance plans. We feel that a 10% difference in the total amount of incentive earnings disbursed based on AP-EHR data compared to manual review is acceptable given the time and resources required to abstract data from medical records. Of the 71 physicians examined, 46 (66.2%) and 56 (78.9%) had a percent change in earnings less than 10%, for the use of guideline-recommended antihypertensive medications measure and the BP control or appropriate response to uncontrolled BP measure, respectively.

Our data are consistent with several studies that suggest that discrepancies between EHR-derived data and manual review are due to information documented in unstructured fields or as unstructured text in the health record.1-3,30,31 Understanding that the AP-EHR data sources are not a perfect reflection of the information in the patient's health record, we were not surprised that only 64.2% patients were identified as receiving guideline-recommended medications compared to 72.3% per the manual review. We assessed up to 19 elements in the medication algorithms for the compelling indication classifications, including coexisting conditions, antihypertensive medication classes, and allergies, contraindications, and refusals. Because the medication algorithm requires many pieces of clinical information that may not be documented in structured fields, it is more susceptible to missing data. We found in our reexamination of a subset of records that two of the three most prevalent disagreed elements in the medication algorithms were more often captured only in the Notes section of the patient's electronic medical record. If the trial had collected data via the AP-EHR approach instead of manual record review, our findings might have been different given that providers might have been more likely to document clinical information and interactions in structured fields in the electronic health record.

Our study has several limitations. First, we compared agreement for performance measures and the incentive payments for the records the reviewers identified as eligible. Manual review found that approximately 20% of the patient records identified from the physicians’ sampling frames developed using AP-EHR data were not eligible for the study. If our study used only AP-EHR data sources, we would have provided performance audit and feedback and financial rewards that included inappropriate patients, unfairly characterizing the physician's and practice's performance. Second, we were not able to explore why there was low agreement for the appropriate response to uncontrolled BP measure. Most providers received credit for this measure because a lifestyle modification was recommended. We hypothesize that this information is documented in the Notes section of the record, but we were not able to systematically evaluate this. Third, given the VA healthcare system is a unique environment with a national electronic health record and a performance measurement reporting system, VA providers are more likely to be aware of documenting clinical information electronically compared to those delivering care within independent small and medium-sized physician practices. While the unique organizational culture of the VA along with its predominantly male and older Veteran population and the salaried (versus fee-for-service) providers impact the generalizability of our findings, the multiple electronic data sources in the VA, especially the VA CDW, reflect current initiatives in health care systems, both at the national and state levels, to use EHR-derived data to evaluate clinical performance and efficiency.

Pay-for-performance plans that use only EHR-derived data to reward providers should carefully consider the measures and the structure of the EHR and other electronic data sources before data collection and financial incentive disbursement. Quality assessment that includes metrics that require several different data elements, such as guideline-recommended medications, and actions that occur during the patient-provider encounter that are not recorded in a structured field are more likely to underestimate provider performance.

Supplementary Material

Supplemental Appendix Table 1
Supplemental Appendix Table 2

Acknowledgments

Funding/Support

The research reported here was supported in part by the U.S. Department of Veterans Affairs, Veterans Health Administration, Health Services Research & Development (HSR&D) Investigator-Initiated Research (IIR) 04-349 (PI Laura A. Petersen, MD, MPH) and the National Institutes of Health (NIH) RO1 HL079173-01 (PI Laura A. Petersen, MD, MPH). Dr. Petersen is the Director at the Center for Innovations in Quality, Effectiveness and Safety (CIN 13-413), Michael E. DeBakey VA Medical Center, Houston, TX. Dr. Petersen was a recipient of the American Heart Association Established Investigator Award (Grant number 0540043N). Dr. Dudley is a Robert Wood Johnson Investigator Awardee in Health Policy. The views expressed are solely of the authors, and do not necessarily represent those of the authors’ institutions.

Footnotes

Complete author information

Tracy H. Urech, MPH; 2002 Holcombe Blvd. (152), Houston, TX, 77030. Phone: 713-794-8601; Fax: 713-748-7359; tracy.urech@va.gov

LeChauncy D. Woodard, MD, MPH; 2002 Holcombe Blvd. (152), Houston, TX, 77030. Phone: 713-440-4441; Fax: 713-748-7359; lwoodard@bcm.edu

Salim S. Virani, MD, PhD; 2002 Holcombe Blvd. (152), Houston, TX, 77030. Phone: 713-440-4410; Fax: 713-748-7359; virani@bcm.edu

R. Adams Dudley, MD, MBA; 333 California Street Suite 265, Box 0936, San Francisco, CA, 94118. Phone: 415-476-8617; Fax: 415-476-0705; adams.dudley@ucsf.edu

Meghan Z. Lutschg, BA; 2002 Holcombe Blvd. (152), Houston, TX, 77030. Phone: 713-794-8601; Fax: 713-748-7359; mlutschg@gmail.com

Laura A. Petersen, MD, MPH; 2002 Holcombe Blvd. (152), Houston, TX, 77030. Phone: 713-794-8623; Fax: 713-748-7359; laura.petersen@va.gov

References

  • 1.Kern LM, Malhotra S, Barron Y, et al. Accuracy of electronically reported “meaningful use” clinical quality measures. Ann Intern Med. 2013;158:77–83. doi: 10.7326/0003-4819-158-2-201301150-00001. [DOI] [PubMed] [Google Scholar]
  • 2.Parsons A, McCullough C, Wang J, et al. Validity of electronic health record-derived quality measurement for performance monitoring. J Am Med Inform Ass. 2012;19:604–609. doi: 10.1136/amiajnl-2011-000557. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3.Gardner W, Morton S, Byron SC, et al. Using computer-extracted data from electronic health records to measures the quality of adolescent well-care. Health Serv Res. 2014;49:1226–1248. doi: 10.1111/1475-6773.12159. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4.Centers for Medicare & Medicaid Services (CMS), HHS Medicare and Medicaid programs: hospital outpatient prospective payment and ambulatory surgical center payment systems and quality reporting programs; Hospital Value-Based Purchasing Program; organ procurement organizations; quality improvement organizations; Electronic Health Records (EHR) Incentive Program; provider reimbursement determinations and appeals. Final rule with comment period and final rules. Fed Regist. 2013;78:74825–75200. [PubMed] [Google Scholar]
  • 5.Adler-Milstein J, DesRoches CM, Furukawa MF, et al. More than half of US hospitals have at least a basic EHR, but stage 2 criteria remain challenging for most. Health Aff (Millwood) 2014;33:1664–1671. doi: 10.1377/hlthaff.2014.0453. [DOI] [PubMed] [Google Scholar]
  • 6. [March 25, 2015];Electronically Specified Clinical Quality Measures (eCQMs) Reporting Quality Net. Available at: http://www.qualitynet.org.
  • 7. [June 23, 2015];Centers for Medicare and Medicaid Services Physician Quality Reporting System (PQRS) Overview. Available at: http://www.cms.gov/Medicare/Quality-Initiatives-Patient-Assessment-Instruments/PQRS/Downloads/PQRS_OverviewFactSheet_2013_08_06.pdf.
  • 8.Hartocollis A. New York City ties doctors’ income to quality of care. New York Times; Jan 12, 2013. [October 3, 2014]. p. A1. Available at: http://www.nytimes.com/2013/01/12/nyregion/new-york-city-hospitals-to-tie-doctors-performance-pay-to-quality-measures.html?pagewanted=all&_r=2&. [Google Scholar]
  • 9.New York City Health and Hospitals Corporation [June 23, 2015];HHC Signs Contract for New Electronic Medical Record System to Span City's Entire Public Hospital System (press release, dated January 16, 2013. Available at: http://www.nyc.gov/html/hhc/html/news/press-release-2013-01-16-medical-records-contract.shtml.
  • 10. [October 3, 2014];Integrated Health Association (IHA) Pay for Performance (P4P) Program Fact Sheet. Available at: http://www.iha.org/pdfs_documents/p4p_california/P4P-Fact-Sheet-September-2014.pdf.
  • 11.Petersen LA, Urech T, Simpson K, et al. Design, rationale, and baseline characteristics of a cluster randomized controlled trial of pay for performance for hypertension treatment: study protocol. Implement Sci. 2011;6:114. doi: 10.1186/1748-5908-6-114. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12.Petersen LA, Simpson K, Pietz K, et al. Effects of individual physician-level and practice-level financial incentives on hypertension care: a randomized trial. JAMA. 2013;310:1042–1050. doi: 10.1001/jama.2013.276303. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.Hysong SJ, Simpson K, Pietz K, et al. Financial incentives and physician commitment to guideline-recommended hypertension management. Am J Manage Care. 2012;18:e378–391. [PMC free article] [PubMed] [Google Scholar]
  • 14.Chobanian AV, Bakris GL, Black HR, et al. National Heart, Lung, and Blood Institute Joint National Committee on Prevention, Detection, Evaluation, and Treatment of High Blood Pressure; National High Blood Pressure Education Program Coordinating Committee. The Seventh Report of the Joint National Committee on Prevention, Detection, Evaluation, and Treatment of High Blood Pressure: the JNC 7 report. JAMA. 2003;289:2560–2572. doi: 10.1001/jama.289.19.2560. [DOI] [PubMed] [Google Scholar]
  • 15.Brown SH, Lincoln MJ, Groen PJ, et al. VistA—U.S. Department of Veterans Affairs national-scale HIS. Int J Med Inform. 2003;69:135–156. doi: 10.1016/s1386-5056(02)00131-4. [DOI] [PubMed] [Google Scholar]
  • 16.Department of Veterans Affairs Office of Enterprise Development [August 28, 2014];VistA-HealtheVet Monograph. Available at: http://www.va.gov/vista_monograph/docs/2008_2009_vistahealthevet_monograph_fc_0309.pdf.
  • 17.Fihn SD, Francis J, Clancy C, et al. Insights from advanced analytics at the Veterans Health Administration. Health Aff. 2014;33:1203–1211. doi: 10.1377/hlthaff.2014.0054. [DOI] [PubMed] [Google Scholar]
  • 18.Hynes DM, Perrin RA, Rappaport S, et al. Informatics resources to support health care quality improvement in the Veterans Health Administration. J Am Med Inform Assoc. 2004;11:344–350. doi: 10.1197/jamia.M1548. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19.Smith MW, Joseph GJ. Pharmacy data in the VA health care system. Med Care Res Rev. 2003;60:92S–123S. doi: 10.1177/1077558703256726. [DOI] [PubMed] [Google Scholar]
  • 20.Department of Veterans Affairs Health Economics Resource Center (HERC) [March 25, 2015];Non-VA medical care (formerly fee basis files) Available at: http://www.herc.research.va.gov/include/page.asp?id=non-va.
  • 21.Department of Veterans Affairs [September 15, 2014];VHA Directive 2010-019: Access to Centers for Medicare and Medicaid Services (CMS) data for VHA users within the VA information technology (IT) systems. Available at: http://www.va.gov/vhapublications/ViewPublication.asp?pub_ID=2228.
  • 22.Department of Veterans Affairs [September 3, 2014];VHA Directive 2008-084: National Clinical Reminders. http://www.va.gov/vhapublications/ViewPublication.asp?pub_ID=1815.
  • 23.Cicchetti DV, Feinstein AR. High agreement but low kappa: II. Resolving the paradoxes. J Clin Epidemiol. 1990;43:551–558. doi: 10.1016/0895-4356(90)90159-m. [DOI] [PubMed] [Google Scholar]
  • 24.Chen G, Faris P, Hemmelgarn B, et al. Measuring agreement of administrative data with chart data using prevalence unadjusted and adjusted kappa. BMC Med Res Methodol. 2009;9:5. doi: 10.1186/1471-2288-9-5. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25.Barbara AM, Loeb M, Dolovich L, et al. Patient self-report and medical records: Measuring agreement for binary data. Can Fam Physician. 2011;57:737–738. [PMC free article] [PubMed] [Google Scholar]
  • 26.Cohen J. A coefficient of agreement for nominal scales. Edu and Psychol Meas. 1960;20:37–46. [Google Scholar]
  • 27.Landis JR, Koch GG. The measurement of observer agreement for categorical data. Biometrics. 1977;33:159–174. [PubMed] [Google Scholar]
  • 28.Persell SD, Wright JM, Thompson JA, et al. Assessing the validity of national quality measures for coronary artery disease using an electronic health record. Arch Intern Med. 2006;166:2272–2277. doi: 10.1001/archinte.166.20.2272. [DOI] [PubMed] [Google Scholar]
  • 29.Borzecki AM, Wong AT, Hickey EC, et al. Can we use automated data to assess quality of hypertension care? Am J Manag Care. 2004;10:473–479. [PubMed] [Google Scholar]
  • 30.Baker DW, Persell SD, Thompson JA, et al. Automated review of electronic health records to assess quality of care for outpatients with heart failure. Ann Intern Med. 2007;146:270–277. doi: 10.7326/0003-4819-146-4-200702200-00006. [DOI] [PubMed] [Google Scholar]
  • 31.Garrido T, Kumar S, Lekas J, et al. e-Measures: insight into the challenges and opportunities of automating publicly reported quality measures. J Am Med Inform Assoc. 2014;21:181–184. doi: 10.1136/amiajnl-2013-001789. [DOI] [PMC free article] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Supplemental Appendix Table 1
Supplemental Appendix Table 2

RESOURCES