Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2018 Dec 13.
Published in final edited form as: J Am Coll Cardiol. 2016 May 24;67(20):2441–2442. doi: 10.1016/j.jacc.2016.03.511

Assessment of Administrative Data to Identify Acute Myocardial Infarction in Electronic Health Records

Robert J Mentz 1,2,*, L Kristin Newby 1,2, Ben Neely 2, Joseph E Lucas 2, Sean D Pokorney 1, Meena P Rao 1, Larry R Jackson II 2, Maria V Grau-Sepulveda 2, Michelle M Smerek 2, Pamela Barth 3, Charlotte L Nelson 2, Michael J Pencina 2, Bimal R Shah 4
PMCID: PMC6292432  NIHMSID: NIHMS996462  PMID: 27199066

To improve the efficiency of clinical research, there is interest in using electronic health record (EHR) data for outcome ascertainment (1). However, EHR data are designed for clinical care and billing and their utility for event capture is largely untested (2,3). We investigated 4 published computable phenotype definitions on the basis of administrative data for acute myocardial infarction (AMI) compared with a gold standard AMI classification created through chart review. The 4 algorithms were published by:1) the Agency for Healthcare Research and Quality;2) Centers for Medicare & Medicaid Services (CMS) Chronic Conditions Warehouse; 3) CMS Quality Net; and 4) Mini-Sentinel (Table 1).

TABLE 1.

Sensitivity and Specificity of Computable Phenotype Algorithms for AMI Compared With AMI Classification on the Basis of the Third Universal Definition

Administrative Algorithm Algorithm Components Sensitivity (95% CI) Specificity (95% CI)
Agency for Healthcare Research and Quality
  • Inpatient claims/principal (first position) discharge or final diagnosis

  • ICD-9 codes: 410.X, 410.XX

0.85 (0.81–0.89) 0.99 (0.99–1.00)
Centers for Medicare and Medicaid
Services (CMS) Chronic
Conditions Warehouse
  • Inpatient claims/first or second position in discharge or final diagnosis

  • ICD-9 codes: 410.X1

0.99 (0.98–1.00) 0.99 (0.99–1.00)
CMS Quality Net
  • Inpatient claims/principal (first position) discharge or final diagnosis/LOS >1 day

  • ICD-9 codes: 410.X1, 410.X0

0.83 (0.79–0.87) 0.99 (0.99–1.00)
Mini-Sentinel
  • Inpatient claims/principal (first position) discharge or final diagnosis, ICD-9 codes: 410.X1, 410.X0

  • Emergency room claim/death within 1 day/principal (first position) discharge or final diagnosis, ICD-9 codes: 410.X1, 410.X0, 411.1, 411.8, 411.8X, 413.X

0.79 (0.74–0.84) 0.99 (0.99–1.00)

AMI = acute myocardial infarction; CI = confidence interval; CMS = Centers for Medicare and Medicaid Services; ICD-9 = International Classification of Diseases, Ninth Revision; LOS = length of stay.

For the present analysis, 150 encounters at Duke between January 2010 and July 2011 were selected for review via stratified random sampling on the basis of all 4 definitions. The population included Durham County residents age ≥18 years. After implementation of the algorithms, the population was stratified according to 1 of 3 criteria: 1) those encounters for which all 4 phenotypes indicated that an AMI occurred; 2) 1 to 3 phenotypes indicated that an AMI event occurred; and 3) all 4 phenotypes indicated no AMI occurred. The sample size was calculated on the basis of assumptions about positive predictive value and 1 – negative predictive value within each strata and was chosen to minimize the variance estimate of sensitivity (4). Cardiologist reviewer pairs independently assessed AMI events on the basis of the Third Universal Definition of Myocardial Infarction (UDMI) (5) through chart review. Reviewers were blinded to algorithm output. Disagreements were resolved by a third adjudicator. Following adjudication, the true AMI status was established for each encounter against which the administrative algorithms were evaluated. Inter-rater agreement was assessed using Fleiss’s kappa statistics. The sensitivity/specificity and 95% confidence intervals of each definition compared with chart review were calculated (4). Sensitivity/specificity estimates were computed on the basis of proportions in all records during the sampling period.

Of the 150 encounters, 70 (47%) were an AMI by the UDMI and 79 (53%) did not meet the UDMI definition; the specific encounter could not be located for 1 chart. Overall, the physician chart reviewers agreed on the assessment of AMI 100% of the time (kappa = 1) in the no positive phenotype stratum. There was 72% agreement in the strata in which 1 to 3 phenotypes were positive (kappa = 0.51), and 78% agreement in the strata in which all phenotypes were positive (kappa = 0.42). When assessing which components of the UDMI were most strongly associated with disagreements, we found that cardiac biomarker changes and symptoms of ischemia were most strongly associated with disagreements compared with electrocardiogram changes, imaging evidence, or cardiac death/autopsy findings of AMI.

Of the 70 AMIs by chart review, all algorithms were positive for 19 (27%) encounters, and 1 to 3 algorithms were positive for 51 (73%). Table 1 presents the sensitivity/specificity of each algorithm for AMI identification compared with the UDMI.

In summary, the CMS Chronic Conditions Warehouse was found to have the highest sensitivity and a high specificity via assessment of inpatient claims with the International Classification of Diseases, Ninth Revision codes of 410.X1 in the first or second position for discharge or final diagnosis. The high sensitivity supports its use to rule out an AMI in those with negative administrative data, and the high specificity rules in an AMI event in those with positive administrative data. Importantly, these findings were observed in a single-center analysis of Durham County residents. We did not specifically document details related to completeness of electrocardiography data or subsequent revascularization.

Importantly, prior analyses demonstrated that classification of nonfatal events in routinely-recorded EHR-type data is imperfect (2). EHRs are heterogeneous, and data elements are inconsistently defined(3). The complexity of the healthcare system can affect the utility of the EHR for outcome ascertainment, as patients receive care from multiple providers and institutions. An improved understanding of how to make administrative algorithms more accurate would facilitate outcomes research. Future research is needed to validate these observations and investigate other endpoints including revascularization, stroke, and heart failure events.

Acknowledgments

The projects and the work described in this research letter are supported in part by grant number 1C1CMS331018-01-00 from the Department of Health and Human Services, Centers for Medicare & Medicaid Services, and in part by the Bristol-Myers Squibb Foundation Together on Diabetes program, respectively. Dr. Mentz receives research support from the National Institutes of Health (U10HL110312). Dr. Newby has received research funding from PCORI and the National Institutes of Health; has received honoraria from JACC: Basic to Translational Science and the Journal of the American Heart Association; has served as a consultant to Roche Diagnostics, Philips Health-care, Metanomics, Merck, Inc., and BioKier; has served on the advisory board of MedScape/theHeart.org; has performed research for Metanomics, Verily (formerly Google Life Sciences), GlaxoSmithKline, Amylin/Bristol-Myers Squibb, and Sanofi; and has served on the data and safety monitoring board of DemeRx. Dr. Pokorney has received modest research or education grants from Gilead, Boston Scientific, and AstraZeneca; and has received modest consulting support from Boston Scientific and Medtronic. Dr. Rao has received a research grant from Medtronic. Dr. Shah is employed by Premier Inc. All other authors have reported that they have no relationships relevant to the contents of this paper to disclose.

REFERENCES

  • 1.Roe MT, Mahaffey KW, Ezekowitz JA, et al. The future of cardiovascular clinical research in North America and beyond—addressing challenges and leveraging opportunities through unique academic and grassroots collaborations. Am Heart J 2015;169:743–50. [DOI] [PubMed] [Google Scholar]
  • 2.Barry SJ, Dinnett E, Kean S, et al. Are routinely collected NHS administrative records suitable for endpoint identification in clinical trials? Evidence from the West of Scotland Coronary Prevention Study. PloS One 2013;8:e75379. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3.Bayley KB, Belnap T, Savitz L, et al. Challenges in using electronic health record data for CER: experience of 4 learning organizations and solutions applied. Med Care 2013;51:S80–6. [DOI] [PubMed] [Google Scholar]
  • 4.Begg CB, Greenes RA. Assessment of diagnostic tests when disease verification is subject to selection bias. Biometrics 1983;167:565–9. [PubMed] [Google Scholar]
  • 5.Thygesen K, Alpert JS, Jaffe AS, et al. Third universal definition of myocardial infarction. J Am Coll Cardiol 2012;60:1581–98. [DOI] [PubMed] [Google Scholar]

RESOURCES