Skip to main content
HHS Author Manuscripts logoLink to HHS Author Manuscripts
. Author manuscript; available in PMC: 2025 Sep 30.
Published in final edited form as: Qual Manag Health Care. 2024 Sep 30;33(4):289–290. doi: 10.1097/QMH.0000000000000492

The Development and Endorsement of a Performance Measure for Stroke Misdiagnosis in the Emergency Department

J Matthew Austin 1, Yuxin Zhu 2, Krisztian Sebestyen 3, Elizabeth A Fracica 4, David E Newman-Toker 5
PMCID: PMC11733821  NIHMSID: NIHMS2043589  PMID: 39724860

THE NEED FOR DIAGNOSTIC QUALITY PERFORMANCE MEASURES

Diagnostic error remains a problem in public health.1 Three major disease categories—vascular events, infections, and cancer—account for three-fourths of all serious harms from diagnostic error,2 with missed stroke the leading cause of serious harm to patients.3 The misdiagnosis of stroke disproportionately occurs when patients present with symptoms that are non-typical or obvious,4,5 such as dizziness or vertigo, which can be mistaken for inner ear disease.4 Annually in the United States, an estimated 45 000–75 000 patients that present to the emergency department (ED) with dizziness or vertigo caused by stroke are misdiagnosed and erroneously discharged.5

ED patients with dizziness and vertigo could be correctly diagnosed as having a stroke if evidence-based interventions are followed,6,7 but there remains a large evidence-to-practice gap.8 Without a timely and accurate diagnosis, these patients suffer misdiagnosis-related harms9 because they do not receive prompt treatment for this time-sensitive condition.4 The lack of operationally viable performance measures remains a barrier for hospitals to improve their diagnostic quality.10,11

AVOID H.A.R.M.—ED STROKE/DIZZINESS MEASURE

To address this gap, the team at the Armstrong Institute Center for Diagnostic Excellence developed the Avoid Hospitalization After Release with a Misdiagnosis (Avoid H.A.R.M.)—ED Stroke/Dizziness quality measure, which tracks the rate of missed strokes in the ED. Using the Symptom-disease Pair Analysis of Diagnostic Error (SPADE) framework,12 the claims-based measure tracks the rate of adult patients treated and released from the ED with either a non-specific, presumed benign symptom-only dizziness diagnosis or a specific inner ear/vestibular diagnosis (collectively referred to as “benign dizziness”) who were subsequently admitted to a hospital for a stroke within 30 days of their ED visit. The measure accounts for the epidemiological base rate of stroke in the population under study using a risk difference approach, comparing the short-term (0–30 days) and long-term (91–360 days) incidence rates of stroke.

MEASURE ENDORSEMENT

The measure was put through the Centers for Medicare and Medicaid Services (CMS) consensus-based entity measure endorsement process, undergoing rigorous reviews by multistakeholder committees against established criteria. Most performance measures used in accountability programs at a national level are endorsed. The endorsement process evaluates measures on four criteria: Importance to Measure, Scientific Acceptability, Feasibility, and Use and Usability.

While the measure was eventually endorsed, our path to endorsement was not straightforward. At two different points in the process, we encountered concerns from the committees charged with evaluating the first two criteria. We needed to submit and resubmit the measure three times before final endorsement.

To establish the Importance to Measure, we needed to demonstrate a relationship between an existing health care structure or process and the outcome of interest. We initially provided evidence that providers trained in eye-movement-based bedside diagnosis of stroke using the head impulse, nystagmus, test of skew (HINTS) exam had improved identification of stroke compared to standard imaging approaches. Use of these bedside diagnostic maneuvers is endorsed by the Society for Academic Emergency Medicine as current best practice.13 The committee reviewing this criterion expressed concerns about the limited number of providers currently trained on HINTS, challenges with scaling training, and eventually discounted HINTS as insufficient to produce a difference in reducing misdiagnosed strokes. In response, we submitted data that examined the relationship between a hospital’s use of neuroimaging in patients with “benign dizziness” and the hospital’s performance on the Avoid H.A.R.M measure. We provided comparisons of measure-based outcomes when hospitals used MRI scans versus CT scans, as MRIs are more sensitive for acute strokes in the back part of the brain than CT.

For the Scientific Acceptability criterion, we needed to demonstrate the measure’s validity and reliability. To demonstrate the reliability, we presented a signal-to-noise analysis across all measured entities derived from the full Medicare Fee-For-Service (FFS) dataset for a four-year period. The committee raised concerns about the presented data, both the measure’s signal and its ability to discriminate high-performing from low-performing hospitals. We reran our analyses with a dataset that included a fuller group of patients at a hospital, as Medicare FFS beneficiaries only reflect approximately 20% of a hospital’s total ED visits. For the updated analysis we used the Agency for Healthcare Research and Quality’s Healthcare Cost and Utilization Project (HCUP) data for Florida hospitals, which includes all patients seen at a hospital. With this greater capture of patients, our statistics improved and we were able to better demonstrate the measure’s signal and discrimination.

The respective committees found our updated submissions to meet the criteria and the measure became the first diagnosis-related outcome measure to be endorsed.

OUR LEARNING

The endorsement process was helpful in our work to advance diagnosis-related performance measurement. Our experience could inform others in two primary ways: First, the committee reviewing the Importance to Measure criterion wanted wide-spread and scalable approaches that hospitals could use to improve diagnostic performance, so as others think about developing diagnosis-related outcome measures, having clear, scalable improvement interventions is important. Second, for us to demonstrate the Scientific Acceptability criterion, we needed to employ datasets that were more encompassing than what Medicare FFS alone could offer. We have incorporated this learning into choosing which datasets we’ll use moving forward for this work.

NEXT STEPS

Our next phase focuses on the use of the measure in accountability programs. We are speaking with states, payers, and health care quality organizations about how the measure could be used in their programs. Although endorsement of the measure is helpful for a measure’s utility, endorsement does not guarantee the measure’s real-world use.

The endorsement process highlighted that the uptake of the measure would benefit from the accompaniment of targeted “balancing” measures. For example, concerns were raised during the endorsement process that diagnostic quality measures might prompt inappropriate increases in testing. Specific to the ED-stroke/dizziness measure, there were concerns about the increased use of neuroimaging for “benign dizzy” patients. It would be ideal to create companion measures that track CT and MRI utilization in patients with “benign dizziness” to ensure that there are no unintended consequences of inappropriate imaging.

Acknowledgments:

The authors thank Shannon L. Cole, PhD, and Megan Clark, MWC, from the Johns Hopkins Armstrong Institute for Patient Safety and Quality for their thoughtful reviews and editing of this paper.

Funding and Disclosures:

This research was funded by US Agency for Healthcare Research and Quality (AHRQ) grant R01HS027614 and also supported by AHRQ grant R18HS029350, the Gordon and Betty Moore Foundation grant #5756, and the Armstrong Institute Center for Diagnostic Excellence.

Footnotes

The authors declare no conflicts of interest.

Contributor Information

J. Matthew Austin, Johns Hopkins Armstrong Institute for Patient Safety and Quality, Johns Hopkins University School of Medicine, Baltimore, MD.

Yuxin Zhu, Johns Hopkins Armstrong Institute for Patient Safety and Quality, Johns Hopkins University School of Medicine, Baltimore, MD.

Krisztian Sebestyen, Johns Hopkins Armstrong Institute for Patient Safety and Quality, Johns Hopkins University School of Medicine, Baltimore, MD.

Elizabeth A. Fracica, Department of Neurology, Harvard University, Boston, MA.

David E. Newman-Toker, Department of Neurology, Johns Hopkins Armstrong Institute for Patient Safety and Quality, Johns Hopkins University School of Medicine, Baltimore, MD.

REFERENCES

  • 1.National Academies of Sciences, Engineering, and Medicine. Improving Diagnosis in Health Care. The National Academies Press; 2015:472. [Google Scholar]
  • 2.Newman-Toker DE, Tucker L, et al. Roadmap for Research to Improve Diagnosis, Part 1: Converting National Academy of Medicine Recommendations into Policy Action. Society to Improve Diagnosis in Medicine. 2018. https://www.improvediagnosis.org/page/roadmap [Google Scholar]
  • 3.Newman-Toker DE, Nassery N, Schaffer AC, et al. Burden of serious harms from diagnostic error in the USA. BMJ Qual Saf. 2024;33:109–120. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4.Tarnutzer AA, Lee SH, Robinson KA, Wang Z, Edlow JA, Newman-Toker DE. ED misdiagnosis of cerebrovascular events in the era of modern neuroimaging: A meta-analysis. Neurology. 2017;88:1468–1477. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5.Newman-Toker DE. Missed stroke in acute vertigo and dizziness: It is time for action, not debate. Ann Neurol. 2016;79:27–31. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6.Tarnutzer AA, Berkowitz AL, Robinson KA, Hsieh YH, Newman-Toker DE. Does my dizzy patient have a stroke? A systematic review of bedside diagnosis in acute vestibular syndrome. CMAJ. 2011;183:E571–92. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7.Newman-Toker DE, Edlow JA. TiTrATE: A Novel, Evidence-Based Approach to Diagnosing Acute Dizziness and Vertigo. Neurol Clin. 2015;33:577–99. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.Kerber KA, Newman-Toker DE. Misdiagnosing dizzy patients: common pitfalls in clinical practice. Neurol Clin. 2015;33:565–75. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9.Newman-Toker DE, Pronovost PJ. Diagnostic errors--the next frontier for patient safety. JAMA. 2009;301:1060–2. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10.Henriksen K, Dymek C, Harrison MI, Brady PJ, Arnold SB. Challenges and opportunities from the Agency for Healthcare Research and Quality (AHRQ) research summit on improving diagnosis: a proceedings review. Diagnosis (Berl). 2017;4:57–66. [DOI] [PubMed] [Google Scholar]
  • 11.National Quality Forum. Improving Diagnostic Quality and Safety. 2017. https://www.qualityforum.org/Projects/i-m/Improving_Diagnostic_Accuracy/Final_Report.aspx
  • 12.Liberman AL, Newman-Toker DE. Symptom-Disease Pair Analysis of Diagnostic Error (SPADE): a conceptual framework and methodological approach for unearthing misdiagnosis-related harms using big data. BMJ Qual Saf. 2018;27:557–566. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.Edlow JA, Carpenter C, Akhter M, et al. Guidelines for reasonable and appropriate care in the emergency department 3 (GRACE-3): Acute dizziness and vertigo in the emergency department. Acad Emerg Med. 2023;30:442–486. [DOI] [PubMed] [Google Scholar]

RESOURCES