Abstract
Background
The public health burden associated with diagnostic errors is likely enormous, with some estimates suggesting millions of individuals are harmed each year in the USA, and presumably many more worldwide. According to the US National Academy of Medicine, improving diagnosis in healthcare is now considered ‘a moral, professional, and public health imperative.’ Unfortunately, well-established, valid and readily available operational measures of diagnostic performance and misdiagnosis-related harms are lacking, hampering progress. Existing methods often rely on judging errors through labour-intensive human reviews of medical records that are constrained by poor clinical documentation, low reliability and hindsight bias.
Methods
Key gaps in operational measurement might be filled via thoughtful statistical analysis of existing large clinical, billing, administrative claims or similar data sets. In this manuscript, we describe a method to guantify and monitor diagnostic errors using an approach we call ‘Symptom-Disease Pair Analysis of Diagnostic Error’ (SPADE).
Results
We first offer a conceptual framework for establishing valid symptom-disease pairs illustrated using the well-known diagnostic error dyad of dizziness-stroke. We then describe analytical methods for both look-back (casecontrol) and look-forward (cohort) measures of diagnostic error and misdiagnosis-related harms using ‘big data’. After discussing the strengths and limitations of the SPADE approach by comparing it to other strategies for detecting diagnostic errors, we identify the sources of validity and reliability that undergird our approach.
Conclusion
SPADE-derived metrics could eventually be used for operational diagnostic performance dashboards and national benchmarking. This approach has the potential to transform diagnostic quality and safety across a broad range of clinical problems and settings.
INTRODUCTION
According to the US National Academy of Medicine (NAM), diagnostic errors represent a major public health problem likely to affect each of us in our lifetime.1 The 2015 NAM report, Improving Diagnosis in Healthcare, goes on to state that, ‘improving the diagnostic process is not only possible, but it also represents a moral, professional, and public health imperative.’1 Annually in the USA, there may be more than 12 million diagnostic errors2 with one in three such errors causing serious patient harm.3 The aggregate annual costs to the US healthcare system could be as high as US$100–US$500 billion.4 The global problem is likely even bigger.5–8
Diagnostic errors represent the ‘bottom of the iceberg’ of patient safety—a hidden, yet large, source of morbidity and mortality. Valid operational measures are badly needed to surface this problem so that it can be quantified, monitored and tracked.9 Existing measures of diagnostic error that rely on manual chart review to confirm diagnostic errors suffer from problems of poor chart documentation,10,11 low inter-rater reliability,12,13 hindsight bias14 and the high costs of human labour needed for chart abstraction. Additionally, reliance on chart review alone will likely lead to an underestimation of diagnostic error since key clinical features necessary to identify errors are preferentially missing from charts where errors occur.15,16 We believe that key gaps in operational measures of diagnostic error can be filled via thoughtful statistical analysis of large clinical (electronic health record (EHR)) and administrative (billing, insurance claims) data sets.
In this manuscript, we describe a novel conceptual framework and methodological approach to measuring diagnostic quality and safety using ‘big data’: Symptom-Disease Pair Analysis of Diagnostic Error (SPADE). We illustrate our approach predominantly using a single well-studied example (dizziness-stroke), but provide evidence that SPADE could be used to develop a scientifically valid set of diagnostic performance metrics across a broad range of conditions.
DIAGNOSTIC ERROR AND MISDIAGNOSIS-RELATED HARM DEFINITIONS
The NAM defines diagnostic error as failure to (A) establish an accurate and timely explanation of the patient’s health problem(s) or (B) communicate that explanation to the patient.1 Harms resulting from the delay or failure to treat a condition actually present (false-negative diagnosis) or from treatment provided for a condition not actually present (false-positive diagnosis) are known as misdiagnosis-related harms.17,18 A key feature of the NAM definition is that it does not require the presence of a diagnostic process failure (eg, failure to perform a specific diagnostic test)17 nor that the error could have been prevented. This patient-centred definition is agnostic as to the correctness of the diagnostic processes; it relies only on the outcome of a patient receiving an inaccurate or delayed diagnosis as opposed to an accurate and timely diagnosis.1
The SPADE approach, described in detail below, uses unexpected adverse health events (eg, stroke, myocardial infarction (MI), death) to measure misdiagnosis-related harms.19–25 SPADE methods maintain core consistency with the NAM definition of diagnostic error by identifying inaccurate or delayed diagnoses, regardless of cause or preventability. Although SPADE does not specifically address communication with patients (part ‘B’ of the NAM definition), if failure to communicate a diagnosis to a patient results in a clinically relevant and harmful health event (ie, misdiagnosis-related harm), the SPADE approach will detect it. A key advantage of this approach is that using ‘hard’ clinical outcomes avoids much of the subjectivity12–14 inherent in other methods that rely on detailed, human medical record reviews to assess for errors.
THE SYMPTOM-DISEASE PAIR FRAMEWORK FOR MEASUREMENT
The SPADE approach is premised on three principles: (1) patients with symptoms seek medical attention; (2) the object of the medical diagnostic process is to identify the underlying cause (ie, the condition(s) responsible for the patient’s symptom(s)); and (3) failure to correctly diagnose the underlying disease(s) in a timely manner (NAM-defined diagnostic error) may be followed by illness progression that might have been avoided through prompt diagnosis and treatment (preventable misdiagnosis-related harm). In this approach, we combine what is known about disease natural history and pathophysiology to develop an inferential model for identifying misdiagnosis-related harms based on time-linked markers of diagnostic delay that are clinically sensible, biologically plausible and specific to symptom-disease pairs (figure 1).
Symptom-disease pairs that may be ‘diagnostic error dyads’ can be analysed using either a ‘look-back’ or a ‘look-forward’ approach (figure 2). The look-back approach takes an important disease and identifies which clinical presentations of that disease are most likely to be missed. The look-forward approach takes a common symptom and identifies which important diseases are likely to be missed among patients who present with this symptom. When little is known about misdiagnosis of a particular disease, a look-back analysis helps identify promising targets to establish one or more diagnostic error dyads. Once one or more diagnostic error dyads are established, a look-forward analysis can be performed to measure real-world performance.
THE SPADE APPROACH
The SPADE approach relies on having information from at least two discrete points in time. The first time point is an ‘index’ diagnosis and the second time point is an ‘outcome’ diagnosis (figure 1). The outcome diagnosis must plausibly link back to symptoms or signs from the index visit (and diagnosis) yet be unexpected or improbable if the index diagnosis had been correct. The most common and straightforward diagnostic error scenario is one in which an ambulatory index visit (eg, primary care or emergency department (ED)) results in a discharge for a supposedly benign disorder (treat-and-release visit) and a subsequent outcome visit or admission discloses otherwise. For example, the occurrence of an adverse outcome (eg, hospitalisation for a newly diagnosed stroke, MI or sepsis) shortly after a treat-and-release ED visit with a benign diagnosis rendered is a strong indicator of diagnostic error with misdiagnosis-related harm (assuming similar symptoms or signs are associated with both the benign and dangerous diseases).
For illustrative purposes, we will use the case of a patient seen in the ED with a chief complaint of dizziness diagnosed as a benign inner ear condition, but who has dangerous cerebral ischaemia as the true cause of her symptoms.26,27 Imagine we are unsure of whether this symptom-disease pair (dizziness-stroke) is a real dyad26,28 or merely coincidental. We would note that, biologically speaking, dizziness/vertigo can be a manifestation of minor stroke or transient ischaemic attack (TIA).29 With untreated TIA and minor stroke, there is a marked increased short-term risk of major stroke in the subsequent 30 days that tapers off by 90 days.29–31 A clinically relevant and statistically significant temporal association between ED discharge for supposedly ‘benign’ vertigo followed by a stroke diagnosis within 30 days is therefore a biologically plausible marker of diagnostic error.21 If this missed diagnosis of cerebral ischaemia resulted in a clinically meaningful adverse health outcome (eg, stroke hospitalisation), this would suggest misdiagnosis-related harm.
The association of treat-and-release visits for ‘benign’ vertigo and subsequent hospitalisations for stroke can readily be measured using information collected in administrative claims or large EHR data sets.21,22,25 We can employ a bidirectional analysis (figure 3). Using the look-back method, we start with a disease cohort of hospitalised patients with stroke and look back in time to prior treat-and-release ED visits for vertigo.25 We analyse the observed to expected treat-and-release visit frequency and temporal distribution of such visits during a reasonable time window. We employ positive (headache) and negative (abdominal/back pain) symptom controls, finding that vertigo is the most over-represented prestroke admission treat-and-release ED visit (figure 3A).25 Using the look-forward method, we start with a vertigo symptom cohort of discharged ED patients and look forward in time to subsequent stroke admissions. We can employ positive (intracerebral haemorrhage) and negative (MI) disease controls, finding that only short-term cerebrovascular event rates are elevated above the base rate, suggesting that a ‘benign’ vertigo discharge is a meaningful risk factor for missed stroke but not missed MI (figure 3B).21
Together, these analyses statistically support the symptom-disease pair of dizziness-stroke and create strong inferential evidence of an index visit diagnostic error (incorrect diagnosis of benign vertigo rendered) with subsequent misdiagnosis-related harms (worsening or recurrent cerebral ischaemia necessitating hospitalisation). Specific analyses that can be used to establish major aspects of validity and reliability for SPADE are shown in table 1.32–34 Key among these are: (1) the bidirectional relationship in an overlapping temporal profile, which establishes convergent construct validity of the association and a link to biological plausibility34; and (2) the use of negative control comparisons which establishes discriminant construct validity and makes it highly improbable that patients discharged from the ED merely have an elevated short-term risk of all adverse medical events (ie, are non-specifically ‘sick’). These statistical methods highlight the fact that valid measures of diagnostic error need not be exclusively derived from traditional approaches such as chart review, survey data or prospective studies.
Table 1.
Concept | Method | Validity and reliability |
---|---|---|
Symptom-disease pair | Test an association that is clinically plausible, linking a presenting symptom (chief complaint) and specific disease.28 | Face validity and biological plausibility of the target symptom-disease dyad |
Bidirectional analysis | Use both look-back and look-forward methods to assess the same symptom-disease association.21,25 | Convergent construct validity of the symptomdisease dyad |
Baseline comparisons (observed to expected) | Compare event frequency or rate of return with baseline or expected level (look back—OR25; look forward—HR21) or matched control population.23 | Strength of measured association relative to internal or external control |
Temporal profiles | Plot temporal profile or trend (look back—time before index event25; look forward—time after index event21). | Temporality and biological plausibility/gradient based on disease natural history |
Positive control comparisons | Test a similar association that is clinically plausible (look back—linked symptom19,25; look forward—linked disease35). | Coherence of the symptom-disease dyad or alternative form reliability |
Negative control comparisons | Test an association that is not clinically plausible (look back—unlinked symptom25; look forward—unlinked disease21,22). | Discriminant construct validity (specificity) of the symptom-disease dyad |
Subgroup analyses | Test for clinically plausible subgroup associations (eg, dizziness linked to missed ischaemic but not haemorrhagic stroke; headache linked to both25). | Face validity and biological plausibility/gradient o the measured associations |
Associated diagnostic process failures | Correlate specific outcomes with known process failures (eg, missed stroke linked to improper use of CT rather than MRI51). | Coherence of the identified associations and construct validity |
Triangulation of findings | Use alternative methods (eg, chart review, surveys, root cause analyses) to confirm the diagnostic error association.26 | Convergent construct validity, coherence of the measured associations |
Impact analysis | Monitor the impact of interventions designed to reduce error or harms on the measure (‘flattening the hump’). | Predictive (criterion) validity and measure responsiveness |
Reproducibility of analytical results | Repeat the analysis in multiple data sets21–23 or using resampling methods (eg, bootstrapping or split-halves). | Consistency of the measured associations or resampling73 reliability |
Reproducibility of SPADE approach | Repeat the approach across other analogous symptomdisease dyads (eg, chest pain-myocardial infarction,24 fever-meningitis/sepsis36). | Analogy34 of the approach to related problems |
HR, hazard ratio; OR, odds ratio; SPADE, Symptom-Disease Pair Analysis of Diagnostic Error.
OPTIMAL MEASUREMENT CONTEXT FOR SPADE
Disease types and analytical approach
The SPADE method should apply to any condition where the short-term risk of worsening or recurrence is high. SPADE has been used for other symptoms and signs tied to missed stroke (headache-aneurysmal subarachnoid haemorrhage19; facial weakness-ischaemic stroke35); to missed cardiovascular events (eg, chest pain-MI)20,24; and to missed infections (eg, fever-men-ingitis/sepsis36; Bell’s palsy-acute otitis35). Since missed vascular events and infections together account for at least one-third of all misdiagnosis-related harms,37–40 using SPADE to monitor and track such errors would represent a major advance for the field.
SPADE can be used to assess a single symptom tightly linked to a single disease (headache-aneurysm,19 syncope-pulmonary embolus41), but can also be used to measure multiple related symptoms or diseases. For example, if multiple symptoms are associated with a target disease (eg, chest pain, shortness of breath, abdominal pain and syncope for MI), the symptoms may be bundled together in the analysis.20 Likewise, if a single symptom is associated with multiple target diseases (eg, fever for meningitis, toxic shock and sepsis), the diseases may be bundled together in the analysis.36 As proof of concept, a recent SPADE-style analysis of over 10 million ED discharges used multiple symptoms-to-disease mappings to identify misdiagnosis (figure 4).42
Some diseases are less well suited to SPADE. For example, chronic diseases for which the risk of misdiagnosis-related harms is either constant or very slowly increasing over time (eg, diabetes, hypertension) will make patterns of diagnostic error difficult to discern via SPADE. For diseases with a subacute time course presenting non-specific symptoms (eg, tuberculosis8 and cancer43), a more complex analytical approach is required. For example, it might be necessary to bundle symptoms and combine with visit/test ordering patterns over time (eg, increased odds of general practitioner visits for new complaints/tests in the 6 months before a cancer diagnosis43).
Ideal data sets
Large enough data sets are needed to draw statistically valid inferences. Most prior studies using aspects of the SPADE approach have examined data sets containing 20 000–190 000 visits to identify misdiagnosis-related harm rates of ~0.2%–2%.21,25,35 From a statistical standpoint, the total number of diagnostic error-related outcome events (eg, admissions) should ideally not be fewer than 50–100, so this implies minimal sample sizes of 5000–50 000 visits for event rates in the 0.2%–2% range. Thus, even for common symptoms or diseases, data must generally be drawn from a large health system or region over a short period (eg, 6 months) or a small health system or hospital over a longer period (eg, 5 years). Constraints on the spatial and temporal resolution of SPADE make it unlikely that this approach could be used for provider-level feedback. This constraint, however, relates to the frequency of harm, not the SPADE method—in other words, any method that assesses infrequent harms will have to draw from a large sample.
Data sets that include ‘out-of-network’ follow-up provide the most robust estimates of diagnostic error, avoiding the problem of hospital crossover (ie, patient goes to one centre at the index visit but returns to an unaffiliated centre at the outcome visit). In a 1-year study of crossover in ED populations across five health systems, 25% of patients who revisited crossed over.44 In a large study of missed subarachnoid haemorrhage in the ED that used regional health data, hospital crossover occurred in 37% of misdiagnosed patients.19 Taken together, these data suggest that patients who are misdiagnosed may be disproportionately likely to cross over. Thus, SPADE will likely provide the strongest inferences when used with data sets that include crossovers (eg, regional health information exchanges like the Chesapeake Regional Information System for our Patients) or from health systems with integrated insurance plans where patients are tracked when they use outside healthcare facilities (eg, Kaiser Permanente22). Nevertheless, even without data on crossovers, health systems can still track error rates over time—measured rates may be lower than the true rates, but rate changes should still reflect temporal trends.
The best data sets for SPADE will have information on visits and admissions, and on other events, such as intrahospital care escalations (eg, ward to ICU transfers) and deaths. Recently, pairing of non-life-threatening ED discharge diagnoses to subsequent death among Medicare beneficiaries was used to identify misdiagnoses (figure 4).42 However, even without death (or other outcome) data, tracking to monitor diagnostic quality and safety trends and intervene to improve them remains possible. This is because root causes (eg, cognitive biases, knowledge deficits) and process failures (eg, exam findings not elicited, tests not ordered) leading to misdiagnosis of specific dangerous diseases probably do not differ based on the severity of subsequent harms (eg, hospital readmission vs out-of-hospital death). Even for conditions with very high mortality (eg, aortic dissection), many patients would still be captured by a delayed admission-only approach.45 Thus, a diagnostic intervention to improve diagnosis of aortic dissection that reduced misdiagnosis-related readmissions would presumably also reduce misdiagnosis-related deaths.
Having systematically coded EHR data on presenting symptoms (as opposed to inferring these from index visit discharge diagnoses) can enrich a SPADE analysis. However, it is not essential, since it is the benign or non-specific nature of the index visit discharge diagnosis (rather than the presenting symptom, per se) that reflects the diagnostic error. Furthermore, many of the index visit diagnoses are coded as non-specific symptoms (eg, dizziness, not otherwise specified25).
USING SPADE TO ASSESS PREVENTABLE HARMS FROM DIAGNOSTIC PROCESS FAILURES
SPADE measures the frequency of diagnostic errors causing misdiagnosis-related harms, rather than all diagnostic errors. This concept is most intuitive using the look-forward approach. Isolated vertigo of vascular aetiology is the most common early manifestation of brainstem or cerebellar ischaemia and is often missed initially as a stroke sign.29 Since it is unlikely that a patient sent home with an index diagnosis of ‘benign’ vertigo also had other obvious neurological signs (eg, hemiparesis or aphasia), their subsequent hospitalisation for stroke suggests clinical worsening or recurrent ischaemia (eg, major stroke after minor stroke or TIA).31 Thus, graphically, the ‘hump’ (hatched area) shown in figure 3B more accurately reflects misdiagnosis-related harms rather than diagnostic error, per se. Fewer than 20% of patients with TIA or minor stroke go on to suffer a major stroke within 90 days,46,47 so there are likely to be at least fivefold more diagnostic errors (misidentifications of TIA or minor stroke at the index visit) than misdiagnosis-related harms (subsequent, delayed major stroke admissions).
When diagnostic process data (eg, use of imaging, lab tests or consults) are also available, it is possible to identify process failures and test their association with misdiagnosis-related harms. For example, guidelines indicate that benign paroxysmal positional vertigo (BPPV), an inner ear disease, should be diagnosed and treated at the bedside without neuroimaging.48,49 Frequent use of neuroimaging in patients discharged with BPPV suggests knowledge or skill gaps in bedside diagnosis of vertigo.26,50 Such process failures may correlate to misdiagnosis-related harms (eg, use of neuroimaging in ‘benign’ dizziness/vertigo is linked to increased odds of stroke readmission after discharge51). For cancers, process failures can be identified by measuring diagnostic intervals (eg, time from index visit to advanced testing or specialty consultation to treatment)43,52; diagnostic delays can be correlated to outcomes and targeted for disease-specific process improvement.53
The SPADE approach can also facilitate identification of symptom-independent system factors that contribute to misdiagnosis. For example, in the study described above looking at short-term mortality after ED discharge, low hospital admission fraction at the index ED visit was associated with death postdis-charge.42 Other studies have found triage to low acuity care is linked to misdiagnosis.19 Healthcare settings can be compared for risk of misdiagnosis and harms—for example, the risk for missed stroke is greater in ED than primary care, but the magnitude of harms is similar because of greater patient volumes in primary care.22 Important demographic and racial disparities in care can also be measured using SPADE.24,25
USING SPADE TO MEASURE DIAGNOSTIC PERFORMANCE AND IMPACT OF INTERVENTIONS
The operational quality and safety goal is ongoing measurement of diagnostic performance in actual clinical practice.9 A major advantage of SPADE is that the core, essential administrative data are already being collected and could be easily used to track diagnostic performance without significant financial burdens. Because these data are also available from past years, internal performance trend lines could be readily constructed. For relatively common diagnostic problems such as chest pain-MI or vertigo-stroke, health systems could probably monitor their performance semiannually or quarterly using a rolling window of 6–12 months of data. Such monitoring would facilitate assessment of interventions to improve diagnostic performance.
In 2017, a National Quality Forum expert panel highlighted SPADE methods as a key measure concept to assess ‘harms from diagnostic error based on unexpected change in health status’ that holds promise for operational use because of the ready availability of administrative data.54 Relevant data for applying SPADE are already gathered in standard, structured formats (eg, International Classification of Diseases diagnostic codes); thus, cross-institutional benchmarking is a realistic possibility if data are curated through an ‘honest broker’.55 Geographic or institutional variation in diagnostic accuracy could also be detected.25,56 Eventually, SPADE-derived metrics could be incorporated into operational diagnostic performance dashboards.22
DIFFERENCES BETWEEN SPADE AND ELECTRONIC TRIGGER TOOLS
Electronic trigger tools seek to identify missed diagnostic opportunities or failed diagnostic processes.57–59 Trigger tools use specific predetermined EHR events (eg, unplanned revisits to primary care) to ‘trigger’ medical record review by trained personnel.60 These ‘trigger’ events can be similar to outcome events used in SPADE, but trigger tools rely on human chart review for adjudication of diagnostic errors, while SPADE combines biological plausibility with statistical analysis of large data sets to verify errors. Also, trigger tools are typically used to find individual patient errors for process analysis and remediation, while SPADE would be used to understand the overall landscape of misdiagnosis-related harms to prioritise problems for solution-making and to operationally track performance over time, including to assess impact of interventions.
LIMITATIONS OF SPADE
SPADE will not solve all problems in measuring diagnostic errors.17,61–64 The method probably substantially understates the frequency of NAM-defined diagnostic errors, since it focuses on misdiagnosis-related harms. It is also not readily applied to all disease states, including chronic conditions where adverse outcomes are evenly distributed over time. The spatial and temporal resolutions are too low to provide individual provider feedback. Correlating SPADE outcome measurements directly to bedside process failures (eg, flawed history or examination) will still require free-text analysis of records or other granular data. When using coded diagnoses for index and outcome visits, SPADE is potentially susceptible to various types of coding error and bias, including intentional gaming such as mis-specification, unbundling and upcoding.65 Because SPADE uses large data sets to identify diagnostic error patterns, it risks apophenia,66 so appropriate statistical validation checks and controls are critical when using SPADE (table 1).
Finally, SPADE has not been directly validated against an independent ‘gold standard’. The method is strongly supported by the fact that the dizziness-stroke dyad has an extensive body of remarkably coherent and consistent scientific literature26,28 that includes chart reviews,15,16,67 surveys,68,69 cross-sectional health services research studies,50,51,56,70 prospective cohort studies71,72 and SPADE-type studies using look-back25 and look-forward21–23 methods. Problems inherent in human chart reviews, particularly hindsight and observer biases,12–14 and flawed underlying documentation15 suggest that this is probably not an ideal reference standard for SPADE. A better validation strategy might be to vet coding and classification accuracy against review of videotaped encounters or gold-standard randomised trial data, as from AVERT (Acute Video-oculography for Vertigo in Emergency Rooms for Rapid Triage; ClinicalTrials.gov NCT02483429). The most compelling validation of the SPADE method would probably be to ‘flatten the hump’ (figure 3B) through diagnostic quality and safety interventions—this would demonstrate predictive validity of SPADE-based metrics.
CONCLUSIONS
We have elaborated a conceptual framework, SPADE, that could be used to measure and monitor a key subset of misdiagnosis-related harms using pre-existing, administrative ‘big data’. This directly addresses a major patient safety and public health need1,9 which we believe could be transformational for improving diagnosis in healthcare by surfacing otherwise hidden diagnostic errors. The SPADE approach leverages symptom-disease pairs and uses statistically controlled inferential analyses of large data sets to construct operational outcome metrics that could be incorporated into diagnostic performance dash-boards.22 When tested, these metrics have demonstrated multiple aspects of validity and reliability. Broad application of the SPADE approach could facilitate local operational improvements, and large-scale, epidemiological research to assess the breadth and distribution of misdiagnosis-related harms, and international/national benchmarking efforts that establish standards for diagnostic quality and safety. Future research should seek to validate SPADE across a wide range of clinical problems.
Acknowledgments
Funding National Institute on Deafness and Communication Disorders (grant #U01 DC013778) and the Armstrong Institute Center for Diagnostic Excellence.
Footnotes
Competing interests None declared.
Provenance and peer review Not commissioned; externally peer reviewed.
No commercial use is permitted unless otherwise expressly granted.
REFERENCES
- 1.National Academies of Sciences, Engineering, and Medicine. Improving diagnosis in health care. Washington, DC: The National Academies Press, 2015. [Google Scholar]
- 2.Singh H, Meyer AN, Thomas EJ. The frequency of diagnostic errors in outpatient care: estimations from three large observational studies involving US adult populations. BMJ Qual Saf 2014;23:727–31. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 3.Singh H, Giardina TD, Meyer AN, et al. Types and origins of diagnostic errors in primary care settings. JAMA Intern Med 2013;173:418–25. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 4.Newman-Toker DE. Diagnostic value: the economics of high-quality diagnosis and a value-based perspective on diagnostic innovation [lecture]. Modern Healthcare 3rd Annual Patient Safety & Quality Virtual Conference;2015 June 17;online e-conference. [Google Scholar]
- 5.Singh H, Schiff GD, Graber ML, et al. The global burden of diagnostic errors in primary care. BMJ Qual Saf 2017;26:484–94. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 6.Johnston SC, Mendis S, Mathers CD. Global variation in stroke burden and mortality: estimates from monitoring, surveillance, and modelling. Lancet Neurol 2009;8:345–54. [DOI] [PubMed] [Google Scholar]
- 7.Davies PD, Pai M. The diagnosis and misdiagnosis of tuberculosis. Int J Tuberc Lung Dis 2008;12:1226–34. [PubMed] [Google Scholar]
- 8.Storla DG, Yimer S, Bjune GA, et al. A systematic review of delay in the diagnosis and treatment of tuberculosis. BMC Public Health 2008;8:15. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 9.Henriksen K, Dymek C, Harrison MI, et al. Challenges and opportunities from the Agency for Healthcare Research and Quality (AHRQ) research summit on improving diagnosis: a proceedings review. Diagnosis 2017;4:57–66. [DOI] [PubMed] [Google Scholar]
- 10.Luck J, Peabody JW, Dresselhaus TR, et al. How well does chart abstraction measure quality? A prospective comparison of standardized patients with the medical record. Am J Med 2000;108:642–9. [DOI] [PubMed] [Google Scholar]
- 11.Thomas EJ, Petersen LA. Measuring errors and adverse events in health care. J Gen Intern Med 2003;18:61–7. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 12.Hayward RA, Hofer TP. Estimating hospital deaths due to medical errors: preventability is in the eye of the reviewer. JAMA 2001;286:415–20. [DOI] [PubMed] [Google Scholar]
- 13.Weingart SN, Davis RB, Palmer RH, et al. Discrepancies between explicit and implicit review: physician and nurse assessments of complications and quality. Health Serv Res 2002;37:483–98. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14.Wears RL, Nemeth CP Replacing hindsight with insight: toward better understanding of diagnostic failures. Ann Emerg Med 2007;49:206–9. [DOI] [PubMed] [Google Scholar]
- 15.Newman-Toker DE. Charted records of dizzy patients suggest emergency physicians emphasize symptom quality in diagnostic assessment. Ann Emerg Med 2007;50:204–5. [DOI] [PubMed] [Google Scholar]
- 16.Kerber KA, Morgenstern LB, Meurer WJ, et al. Nystagmus assessments documented by emergency physicians in acute dizziness presentations: a target for decision support? Acad Emerg Med 2011;18:619–26. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 17.Newman-Toker DE, Pronovost PJ. Diagnostic errors--the next frontier for patient safety. JAMA 2009;301:1060–2. [DOI] [PubMed] [Google Scholar]
- 18.Newman-Toker DE. A unified conceptual model for diagnostic errors: underdiagnosis, overdiagnosis, and misdiagnosis. Diagnosis 2014;1:43–8. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 19.Schull MJ, Vermeulen MJ, Stukel TA. The risk of missed diagnosis of acute myocardial infarction associated with emergency department volume. Ann Emerg Med 2006;48:647–55. [DOI] [PubMed] [Google Scholar]
- 20.Vermeulen MJ, Schull MJ. Missed diagnosis of subarachnoid hemorrhage in the emergency department. Stroke 2007;38:1216–21. [DOI] [PubMed] [Google Scholar]
- 21.Kim AS, Fullerton HJ, Johnston SC. Risk of vascular events in emergency department patients discharged home with diagnosis of dizziness or vertigo. Ann Emerg Med 2011;57:34–41. [DOI] [PubMed] [Google Scholar]
- 22.Newman-Toker DE, Moy E, Valente E, et al. Missed diagnosis of stroke in the emergency department: a cross-sectional analysis of a large population-based sample. Diagnosis 2014;1:155–66. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 23.Moy E, Barrett M, Coffey R, et al. Missed diagnoses of acute myocardial infarction in the emergency department: variation by patient and facility characteristics. Diagnosis 2015;2:29–40. [DOI] [PubMed] [Google Scholar]
- 24.Atzema CL, Grewal K, Lu H, et al. Outcomes among patients discharged from the emergency department with a diagnosis of peripheral vertigo. Ann Neurol 2016;79:32–41. [DOI] [PubMed] [Google Scholar]
- 25.Nassery NMK, Liu F, Sangha NX, et al. Measuring missed strokes using administrative and claims data: towards a diagnostic performance dashboard to monitor diagnostic errors [oral abstract]. 9th Annual Diagnostic Error in Medicine Conference Hollywood, CA;2016 November 6–8. [Google Scholar]
- 26.Savitz SI, Caplan LR, Edlow JA. Pitfalls in the diagnosis of cerebellar infarction. Acad Emerg Med 2007;14:63–8. [DOI] [PubMed] [Google Scholar]
- 27.Kerber KA, Newman-Toker DE. Misdiagnosing Dizzy Patients: Common Pitfalls in Clinical Practice. Neurol Clin 2015;33:565–75, viii. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 28.Newman-Toker DE. Missed stroke in acute vertigo and dizziness: It is time for action, not debate. Ann Neurol 2016;79:27–31. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 29.Paul NL, Simoni M, Rothwell PM, et al. Transient isolated brainstem symptoms preceding posterior circulation stroke: a population-based study. Lancet Neurol 2013;12:65–71. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 30.Valls J, Peiro-Chamarro M, Cambray S, et al. A Current Estimation of the Early Risk of Stroke after Transient Ischemic Attack: A Systematic Review and Meta-Analysis of Recent Intervention Studies. Cerebrovasc Dis 2017;43:90–8. [DOI] [PubMed] [Google Scholar]
- 31.Rothwell PM, Buchan A, Johnston SC. Recent advances in management of transient ischaemic attacks and minor ischaemic strokes. Lancet Neurol 2006;5:323–31. [DOI] [PubMed] [Google Scholar]
- 32.Karras DJ, methodology: S II. Reliability and variability assessment in study design, Part A. Acad Emerg Med 1997;4:64–71. [DOI] [PubMed] [Google Scholar]
- 33.Karras DJ, methodology: S II. Reliability and validity assessment in study design, Part B. Acad Emerg Med 1997;4:144–7. [DOI] [PubMed] [Google Scholar]
- 34.Rothman KJ, Greenland S. Causation and causal inference in epidemiology. Am J Public Health 2005;95(Suppl 1):S144–150. [DOI] [PubMed] [Google Scholar]
- 35.Fahimi J, Navi BB, Kamel H. Potential misdiagnoses of Bell’s palsy in the emergency department. Ann Emerg Med 2014;63:428–34. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 36.Vaillancourt S, Guttmann A, Li Q, et al. Repeated emergency department visits among children admitted with meningitis or septicemia: a population-based study. Ann Emerg Med 2015;65:625–32. [DOI] [PubMed] [Google Scholar]
- 37.Kachalia A, Gandhi TK, Puopolo AL, et al. Missed and delayed diagnoses in the emergency department: a study of closed malpractice claims from 4 liability insurers. Ann Emerg Med 2007;49:196–205. [DOI] [PubMed] [Google Scholar]
- 38.Winters B, Custer J, Galvagno SM, et al. Diagnostic errors in the intensive care unit: a systematic review of autopsy studies. BMJ Qual Saf 2012;21:894–902. [DOI] [PubMed] [Google Scholar]
- 39.Custer JW, Winters BD, Goode V, et al. Diagnostic errors in the pediatric and neonatal ICU: a systematic review. Pediatr Crit Care Med 2015;16:29–36. [DOI] [PubMed] [Google Scholar]
- 40.Troxel DB. the Doctor’s company Diagnostic Error in Medical Practice by Specialty. 2014. http://www.thedoctors.com/KnowledgeCenter/Publications/TheDoctorsAdvocate/Diagnostic-Error-in-Medical-Practice-by-Specialty (accessed 1 Sep 2017).
- 41.Prandoni P, Lensing AW, Prins MH, et al. Prevalence of Pulmonary Embolism among Patients Hospitalized for Syncope. N Engl J Med 2016;375:1524–31. [DOI] [PubMed] [Google Scholar]
- 42.Obermeyer Z, Cohn B, Wilson M, et al. Early death after discharge from emergency departments: analysis of national US insurance claims data. BMJ 2017;356:j239. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 43.Lyratzopoulos G, Vedsted P, Singh H. Understanding missed opportunities for more timely diagnosis of cancer in symptomatic patients after presentation. Br J Cancer 2015;112(Suppl 1):S84–91. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 44.Finnell JT, Overhage JM, Dexter PR, et al. Community clinical data exchange for emergency medicine patients. AMIA Annu Symp Proc 2003:235–8. [PMC free article] [PubMed] [Google Scholar]
- 45.Hansen MS, Nogareda GJ, Hutchison SJ. Frequency of and inappropriate treatment of misdiagnosis of acute aortic dissection. Am J Cardiol 2007;99:852–6. [DOI] [PubMed] [Google Scholar]
- 46.Giles MF, Rothwell PM. Risk of stroke early after transient ischaemic attack: a systematic review and meta-analysis. Lancet Neurol 2007;6:1063–72. [DOI] [PubMed] [Google Scholar]
- 47.Flossmann E, Rothwell PM. Prognosis of vertebrobasilar transient ischaemic attack and minor stroke. Brain 2003;126:1940–54. [DOI] [PubMed] [Google Scholar]
- 48.Bhattacharyya N, Baugh RF, Orvidas L, et al. Clinical practice guideline: benign paroxysmal positional vertigo. Otolaryngol Head Neck Surg 2008;139:47–81. [DOI] [PubMed] [Google Scholar]
- 49.Fife TD, Iverson DJ, Lempert T, et al. Practice parameter: therapies for benign paroxysmal positional vertigo (an evidence-based review): report of the Quality Standards Subcommittee of the American Academy of Neurology. Neurology 2008;70:2067–74. [DOI] [PubMed] [Google Scholar]
- 50.Newman-Toker DE, Camargo CA, Hsieh YH, et al. Disconnect between charted vestibular diagnoses and emergency department management decisions: a cross-sectional analysis from a nationally representative sample. Acad Emerg Med 2009;16:970–7. [DOI] [PubMed] [Google Scholar]
- 51.Grewal K, Austin PC, Kapral MK, et al. Missed strokes using computed tomography imaging in patients with vertigo: population-based cohort study. Stroke 2015;46:108–13. [DOI] [PubMed] [Google Scholar]
- 52.Lyratzopoulos G, Saunders CL, Abel GA, et al. The relative length of the patient and the primary care interval in patients with 28 common and rarer cancers. Br J Cancer 2015;112(Suppl 1):S35–40. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 53.Senter J, Hooker C, Lang M, et al. Thoracic Multidisciplinary Clinic Improves Survival in Patients With Lung Cancer. Int J Radiat Oncol Biol Phys 2016;96:S134. [Google Scholar]
- 54.Improving Diagnostic Quality and Safety. 2016. http://www.qualityforum.org/Improving_Diagnostic_Quality_and_Safety.aspx (accessed 7 Jun 2017). [Google Scholar]
- 55.Boyd AD, Saxman PR, Hunscher DA, et al. The University of Michigan Honest Broker: a Web-based service for clinical and translational research and practice. J Am Med Inform Assoc 2009;16:784–91. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 56.Kim AS, Sidney S, Klingman JG, et al. Practice variation in neuroimaging to evaluate dizziness in the ED. Am J Emerg Med 2012;30:665–72. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 57.Singh H, Giardina TD, Forjuoh SN, et al. Electronic health record-based surveillance of diagnostic errors in primary care. BMJ Qual Saf 2012;21:93–100. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 58.Murphy DR, Laxmisan A, Reis BA, et al. Electronic health record-based triggers to detect potential delays in cancer diagnosis. BMJ Qual Saf 2014;23:8–16. [DOI] [PubMed] [Google Scholar]
- 59.Murphy DR, Wu L, Thomas EJ, et al. Electronic Trigger-Based Intervention to Reduce Delays in Diagnostic Evaluation for Cancer: A Cluster Randomized Controlled Trial . J Clin Oncol 2015;33:3560–7. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 60.Al-Mutairi A, Meyer AN, Thomas EJ, et al. Accuracy of the Safer Dx Instrument to Identify Diagnostic Errors in Primary Care. J Gen Intern Med 2016;31:602–8. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 61.Graber M Diagnostic errors in medicine: a case of neglect. Jt Comm J Qual Patient Saf 2005;31:106–13. [DOI] [PubMed] [Google Scholar]
- 62.Shojania K 3rd Annual Diagnostic Error in Medicine Conference, Toronto, Canada;2010 September 11–12;Diagnostic mistakes - past, present, and future [lecture]. [Google Scholar]
- 63.Graber ML. The incidence of diagnostic error in medicine. BMJ Qual Saf 2013;22 Suppl 2:ii21–ii27. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 64.Zwaan L, Singh H. The challenges in defining and measuring diagnostic error. Diagnosis 2015;2:97–103. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 65.O’Malley KJ, Cook KF, Price MD, et al. Measuring diagnoses: ICD code accuracy. Health Serv Res 2005;40:1620–39. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 66.Wears RL, Williams DJ. Big Questions for “Big Data”. Ann Emerg Med 2016;67:237–9. [DOI] [PubMed] [Google Scholar]
- 67.Kerber KA, Burke JF, Skolarus LE, et al. Use of BPPV processes in emergency department dizziness presentations: a population-based study. Otolaryngol Head Neck Surg 2013;148:425–30. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 68.Stanton VA, Hsieh YH, Camargo CA, et al. Overreliance on symptom quality in diagnosing dizziness: results of a multicenter survey of emergency physicians. Mayo Clin Proc 2007;82:1319–28. [DOI] [PubMed] [Google Scholar]
- 69.Newman-Toker DE, Stanton VA, Hsieh YH, et al. Frontline providers harbor misconceptions about the bedside evaluation of dizzy patients. Acta Otolaryngol 2008;128:601–4. [DOI] [PubMed] [Google Scholar]
- 70.Kerber KA, Schweigler L, West BT, et al. Value of computed tomography scans in ED dizziness visits: analysis from a nationally representative sample. Am J Emerg Med 2010;28:1030–6. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 71.Kerber KA, Brown DL, Lisabeth LD, et al. Stroke among patients with dizziness, vertigo, and imbalance in the emergency department: a population-based study. Stroke 2006;37:2484–7. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 72.Kerber KA, Zahuranec DB, Brown DL, et al. Stroke risk after nonstroke emergency department dizziness presentations: a population-based cohort study. Ann Neurol 2014;75:899–907. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 73.De Bin R, Janitza S, Sauerbrei W, et al. Subsampling versus bootstrapping in resampling-based model selection for multivariable regression. Biometrics 2016;72:272–80. [DOI] [PubMed] [Google Scholar]