Abstract
Diagnostic errors can pose a serious threat to patient safety, leading to serious harm and even death. Efforts are being made to develop interventions that allow physicians to reassess for errors and improve diagnostic accuracy. Our study presents an exploration of misdiagnosis patterns mined from PubMed abstracts. Article titles containing certain phrases indicating misdiagnosis were selected and frequencies of these misdiagnoses calculated. We present the resulting patterns in the form of a directed graph with frequency-weighted misdiagnosis edges connecting diagnosis vertices. We find that the most commonly misdiagnosed diseases were often misdiagnosed as many different diseases, with each misdiagnosis having a relatively low frequency, rather than as a single disease with greater probability. Additionally, while a misdiagnosis relationship may generally exist, the relationship was often found to be one-sided.
Introduction
Diagnostic errors can be frequent, costly, and sometimes fatal in medicine. It is one of the most pressing issues on patients’ minds, with 22% of all patients seeking emergency treatment expressing concern over misdiagnoses1. In a randomized survey, more than 1 in 10 respondents reported having experienced an issue with their diagnosis in the past. It is estimated that about 12 million Americans may face a diagnostic error each year, and half of these errors have the potential to cause serious harm2. Worse, according to one study, 83.3% of diagnostic errors were preventable3.
Diagnostic errors can cause serious costs for both the healthcare provider and the patient. Medical errors, among which diagnostic errors were found to make up around 28.7% based on malpractice claims, cost the United States $19.5 billion in 2008 alone4,5. The extra cost for additional treatment, more deaths, and lost productivity is driving up the cost of healthcare, hurting both healthcare providers and patients seeking treatments5. Malpractice claims are also common and costly, with sometimes exorbitant litigation fees6. Diagnostic errors contribute a significant portion to these cases, ranging from 30 to 59%7,8.
For patients, the consequences of diagnostic errors are even more direct. In 59% of the malpractice claims resulting from diagnostic errors, serious harm was caused by these errors and 30% resulted in death7. Even in less serious cases, diagnostic errors may lead to patients needing more doctor visits, longer hospital stays, or receiving inappropriate treatment9,10.
Physicians can be affected by cognitive biases and personality traits when making medical decisions. Factors such as overconfidence or risk aversion can affect the accuracy of their diagnoses. These cognitive biases were associated with inaccuracies in diagnosis in 36.5% to 77% of case scenarios11. Furthermore, physicians are often unaware of mistakes they make. Interventions that force doctors to reassess and be more aware of possible errors, may thus be very useful in reducing the frequency of diagnostic errors8.
One study described an intervention that has already shown to improve diagnostic accuracy through a computerised diagnostic support system, where physicians code in symptoms, and a list of possible diagnoses along with their likelihood is returned12. In another study, 74% of physicians felt that these systems were useful, helping them consider more diagnoses and ask more specific questions13. Furthermore, these systems did not increase the duration of the consultation nor the number of tests ordered. Yet another study’s system showed an improvement of 8-9% in diagnostic accuracy, which, given the many consultations that occur every day, would benefit a significant number of patients12. We hypothesize that these approaches may be improved by integrating patterns of misdiagnosis, allowing them to return not only the likelihood of a diagnosis but also the likelihood of its misdiagnosis as another disease.
In this study we explore accounts of misdiagnoses in the PubMed database as a resource for potentially flagging misdiagnosis.
Methodology
We used the PubMed database, specifically all articles contained in the PubMed 2018 annual baseline (N=27,837,540 citations). Each article is an XML file with title, author, and abstract tags among others. Using Python, we parsed through these articles, selecting only those whose titles contained the phrases, “misdiagnosed as” or “masquerading as”. These phrases were chosen after testing phrases using the PubMed database search tool to determine which ones produced the most results. Titles were examined for relevancy and those that did not pertain to medical terms were filtered out when matching.
From the selected titles, we used QuickUMLS14, a medical named entity resolution library, to extract disease names and their Universal Medical Language System concept unique identifiers (UMLS CUIs), each of which is denoted by a “C” followed by 7 digits, selecting specifically for semantic types T047 (Disease or Syndrome) and T191 (Neoplastic Process). QuickUMLS returned a list of all matches, with each entry in the list representing a list of all possible matches to a medical term found in the title. Each match was represented by a dictionary with the start and end of the medical term being matched, the UMLS CUI, the term associated with the UMLS CUI, the degree of similarity from 0 to 1, and the UMLS semantic type. Since a given medical term may match with multiple CUIs, we prioritized the longest match, followed by the greatest similarity, followed by the CUI with the lowest 7-digit number. We considered only titles that contained exactly two matches, one preceding our selected phrase and one following after it. Figure 1 illustrates this process.
Figure 1.
Examples of titles matching our condition (left) and not matching our condition (right). The selection criterion was that there is only a single medical concept recognized preceding the chosen phrase (red) and only a single one following after it. For the title on the right, there were two recognized medical concepts preceding the chosen phrase, so the title was not selected.
The observation frequencies of each disease pair and of each disease being misdiagnosed as another were calculated. Using the UMLS API15, each CUI extracted was checked for parent/child or synonymy relationships with other CUIs. If a CUI had a parent-child (PAR or RB and CHD or RN, respectively) or synonymous (SYN or RL) relationship with another CUI extracted from one of our titles, it was counted as that parent or synonymous CUI16 (Figure 2). Each CUI pair frequency was then normalized by the frequency of the first CUI, which represents the correct diagnosis, and a graph was generated with the normalized frequencies as the weights of the directed edges. The graph was created and graph statistics analyzed using NetworkX17.
Figure 2.
Diagram of parent/child and synonymy CUI relationships and the resulting chosen CUI. In the case of parent/child relationships, the parent CUI was always chosen. For synonymy relationships, the first synonymous CUI to appear in the selected titles was chosen.
Results
Of 5,105 titles that contained our given phrase, 2,502 misdiagnosis pairs were extracted. Each node represents a unique CUI and the arrows indicate a misdiagnosis relationship, where the source node is the correct diagnosis, and the destination node is the incorrect diagnosis. The darker node indicates that the source node was more often misdiagnosed as the destination node than any of its other destination nodes.
The larger nodes in the center indicate that several CUIs are misdiagnosed as other diseases relatively frequently whereas the smallest nodes along the edges indicate that there are diseases that were rarely or never reported misdiagnosed based on the articles extracted. The darker edges suggest that when the source node is misdiagnosed, it is more often misdiagnosed as the destination node; however, in many cases, especially in the case of the nodes on the edge of the graph, the greater weight is due to the frequency of the source node being very low to begin with (Figure 3).
Figure 3.
Graphs of all CUIs and misdiagnosis relationships extracted. Graph is drawn using a force-directed algorithm. The node closest to the arrowhead is the disease that another is being misdiagnosed as. A larger node means a greater out-degree centrality while a darker arrow means the relationship has a greater normalized frequency. (A) All nodes shown. (B) Nodes in the center tend to be larger. (C) Nodes along the edge tend to be smaller.
While many diseases only had a few incidences of misdiagnosis cited in the literature, several were misdiagnosed relatively frequently, the top five of which can be found in Table 1. The single most commonly misdiagnosed disease was tuberculosis (C0041296), which had both the greatest source frequency, or number of mentions of being misdiagnosed, as well as the greatest out-degree (i.e. the number of different diseases it was wrongly diagnosed as). This is not surprising from a clinical viewpoint, tuberculosis has long been known as a mimic of other diseases. It was most often misdiagnosed as a form of carcinoma followed by malignancies and other tumors. The frequency with which tuberculosis was misdiagnosed as each disease were not high, though, with tuberculosis being misdiagnosed as carcinoma in about 10% of cases of tuberculosis misdiagnosis.
Table 1.
Top 5 most commonly misdiagnosed diseases according to our results, along with their out-degree, source frequency, or how often a given disease was misdiagnosed as another, and the disease they are most mistaken as.
Name (CUI) | Source Frequency | Out Degree | Most Frequently Mistaken as (CUI) & Frequency |
---|---|---|---|
Tuberculosis (C0041296) | 68 | 46 | Carcinoma (C0041296) 0.1029 |
Cyst (C0010709) | 44 | 34 | Pericarditis (C0031046) 0.0455 |
Tumor (C0027651) | 40 | 37 | Liver Secondaries (C0494165) 0.0500 |
Nerve Sheath Tumor (C0206727) | 32 | 24 | Carotid Body Tumor (C0007279) 0.1250 |
Neuroendocrine Tumor (C0206754) | 31 | 28 | Diabetic Foot Ulcer (C1456868) 0.1290 |
Similarly, for the other most commonly misdiagnosed diseases, the frequency of each misdiagnosis pair was fairly low, with the highest being about 13% of all cases of that disease being misdiagnosed (Table 1). Overall, the normalized frequencies were relatively low, except in the cases where the occurrence of the correct diagnosis was low to begin with.
When inspecting reversed disease pair relationships, it was found that while one disease may have been most frequently misdiagnosed as another, the same did not often hold in the reverse direction. For example, while tuberculosis was most frequently misdiagnosed as a carcinoma, there were no mentions at all of carcinoma misdiagnosed as tuberculosis in our results. Of the top 20 most misdiagnosed diseases, for all but two diseases, the CUI that the disease was most misdiagnosed as was never misdiagnosed as the disease. In those two other cases, the frequency only went up to about 8.7%.
Just as with the misdiagnosed diseases, most of the diagnosed diseases that were later found to be incorrect only had a few incidences, although a few were relatively often diagnosed when it was that disease. The disease with highest in-degree, or the one that other diseases are most commonly misdiagnosed as, are tumors (C0027651). As with the misdiagnosed diseases, though, when a disease was incorrectly diagnosed, there was no single disease that was very frequently the correct diagnosis. In only about 8.21% of cases where a tumor was a destination node, or incorrect diagnosis, the correct diagnosis was tuberculosis and this was the highest frequency for tumors. For the other most often incorrectly diagnosed diseases, the frequency only reached about 12% of all cases where that disease was the incorrect diagnosis (Table 2).
Table 2.
Top 5 diseases most often wrongly diagnosed according to our results, along with their in-degree, destination frequency, or how often other diseases were misdiagnosed as a given disease, and the most frequent correct diagnosis.
Name (CUI) | Destination Frequency | In Degree | Most Frequent Correct (CUI) & Frequency |
---|---|---|---|
Tumor (C0027651) | 73 | 62 | Tuberculosis (C0041296) 0.0821 |
Malignancy (C0006826) | 59 | 45 | Actinomycosis (C0001261) 0.1186 |
Cyst (C0010709) | 42 | 31 | Connective Tissue Tumor (C0027656) 0.0952 |
Tuberculosis (C0041296) | 35 | 33 | Carcinoma of Lung (C0684249) 0.0857 |
Skin Conditions (C0037274) | 34 | 30 | Skin Ulcers (C0037299) 0.1176 |
As with the normalized frequencies, the frequency of each disease given the incorrect diagnosis was relatively low for most misdiagnoses except in cases where the wrong disease had a low frequency. To facilitate further research, we share all extracted misdiagnosis pairs alongside the computed normalized frequencies and graphs with the research community (https://github.com/bcbi-edu/p_eickhoff-misdiagnosis).
Discussion
Our study shows that diagnoses are often mistaken for a variety of diseases rather than a single recurring one. A disease pair often occurred only once, and for disease pairs with a frequency greater than one, the normalized count was often low. Thus, the frequency of the disease being misdiagnosed as a single specific disease was relatively low. Likewise, incorrect diagnoses link to multiple correct diagnoses.
This one-sided directionality of misdiagnosis relationships suggests that while symptoms for the two diseases may be similar, leading to the first disease often being misdiagnosed as the second, the second disease may not be frequently misdiagnosed as the first. This may in some cases be due to the rarity of the diseases.
In our current analysis, much data (about 51%) was lost due to the often complicated structures of the article titles. More data may also be gained by looking at more than just article titles when selecting those to evaluate initially as articles may discuss misdiagnoses without explicitly stating it in the title. In this regard, abstracts, full texts and annotated resources such as SemRep18 or SemMedDB119 will be of great value.
Our work provides a first step towards obtaining misdiagnosis patterns and frequencies. Future work may include factoring in overall prevalence rates of diseases instead of just cases where the diagnosis was erroneous. Our observation frequencies currently only reflect the probability of one disease given an incorrect diagnosis. Healthcare providers may benefit more from also knowing the frequency of a diagnosis being incorrect. Additionally, studying orthogonal sources, possibly open medical records or malpractice claims, as well as additional phrases indicating misdiagnoses may offer more insights.
In the long run we are excited to use these early insights towards diagnostic decision support. For example, if meningitis is often misdiagnosed as the flu - then that should direct us to ensure that patients with an initial diagnosis of flu receive the correct screening questions to help discriminate them. One also might use this data to devise more effective and cost efficient strategies for investigation of patients with certain provisional diagnoses. Finally, this approach might be an effective way to inform a specialized systematic review where the search tools identify the cohort of potential studies to include which then undergo detailed review to extract important relationships between diagnoses.
Conclusion
Diagnostic errors can pose a serious threat to patients. This paper mines patterns from thousands of published misdiagnosis reports in the biomedical literature and structures the derived information in the form of a directed graph with frequency-weighted edges. It offers an additional way to understand misdiagnosis from a clinical or diagnostic decision support viewpoint The resulting materials are made available to the research community to inform clinical practice and research.
Figures & Table
References
- 1.Burroughs TE, Waterman AD, Gallagher TH, et al. Patient concerns about medical errors in emergency departments. Acad Emerg Med. 2008;12(1):57–64. doi: 10.1197/j.aem.2004.08.052. [DOI] [PubMed] [Google Scholar]
- 2.Singh H, Meyer AN, Thomas EJ. The frequency of diagnostic errors in outpatient care: estimations from three large observational studies involving US adult populations. BMJ Qual Saf. 2014;23:727–731. doi: 10.1136/bmjqs-2013-002627. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 3.Zwaan L, Thijs A, Wagner C, van der Wal G, Timmermans DR. Relating faults in diagnostic reasoning with diagnostic errors and patient harm. Acad Med. 2012;87(2):149–156. doi: 10.1097/ACM.0b013e31823f71e6. [DOI] [PubMed] [Google Scholar]
- 4.Newman-Toker DE, Schaffer AC, Yu-Moe CW, et al. Serious misdiagnosis-related harms in malpractice claims: the “Big Three” - vascular events, infections, and cancers. Diagnosis. 2019;6(3):227–240. doi: 10.1515/dx-2019-0019. [DOI] [PubMed] [Google Scholar]
- 5.Andel C, Davidow SL, Hollander M, Moreno DA. The economics of health care quality and medical errors. J Health Care Finance. 2012;39(1):39–50. [PubMed] [Google Scholar]
- 6.Studdert DM, Mello MM, Gawande AA, et al. Claims, errors, and compensation payments in medical malpractice litigation. N Engl J Med. 2006;354:2024–2033. doi: 10.1056/NEJMsa054479. [DOI] [PubMed] [Google Scholar]
- 7.Gandhi TK, Kachalia A, Thomas EJ, et al. Missed and delayed diagnoses in the ambulatory setting: a study of closed malpractice claims. Ann Intern Med. 2006;145:488–496. doi: 10.7326/0003-4819-145-7-200610030-00006. [DOI] [PubMed] [Google Scholar]
- 8.Zwaan L, Bruijne M, Wagner C, et al. Patient record review of the incidence, consequences, and causes of diagnostic adverse events. JAMA Intern Med. 2010;170:1015–1021. doi: 10.1001/archinternmed.2010.146. [DOI] [PubMed] [Google Scholar]
- 9.Carberry AR, Hanson K, Flannery A, et al. Diagnostic error in pediatric cancer. Clin Pediatr (Phila) 2018;57(1):11–18. doi: 10.1177/0009922816687325. [DOI] [PubMed] [Google Scholar]
- 10.Hautz WE, Kämmer JE, Hautz SC, et al. Diagnostic error increases mortality and length of hospital stay in patients presenting through the emergency room. Scand J Trauma Resusc Emerg Med. 2019;27(1) doi: 10.1186/s13049-019-0629-z. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 11.Saposnik G, Redelmeier D, Ruff CC, Tobler PN. Cognitive biases associated with medical decisions: a systematic review. BMC Med Inform Decis Mak. 2016;16:138. doi: 10.1186/s12911-016-0377-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 12.Kostopoulou O, Porat T, Corrigan D, Mahmoud S, Delaney BC. Diagnostic accuracy of GPs when using an early-intervention decision support system: a high-fidelity simulation. Br J Gen Pract. 2017;67:e201–e208. doi: 10.3399/bjgp16X688417. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 13.Porat T, Delaney B, Kostopoulou O. The impact of a diagnostic decision support system on the consultation: perception of GPs and patients. BMC Med Inform Decis Mak. 2017:17. doi: 10.1186/s12911-017-0477-6. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14.Soldaini L, Goharian N. QuickUMLS: a fast, unsupervised approach for medical concept extraction. Proceedings of the 2nd SIGIR workshop on Medical Information Retrieval (MedIR) 2016 [Google Scholar]
- 15.Bethesda, MD: NIH US National Library of Medicine; NIH US National Library of Medicine. UMLS REST API Home Page [Internet] [cited 2019 Aug 09]. Available from: https://documentation.uts.nlm.nih.gov/rest/home.html. [Google Scholar]
- 16.Bethesda, MD: NIH US National Library of Medicine; NIH US National LIbrary of Medicine. Abbreviations Used in Data Elements - 2019AB Release [Internet] [cited 2019 Nov 20]. Available from: https://www.nlm.nih.gov/research/umls/knowledge_sources/metathesaurus/release/abbreviations.html. [Google Scholar]
- 17.Hagberg AA, Schult DA, Swart PJ. Varoquaux G, Vaught T, Millman J. Exploring network structure, dynamics, and function using NetworkX. Proceedings of the 7th Python in Science Conference (SciPy2008) 2008:11–15. Aug; Pasadena, CA. [Google Scholar]
- 18.Bethesda, MD: NIH US National Library of Medicine; Lister Hill National Center for Biomedical Communications. SemRep [Internet]. [updated 2017 Dec 07; cited 2019 Aug 11]. Available from: https://semrep.nlm.nih.gov. [Google Scholar]
- 19.Bethesda, MD: NIH US National Library of Medicine; Lister Hill National Center for Biomedical Communications. Semantic MEDLINE Database [Internet] [updated 2018 May 03; cited 2019 Aug 11]. Available from: https://skr3.nlm.nih.gov/SemMedDB/ [Google Scholar]