Skip to main content
The British Journal of Surgery logoLink to The British Journal of Surgery
letter
. 2025 Jul 17;112(7):znaf139. doi: 10.1093/bjs/znaf139

Potential for error when relying on administrative data

Hila Zelicha 1, Douglas S Bell 2,3, Yijun Chen 4, Edward H Livingston 5,
PMCID: PMC12268322  PMID: 40673449

The widespread use of electronic medical records (EMR) and advanced billing systems has made large databases available for analysing medical care outcomes1–3. Most administrative databases, such as the National Inpatient Survey and the Centers for Medicare & Medicaid Services data, rely on billing codes to identify diseases or procedures associated with patient care. However, a fundamental limitation of administrative database research is its reliance on billing codes. For any analyses of these databases, the precision and accuracy of disease identification by billing codes should be verified. It is uncommon for this validation step to be included in research reports relying on administrative data.

To assess the accuracy of diagnostic coding as reflecting the true presence of disease, we analysed the EMRs of a large, integrated health system, extracting records from primary care encounters with a primary diagnosis of hernia, with a subset of patients who had abdominal imaging. We focused on abdominal hernia, a common condition that can be reliably confirmed through imaging, making it an ideal model for assessing misclassification. We hypothesized that reliance on billing records in administrative data frequently misidentifies the presence of actual disease.

Patient-level data were extracted from the UCLA Epic Clarity EHR database, excluding restricted individuals (for example celebrities) under UCLA IRB approval (IRB#23-000174). The analytic sample included primary care clinic visits of patients >16 years diagnosed with a hernia between 1 January 2018 and 6 June 2023. Abdominal hernias were identified by ICD-10 codes K40–K46 as the primary diagnosis. Subtypes included diaphragmatic (K44.9), ventral (K42–K43, K45–K46), and inguinal/femoral (K40–K41). When there were multiple hernia diagnoses, the diagnosis listed first was used.

For patients with abdominal imaging (CT/US), hernia diagnoses were confirmed by natural language processing (NLP) of dictated radiology reports. Sentences were parsed, lowercased, and searched for hernia-related terms. Words used to identify hernia from radiology reports: diaphragmatic—hiatal hernia, diaphragmatic hernia, hiatus, Bochdalek hernia, and paraesophageal hernia; ventral—umbilical hernia, umbilicus hernia, ventral hernia, abdominal wall hernia, Spigelian hernia, Richter hernia, quadrant hernia, lateral abdominal hernia, small fat containing upper abdominal hernia, epigastric hernia, and incisional hernia; inguinal—inguinal, groin, and femoral. Presence was determined by keyword context (for example ‘was present’ versus ‘no evidence of’), followed by manual review by two authors (H.Z., E.L.). For multiple imaging records, the earliest date was used as the diagnosis. Analyses were conducted using R Studio v4.3.1.

Of 1 362 440 patients whose records were examined, 41 703 had a hernia diagnosis based on encounter ICD-10 diagnostic coding (Fig. 1). Of these patients, 28 555 (68%) had corresponding imaging records available. Based on ICD-10 diagnostic codes for outpatient encounters there were: 12 819 (45%) diaphragmatic hernias, 6979 (24%) ventral hernias, and 8757 (31%) inguinal hernias. Imaging verified the presence of hernia in 10 234 cases, yielding a true positive rate of 368% for the diagnosis. Diaphragmatic hernias were verified in 34% of cases labelled as such in clinic encounters, versus 44% for ventral hernias and 32% for inguinal hernias (Fig. 1).

Fig. 1.

Fig. 1

Flow diagram of the clinic encounter ICD coding validated by imaging database

These findings show a very high error rate when assuming a disease entity is present based on ambulatory billing codes found in administrative records. This probably results from clinicians coding visit diagnoses based on a clinical problem being considered, irrespective of its actual presence. For example, if a patient is referred for groin pain because of a possible hernia, the visit will be coded as ‘hernia’ until the diagnosis is ruled out. Although subsequent visits may refine the diagnosis, the erroneous diagnosis remains in the record when retrospectively analysed.

These findings highlight a fundamental weakness in using administrative data for disease identification. Encounter coding occurs because a diagnosis is considered and not necessarily proven. We found that reliance on billing codes for hernia identification could result in two-thirds of cases being erroneously identified. This issue extends beyond hernia1–5, highlighting a serious limitation in using administrative data for clinical research. Validation of coding accuracy against actual disease presence is essential before assuming diagnosis validity.

Contributor Information

Hila Zelicha, Department of Surgery, Faculty of Health Sciences, UCLA, Los Angeles, California, USA.

Douglas S Bell, Department of Medicine, Division of General Internal Medicine, UCLA, Los Angeles, California, USA; Biomedical Informatics Program of the UCLA Clinical and Translational Science Institute (CTSI), UCLA, Los Angeles, California, USA.

Yijun Chen, Department of Surgery, Faculty of Health Sciences, UCLA, Los Angeles, California, USA.

Edward H Livingston, Department of Surgery, Faculty of Health Sciences, UCLA, Los Angeles, California, USA.

Funding

UCLA Department of Surgery Research Funds; National Center for Advancing Translational Sciences (NCATS) of the National Institutes of Health under the UCLA Clinical and Translational Science Institute grant number UL1TR001881.

Disclosure

The authors have no relevant conflicts of interest to report.

Ethical Consideration: UCLA IRB 23-000174.

Data Access: All data collected for this research are reported in the manuscript and accompanying supplement.

Data availability

The de-identified data underlying this article cannot be shared publicly due to the inclusion of personally identifying information.

References

  • 1. Rios-Diaz  A, Hsu  JY, Broach  R, Bormann  B, Fischer  J. P057 The accuracy of incisional hernia codes in administrative data: a validation study. Br J Surg  2021;108:znab395-053 [Google Scholar]
  • 2. Alotaibi  GS, Wu  C, Senthilselvan  A, McMurtry  MS. The validity of ICD codes coupled with imaging procedure codes for identifying acute venous thromboembolism using administrative data. Vasc Med  2015;20:364–368 [DOI] [PubMed] [Google Scholar]
  • 3. Thigpen  JL, Dillon  C, Forster  KB, Henault  L, Quinn  EK, Tripodis  Y  et al.  Validity of international classification of disease codes to identify ischemic stroke and intracranial hemorrhage among individuals with associated diagnosis of atrial fibrillation. Circ Cardiovasc Qual Outcomes  2015;8:8–14 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4. Lee  S, Shaheen  AA, Campbell  DJT, Naugler  C, Jiang  J, Walker  RL  et al.  Evaluating the coding accuracy of type 2 diabetes mellitus among patients with non-alcoholic fatty liver disease. BMC Health Serv Res  2024;24:218. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5. Widdifield  J, Ivers  NM, Young  J, Green  D, Jaakkimainen  L, Butt  DA  et al.  Development and validation of an administrative data algorithm to estimate the disease burden and epidemiology of multiple sclerosis in Ontario, Canada. Multiple Sclerosis Journal  2015;21:1045–1054 [DOI] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Data Availability Statement

The de-identified data underlying this article cannot be shared publicly due to the inclusion of personally identifying information.


Articles from The British Journal of Surgery are provided here courtesy of Oxford University Press

RESOURCES