Skip to main content
Springer logoLink to Springer
. 2025 Jul 31;167(1):209. doi: 10.1007/s00701-025-06631-3

Large language models for extraction of OPS-codes from operative reports in meningioma surgery

Sebastian Lehmann 1,, Florian Wilhelmy 1, Nikolaus von Dercks 2, Erdem Güresir 1, Johannes Wach 1
PMCID: PMC12313760  PMID: 40742565

Abstract

Background

In the German medical billing system, surgical departments encode their procedures in OPS-codes. These OPS-codes have major impact on DRG grouping and thus mainly determine each case´s revenue. In our study, we investigate the ability of the Large Language Model (LLM) GPT to derive correct OPS codes from the surgical report.

Methods

For our study, 100 patients who underwent meningioma surgery at our clinic between 2023 and 2024 were examined. We recorded the OPS codes assigned by the surgeon after the procedure, as well as the final coding by the hospital´s coders before case closure. In addition, the surgical report was extracted and anonymously provided to GPT-4o and GPT CodeMedic together with the current OPS-catalogue. The coding of each group was analyzed descriptively and compared using the Chi-Square test. Additionally, errors and deviations were assessed and analyzed.

Results

In our analyses, coders (100%) and surgeons (99%) demonstrated to significantly perform higher than LLMs in sufficient coding, for which the basic coding must be correct and unquestionable (GPT-4o 78%, GPT CodeMedic 89%; p < 0.01). For optimal coding, where every code potentially contributing to increase the revenue must be included, only the coders (94%) achieved superiority (GPT-4o p < 0.01; GPT CodeMedic p = 0.02), whereas GPT CodeMedic (83%) even outperformed surgeons (69%) (p = 0.03). The specialized GPT CodeMedic tends to show fewer hallucinations compared to GPT-4o (7% vs. 15%).

Conclusion

GPT is capable of extracting OPS codes from surgical reports. The most frequent errors made by LLMs can be attributed to a lack of specialized training. Currently, professional coders still significantly outperform LLMs in sufficient and optimal coding. For optimal coding however, GPT shows to perform comparably to surgeon´s coding skills. This indicates, that in near future after further training, LLMs may take over this task from surgeons without loss in quality.

Graphical abstract

Large language models for extraction of OPS-codes from operative reports in meningioma surgery

graphic file with name 701_2025_6631_Figa_HTML.jpg

Supplementary Information

The online version contains supplementary material available at 10.1007/s00701-025-06631-3.

Keywords: OPS-Coding, Artificial intelligence, GPT, Meningioma surgery

Introduction

The German system for accounting revenue for hospital services is based on classification of cases into “Diagnosis Related Groups” (DRGs). Each DRG is based on a multitude of patient-specific and disease-specific factors [12, 18, 21]. In surgical disciplines, the procedure performed mainly contributes to the classification in the respective DRG. These procedures are coded using OPS-codes that base on the OPS catalogue, annually published by the Federal Institute for Drugs and Medical Devices (BfArM) [7, 13].

In hospital routine, correctly coding a patient’s case is a multi-step procedure. The initial OPS coding is performed by the surgeon that carried out the operation. This initial coding is implemented into the DRG by the hospitals grouping system and cross-checked by the hospital´s coding professionals. Before revenue can be paid, a subset of cases is reviewed by the MDK (Medical Service of Health insurance providers). If incorrect coding detected, it may lead to a reduction in the case value and a fine. Thus correct and efficient coding is a vital process in clinical routine [11].

At a time when medical practice is highly bureaucratized, artificial intelligence is increasingly finding its way into clinical practice [29]. AIs that are able to analyze complex information, understanding context and processing content may be able to make our everyday life in hospitals more efficient, less bureaucratic and more patient-oriented in the future [2, 5, 30].

First studies testing LLMs coding abilities were recently published [25]. While we see an improvement in coding while the LLM progress [16], first studies indicate a superiority of GPT in ICD (international classification of diseases)- coding compared to humans and established coding systems [1, 28].

To our knowledge, this study is the first to analyze the capability of GPT extracting OPS codes from brain tumor related operational reports.

Methods

Study design

In this retrospective study, we analyzed 100 surgical reports of patients who underwent meningioma resection (ICD D32.0) in our clinic in the time from January 2023 to December 2024. Sphenoorbital, spinal, intraventricular meningiomas and tumors of the posterior fossa were not included.

Human and artificial coders of surgical reports

We assessed the four groups for comparison of coding accuracy: 1. The surgeon assigning OPS codes to the performed procedure. 2. The professional coders revising the coding before closing the case. 3. The LLM GPT in official Version GPT-4o [23]. GPT in the custom-refined version “CodeMedic”, that was published in 2023 [8]. This GPT Version has been trained on medical datasets and fine-tuned by supervised learning specialising in medical ICD and OPS coding At the time of our analysis, GPT CodeMedic was uniquely described to be optimised for processing OPS coding.

Endpoints

For our analysis we defined two endpoints. We defined"sufficient coding", that must contain a correct procedure code (e.g. 5–015.3, removal of tumor of the meninges without infiltration of adjacent tissue), as well as a correct approach code (e.g. 5–010.11, craniotomy via the midline). Additional codes for the description of accessory devices and techniques (e.g. 5–984, microsurgical procedure, 5–988 neuronavigation or 8–925 neuromonitoring) are optional [6]. No additional code may be specified that is not explicitly described in the OR. Thus, sufficient coding includes basic coding that cannot be questioned but potentially does not utilize the revenue of the procedure."Optimal coding"however includes the correct representation of the procedure code, the approach code and all additional codes that comprehensively describe the procedure. Likewise, no codes may be used that are not explicitly described in the surgical report. After extraction of the surgeons and coders coding from each file’s records, the anonymized operational report was provided to GPT (exemplary Chat prompt see suppl. Figure 1). To support GPT´s context understanding, the OPS-Catalogue and our internal list of frequently used codes were provided [6] (suppl. Table 1/2). In the time period from 2022 to 2025, no changes relevant for our research were made in the OPS-catalogue [20]. We generated a prompt instructing GPT to extract a list of the correct OPS-Coding from the operational report. In case where neuromonitoring was used, an additional note was provided as the use and duration is not documented as standard in our operative reports (Suppl. Figure 1). The consultation of GPT took place in February and March 2025.

Statistics

For analyses, the data were anonymously dichotomized into correct or incorrect extraction and recorded in a computerized database. We carried out further analyses using SPSS (IBM Corp., IBM SPSS Statistics for Windows, Version 29.0.2.0, Armonk, NY: USA) for descriptive analyses of coding capabilities and error distributions. Subsequently, group comparisons were performed for statistical significance using chi-square testing (two-sided p-values are reported). Additionally, the hallucination rates of the LLMs were assessed.

Results

We analyzed a total of 100 surgical reports. Tumor locations were distributed as follows: 44 convexity meningiomas (including 6 in the central region), 30 sphenoid wing meningiomas, 10 falx meningiomas, 9 located at the sphenoid plane or orbital roof, 6 in the olfactory groove and 1 at the tuberculum sellae. Two operations were performed for tumor recurrence.

Sufficient coding was achieved in 99–100% by surgeons and professional coders compared to 78–89% by the LLMs. GPT CodeMedic outperformed GPT-4o by more than 11% in this category. For optimal coding, professional coders achieve the highest performance (94%), while surgeons showing the highest error rate (69%). Again, GPT CodeMedic outperformed GPT-4o by more than 10% (Fig. 1).

Fig. 1.

Fig. 1

Illustration of sufficient and optimal coding for patients operated on with ICD32.0 (benign meningeal tumor); Sufficient coding: Coder vs. GPT-4o p < 0.01; Surgeon vs. GPT-4o p < 0.01; Coder vs. GPT CodeMedic p < 0.01; Surgeon vs. GPT CodeMedic p < 0.01; Coder vs. Surgeon p = 1.0; GPT-4o vs. GPT CodeMedic p = 0.06. Optimal coding: Coder vs. GPT-4o p < 0.01; Surgeon vs. GPT-4o p = 0.88; Coder vs. GPT CodeMedic p = 0.02; GPT CodeMedic vs. Surgeon p = 0.03; Coder vs. Surgeon p < 0.01; GPT-4o vs. GPT CodeMedic p = 0.06

In cross-tabulation analysis, professional coders were significantly superior to both LLMs in sufficient (p < 0.01) and optimal coding (GPT-4o: p < 0.01; GPT CodeMedic: p = 0.02). Surgeons showed significant superiority in sufficient coding (p < 0.01), but performed significantly worse in optimal coding. In this category, GPT CodeMedic was even significantly superior to surgeons (GPT-4o: p = 0.88; GPT CodeMedic: p = 0.03), as well as Coders (p < 0.01). In sufficient coding, the performance of surgeons did not differ significantly from GPT-4o (p = 0.88). For optimal as well as sufficient coding, GPT CodeMedic performs better than GPT-4o, with statistical significance only narrowly missed (p = 0.06) (Fig. 1; Suppl. Table 3).

Error source analysis revealed distinct patterns between human and machine coders. For surgeons, the most common error was the inaccurate or missing coding of the neuromonitoring (84%). For LLMs, most errors resulted from miscoding of procedures (66–80%). Among professional coders, the most frequent errors were uncoded tumor recurrences. However, this and missing coding of duroplasties (3%) hat minor overall impact. Hallucinations were observed in 7% (GPT CodeMedic) to 15% (GPT-4o), contributing notably to the incorrectly coded procedures (Fig. 2).

Fig. 2.

Fig. 2

Illustration of the error distribution in the coding for patients operated on with ICD D32.0 (benign neoplasm of the meninges)

In the present analysis including a dataset of 100 operative cranial meningioma procedures, each requiring approximately three minutes of a senior physician’s time for accurate coding. This totals 300 min, or five hours of labor. With a monthly salary of €9,302.27 (TVÄ3-Stage 1) and an estimated 160 working hours per month, the hourly wage of a senior physician is approximately €58.14. Consequently, the total cost incurred for manual coding across 100 procedures amounts to €290.70 in physician time (Table 1). In a high-volume university hospital, where approximately 300 brain tumor surgeries are performed each year, this results in a total of 900 min or 15 h of specialized labor annually. Therefore, the direct labor cost of manual coding for 300 patients amounts to €872.10 per year in a high-volume neurosurgical center. Calculated for procedures in a mid-sized department with approximately 1800 surgical Interventions/year the costs add up to €5,232.6.

Table 1.

Estimated cost savings from automated coding for the present meningioma cohort

Item Value
Number of procedures 100
Time per procedure (manual coding) 3 min
Total time saved 300 min (5 h)
Senior physician monthly salary (TVÄ3-Stage 1) €9,302.27
Estimated monthly working hours 160 h
Hourly wage €58.14
Total labor cost saved €290.70

Table 1: Potential cost savings from automated coding for the present meningioma cohort and in a high-volume brain tumor center

Discussion

Our analysis shows a clear superiority of the professional coders to GPT in both sufficient and optimal coding. Surgeons however exhibit a 30% gap between sufficient and optimal coding. Here the surgeons do not perform better than GPT-4o, on the contrary, GPT CodeMedic even achieves statistical superiority.

The error analysis helps to explain these findings. While coders achieve near-perfect results, the errors of surgeon´s and LLM´s are attributable to divergent causes. Coding for meningioma resection is generally straightforward and standardized. Common sources of error—such as recurrence coding—are less frequent in this context.

The most common error for surgeons coding meningioma resections was the correct coding of the neuromonitoring used. However, as this is specifically documented in the casefile, these errors were almost always recognized and corrected by the professional coders.

For LLMs, the most common error was the incorrect application of a code for cranioplasty. This appears to be triggered by a detailed description of bone-flap reinsertion. Tough not entirely being hallucinatory, as the code itself exists, it is wrongly addressed rendering the coding invalid as it is challengeable by the MDK. This illustrated a key limitation of the LLM, where they are not able to distinguish subtle but coding relevant distinctions in surgical language. Here, precise descriptions in the operational reports are vital to prevent miscoding.

Another issue observed is that of true hallucination, where fictional yet superficially plausible codes like: “5–830.4: Haemostasis with bipolar coagulation” are invented by the model. This phenomenon is especially prominent when a LLM assesses a task it is not specifically trained for [3, 4], The possible reduction of hallucination by training may be depicted by our data, as GPT CodeMedic, that was modified to improve interaction with medical coding, especially the German OPS-coding system, shows fewer hallucinations (7%) than the untrained GPT-4o (15%).

LLM use in medical coding is a rapidly evolving field which was still in its early stages only a few years ago [9]. 2024 Soroush et al. [27] published a large series of several LLM performing code extraction for medical diagnosis across different medical coding systems showing results from 1.2–45.9% exact match rates, identifying GPT as the best performing LLM. The ICD code extraction abilities were shown to be improved by provision of additional context enhancement in terms of lead term definition or RAG-based code assignment [15, 24]. If the code extraction was not performed from a digital patient file but from a single report such as discharge summaries, results with high rates of accuracy from over 90% were achieved [24, 28].

While the extraction of ICD 10-based diagnosis coding has been well documented in studies [27], there are currently only a few studies on the extraction of procedure codes. Lehnen et al. [17] show that GPT was able to extract procedural data from operative reports. In comparison to diagnostic code extraction, procedure code extraction showed to be associated with a significantly higher error rate [22]. With the use of one comprehensive report and context enhancement we combine previously identified factors increasing coding effectiveness. The individual effect of these measures however is unclear. To what extent our findings can be generalized to other countries'coding systems remains unclear, as our study uniquely evaluates OPS-based code extraction in the relatively unexplored field of LLM-based procedure code analysis in Germany.

It is likely, that with advancing technology as well as the use of tailored approaches including context enrichment, decision pathway instructions and self-review feedback loops, GPTs accuracy will rise further. This may contribute to the reduction of the bureaucratic burden on a surgeons shoulders like hinted by Dubinski et al., in whose study LLMs successfully created operative reports and discharge letters [10]. In Germany, there are several automated coding tools that can generate the appropriate coding for billing based on machine-readable documents from the hospital information system. What they have in common is that procedures (OPS) are less easily recognised than diagnoses (ICD). This is probably due to the simpler assignment of diagnosis and code, but also to the problem of the heterogeneous description of procedures in the original document.

Our results emphasize to main keys: Surgeon’s coding accuracy may be greatly improved by creating awareness for frequently overlooked codes. LLMs however perform at a level comparable to or exceeding surgeons in optimal coding. However, sophisticated and precise domain-specific training appears to be the key to further improve GPTs coding abilities before being introduced into clinical practise. Further studies are required to examine the transferability of the results to other fields of operation and to other classification systems, such as the German-language Swiss CHOP system [26], the Austrian BMSGPK-based system [19] or the ICD-10-PCS system used in the United States [14].

By even implementing conventional software that automates this coding process, the annual 15 h of highly qualified personnel time could be reallocated to more complex clinical tasks or patient care, translating directly into financial savings. Although this estimated example covers only 300 min, the cumulative savings become significantly more impactful when scaled to larger patient volumes including other procedures (spinal, functional, peripheral nerve, trauma and neurovascular surgery) or longer time periods. Additionally, automated systems offer advantages in standardization, reduced error rates, and faster throughput. This underscores the value of investing in AI-driven coding tools, especially in high-volume surgical specialties such as neurosurgery, where accurate and efficient documentation is directly tied to DRG-based reimbursement and administrative workload reduction.

Conclusion

Our study shows that LLM possess substantial potential in extracting procedure depicting codes from operational report. However, sophisticated training is mandatory to eliminate unnecessary mistakes and hallucinations. We believe that with more comprehensive training OPS Coding can be a robust target for LLM to relieve surgeons of bureaucratic tasks in clinical routine.

Supplementary Information

Below is the link to the electronic supplementary material.

Abbreviations

GPT

Generative Pre-trained Transformer

LLM

Large language model

OPS

Operationen und Prozedurenschlüssel

ICD

International Statistical Classification of Diseases

Author contributions

Conceptualization: S.L., F. W., J. W.. Investigation: S. L., Visualization: S. L., J. W.. Supervision: J. W., N.D., E. G.. Writing-original draft: S... Writing-review & editing: S. L., N.D.,J. W.

Funding

Open Access funding enabled and organized by Projekt DEAL.

Data availability

The datasets generated during the current study are available from the corresponding author on reasonable request.

Declarations

Compliance with ethical standards

The manuscript has been prepared in accordance with the instructions provided by the authors and has been approved by all of them. All of the requirements set out by the authors have been met and approved by all of them. The authors have no conflicts of interest to declare. All co-authors have seen and agree with the content of the manuscript and there is no financial interest to report. We certify that the submission is original work and is not under review at any other publication.

Human ethics and consent to participate declarations

Not applicable, as this is a retrospective study and no additional patient data was raised.

Institutional review board statement

The study was conducted in accordance with the Declaration of Helsinki. IRB approval was obtained by the local ethics committee of the University of Leipzig, medical faculty under the chairmanship of Prof. Dr. Dr. Ortrun Riha (No.: 055/25-ek).

Competing interests

The authors declare no competing interests.

Footnotes

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

References

  • 1.Abdelgadir Y, Thongprayoon C, Miao J et al (2024) AI integration in nephrology: evaluating ChatGPT for accurate ICD-10 documentation and coding. Front Artif Intell. 10.3389/frai.2024.1457586 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 2.Al Kuwaiti A, Nazer K, Al-Reedy A et al (2013) A Review of the Role of Artificial Intelligence in Healthcare. J Pers Med 13(6). 10.3390/jpm13060951 [DOI] [PMC free article] [PubMed]
  • 3.Arshad HB, Butt SA, Khan SU, Javed Z, Nasir K (2023) ChatGPT and artificial intelligence in hospital level research: potential, precautions, and prospects. Methodist Debakey Cardiovasc J 19(5):77–84. 10.14797/mdcvj.1290 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4.Athaluri SA, Manthena SV, Kesapragada VSRKM, Yarlagadda V, Dave T, Duddumpudi RTS (2023) Exploring the boundaries of reality: investigating the phenomenon of artificial intelligence hallucination in scientific writing through ChatGPT references. Cureus 15(4):e37432. 10.7759/cureus.37432 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5.Bhattacharya K, Bhattacharya AS, Bhattacharya N, Yagnik VD, Garg P, Kumar S (2023) ChatGPT in surgical practice—a new kid on the block. Indian J Surg 85(6):1346–1349. 10.1007/s12262-023-03727-x [Google Scholar]
  • 6.BfArM. OPS Version 2025 Systematisches Verzeichnis. Accessed March 31, 2025. https://www.dkgev.de/fileadmin/default/BfArM_-_OPS_Version_2025.pdf
  • 7.Bundesinstitut für Arzneimittel und Medizinprodukte. (2024) Operationen- und Prozedurenschlüssel: internationale Klassifikation der Prozeduren in der Medizin. [(OPS)Systematisches Verzeichnis];18. Oktober
  • 8.ChatGPT. Published March 4, 2025. Accessed March 4, 2025. https://chatgpt.com/g/g-E3rN5KmCT-codemedic
  • 9.Dong H, Falis M, Whiteley W et al (2022) Automated clinical coding: what, why, and where we are? NPJ Digit Med. 10.1038/s41746-022-00705-7 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10.Dubinski D, Won S-Y, Trnovec S et al (2024) Leveraging artificial intelligence in neurosurgery-unveiling ChatGPT for neurosurgical discharge summaries and operative reports. Acta Neurochir 166(1):38. 10.1007/s00701-024-05908-3 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11.Gaertner T, Lehr F, Blum B, van Essen J (2015) Einheitlichkeit der richtlinienkonformen Begutachtung durch den Medizinischen Dienst der Krankenversicherung bei der Prüfung der Kodierung im G-DRG-System. Gesundheitswesen 77:559–564. 10.1055/s-0034-1398553 [DOI] [PubMed] [Google Scholar]
  • 12.Goldfield N (2010) The evolution of diagnosis-related groups (DRGs): from its beginnings in case-mix and resource use theory, to its implementation for payment and now for its current utilization for quality within and outside the hospital. Qual Manag Health Care 19(1):3–16. 10.1097/QMH.0b013e3181ccbcc3 [DOI] [PubMed] [Google Scholar]
  • 13.Graubner B (2007) ICD und OPS. Historische Entwicklung und aktueller Stand. Bundesgesundheitsbl Gesundheitsforsch Gesundheitsschutz 50(7):932–943. 10.1007/s00103-007-0283-x [DOI] [PubMed] [Google Scholar]
  • 14.ICD-10 | CMS. Published July 16, 2025. Accessed July 16, 2025. https://www.cms.gov/medicare/coding-billing/icd-10-codes
  • 15.Klang E, Tessler I, Apakama DU, et al. Assessing Retrieval-Augmented Large Language Model Performance in Emergency Department ICD-10-CM Coding Compared to Human Coders
  • 16.Lehnen NC, Dorn F, Wiest IC et al (2024) Data extraction from free-text reports on mechanical thrombectomy in acute ischemic stroke using ChatGPT: a retrospective analysis. Radiology 311(1):e232741. 10.1148/radiol.232741 [DOI] [PubMed] [Google Scholar]
  • 17.Lehnen NC, Dorn F, Wiest IC et al (2024) Data extraction from free-text reports on mechanical thrombectomy in acute ischemic stroke using ChatGPT: a retrospective analysis. Radiology. 10.1148/radiol.232741 [DOI] [PubMed] [Google Scholar]
  • 18.Lerch MM, Rathmayer M, Siegmund B et al (2020) Die Grenzen des G-DRG-Systems bei der Abbildung von Komplexität in der Universitätsmedizin. Z Gastroenterol 58:747–753. 10.1055/a-1219-8245 [DOI] [PubMed] [Google Scholar]
  • 19.Mag. Stefan Eichwalder, Leistungskatalog BMSGPK 2025, Bundesministerium für Soziales, Gesundheit, Pflege und Konsumentenschutz (BMSGPK), Stubenring 1, 1010 Wien, 1. Jänner 2025, OID 1.2.40.0.34.5.229
  • 20.Medizinische Klassifikationen zum Download – bifg. Published July 16, 2025. Accessed July 16, 2025. https://www.bifg.de/daten-und-analysen/klassifikationen
  • 21.Munoz E, Boiardo R, Mulloy K, Goldstein J, Tenenbaum N, Wise L (1990) Economies of scale, physician volume for neurosurgery patients, and the diagnosis-related group prospective hospital payment system. Neurosurgery 26(1):156–60. 10.1097/00006123-199001000-00027 [DOI] [PubMed] [Google Scholar]
  • 22.Nascimento Teixeira B, Leitão A, Nascimento G, Campos-Fernandes A, Cercas F (2024) Can ChatGPT support clinical coding using the ICD-10-CM/PCS? Informatics. 10.3390/informatics11040084 [Google Scholar]
  • 23.OpenAI, Achiam J, Adler S et al (2023) GPT-4 Technical Report. http://arxiv.org/pdf/2303.08774
  • 24.Puts S, Zegers CML, Dekker A, Bermejo I. Developing an ICD-10 Coding Assistant: Pilot Study Using RoBERTa and GPT-4 for Term Extraction and Description-Based Code Selection [DOI] [PMC free article] [PubMed]
  • 25.Roy JM, Self DM, Isch E et al (2025) Evaluating large language models for automated CPT code prediction in endovascular neurosurgery. J Med Syst 49(1):15. 10.1007/s10916-025-02149-4 [DOI] [PubMed] [Google Scholar]
  • 26.Schweizerische Eidgenossenschaft, Schweizerische Operationsklassifikation (CHOP), Systematisches Verzeichnis, Bundesamt für Statistik BFS, Eidgenössisches Departement des Innern EDI, Version 2025, BFS-Nummer 659-2400. Accessed 15, July 2025
  • 27.Soroush A, Glicksberg BS, Zimlichman E, et al. Large Language Models Are Poor Medical Coders — Benchmarking of Medical Code Querying
  • 28.Wang Y, Huang Y, Nimma IR et al (2024) Validation of GPT-4 for clinical event classification: a comparative analysis with ICD codes and human reviewers. J Gastroenterol Hepatol 39(8):1535–1543. 10.1111/jgh.16561 [DOI] [PubMed] [Google Scholar]
  • 29.Wójcik S, Rulkiewicz A, Pruszczyk P, Lisik W, Poboży M, Domienik-Karłowicz J (2023) Beyond ChatGPT: What does GPT-4 add to healthcare? The dawn of a new era. Cardiol J 30(6):1018–1025. 10.5603/cj.97515 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 30.Xue VW, Lei P, Cho WC (2023) The potential impact of ChatGPT in clinical and translational medicine. Clin Transl Med 13(3):e1216. 10.1002/ctm2.1216 [DOI] [PMC free article] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Data Availability Statement

The datasets generated during the current study are available from the corresponding author on reasonable request.


Articles from Acta Neurochirurgica are provided here courtesy of Springer

RESOURCES