Skip to main content
Journal of Orthopaedic Surgery and Research logoLink to Journal of Orthopaedic Surgery and Research
. 2025 May 28;20:531. doi: 10.1186/s13018-025-05955-1

The role of large language models in improving the readability of orthopaedic spine patient educational material

Melissa Romoff 1, Madison Brunette 1, Melanie K Peterson 1, Sohaib Z Hashmi 1, Michael S Kim 1,
PMCID: PMC12117680  PMID: 40426209

Abstract

Introduction

Patient education is crucial for informed decision-making. Current educational materials are often written at a higher grade level than the American Medical Association (AMA)-recommended sixth-grade level. Few studies have assessed the readability of orthopaedic materials such as American Academy of Orthopaedic Surgeons (AAOS) OrthoInfo articles, and no studies have suggested efficient methods to improve readability. This study assessed the readability of OrthoInfo spine articles and investigated the ability of large language models (LLMs) to improve readability.

Methods

A cross-sectional study analyzed 19 OrthoInfo articles using validated readability metrics (Flesch-Kincaid Grade Level and Reading Ease). Articles were simplified iteratively in three steps using ChatGPT, Gemini, and CoPilot. LLMs were prompted to summarize text, followed by two clarification prompts simulating patient inquiries. Word count, readability, and accuracy were assessed at each step. Accuracy was rated by two independent reviewers using a three-point scale (3 = fully accurate, 2 = minor inaccuracies, 1 = major inaccuracies). Statistical analysis included one-way and two-way ANOVA, followed by Tukey post-hoc tests for pairwise comparisons.

Results

Baseline readability exceeded AMA recommendations, with a mean Flesch-Kincaid Grade Level of 9.5 and a Reading Ease score of 51.1. LLM summaries provided statistically significant improvement in readability, with the greatest improvements in the first iteration. All three LLMs performed similarly, though ChatGPT achieved statistically significant improvements in Reading Ease scores. Gemini incorporated appropriate disclaimers most consistently. Accuracy remained stable throughout, with no evidence of hallucination or compromise in content quality or medical relevance.

Discussion

LLMs effectively simplify orthopaedic educational content by reducing grade levels, enhancing readability, and maintaining acceptable accuracy. Readability improvements were most significant in initial simplification steps, with all models performing consistently. These findings support the integration of LLMs into patient education workflows, offering a scalable strategy to improve health literacy, enhance patient comprehension, and promote more equitable access to medical information across diverse populations.

Introduction

Effective education is essential for supporting informed decision-making and protecting patient autonomy [15]. However, many educational materials remain inaccessible to the general population due to readability challenges [6]. The National Assessment of Adult Literacy reports that the average U.S. adult reads at an eighth-grade level [7], while Medicaid enrollees average at a fifth-grade level [8]. To address this discrepancy, the American Medical Association (AMA) recommends that patient educational materials be written at or below a sixth-grade reading level to ensure accessibility [9].

Despite these guidelines, a 2018 systematic review found that many health education resources are written at a 10 th- to 15 th-grade level [10]. In the field of orthopaedics, the American Academy of Orthopaedic Surgeons (AAOS) provides public-facing educational content through its OrthoInfo website. Previous studies have demonstrated that orthopaedic educational materials frequently exceed recommended reading levels [1119], but OrthoInfo’s spine content, despite its complexity and clinical importance, has not been systematically evaluated.

Recent work in other medical specialties has shown that large language models (LLMs) can improve the readability of patient education without compromising accuracy [2022]. However, adoption in clinical settings remains limited due to concerns surrounding factual reliability, medicolegal implications, and the absence of formal guidelines or validated implementation strategies [23]. These limitations underscore the need for additional research evaluating LLM performance across diverse clinical domains.

Although LLM-based readability analyses have been conducted in orthopaedics, prior studies have not specifically examined spine-related articles from OrthoInfo, a widely used, patient-facing educational platform. Given the procedural complexity and specialized terminology common in spine surgery, general findings from other fields – or even from broader orthopaedic topics – may not fully apply. In this study, we applied a structured, iterative prompting protocol across multiple LLMs to simplify OrthoInfo spine content, evaluating readability improvements, content accuracy, and the potential for effective integration of LLMs into orthopaedic patient education.

Methods

A cross-sectional observational study of public-facing educational material published by the American Academy of Orthopaedic Surgeons on OrthoInfo.AAOS.org was conducted. Institutional Review Board approval was not required as no protected health information was involved in this study. No large language models (LLMs) were used in drafting this manuscript text itself; however, LLMs were employed for content transformation as part of the study design.

All OrthoInfo web pages associated with the treatment of spinal conditions were identified by selecting “Treatment” from the homepage toolbar and reviewing articles under “Neck” and “Back.” Web pages presenting video content without text substantial enough to enable analysis were excluded. Articles were classified into three categories: 1) Background: Articles primarily providing general education about conditions; 2) Procedure: Articles primarily focused on surgical or non-surgical interventions; 3) Opinion: Articles offering physician perspectives or decision-making advice. Content from each article was extracted as plain text. Words and figures were counted. Date of last review was recorded.

Text was then submitted to the three most used LLMs: ChatGPT GPT-4o (OpenAI, San Francisco, CA), Copilot (Microsoft, Redmond, WA), and Gemini 1.5 Flash-8B (Alphabet, Mountain View, CA), preceded by the request “Please summarize the following.” A new chat session was opened for each article to prevent response bias. Summarization was prompted through three iterative steps: (1) “Please summarize the following,” (2) “I don’t understand, please clarify,” and (3) “I still don’t understand, please clarify.” Each response was documented and analyzed for word count, readability, accuracy, and the presence of an appropriate disclaimer.

Readability was evaluated using Readable (Added Bytes, Hassocks, UK), a software that calculates validated measures of readability. These included the Flesch-Kincaid Grade Level (estimated U.S. school grade level) and Flesch Reading Ease (scored from 0 to 100, with higher scores indicating easier readability). Given the inherent variability of measures of readability, an overall mean readability was calculated by averaging readability measures that provide estimated grade level as had been done in a previous study to allow direct comparison.

Accuracy was independently assessed by two physician authors using a three-point scale: summaries that were fully accurate without requiring correction received three points; those with minor inaccuracies or hallucinations received two points; and those with substantial inaccuracies or hallucinations that affected medical relevance received one point. A three-point scale was selected to facilitate consistent reviewer grading and focus on clinically meaningful content accuracy rather than stylistic variation. In cases of disagreement, scores were reviewed collaboratively, and a consensus was reached between the two reviewers and senior author to ensure consistency and minimize individual bias in the final dataset.

Statistical analysis was performed using R. Descriptive statistics, mean and standard deviations, were calculated for article characteristics and LLM outputs. One-way and two-way ANOVA were used to compare readability, word count, and accuracy across LLMs and steps. Two-way ANOVA was used specifically to evaluate interaction effects between model and simplification step. Assumptions of normality and homogeneity of variants were assessed prior to performing ANOVA tests. Significant ANOVA results were followed by Tukey’s post-hoc tests for pairwise comparisons. Linear regression evaluated trends between readability metrics and baseline article characteristics. Statistical significance was defined as p < 0.05, and visualizations were generated using GraphPad Prism (GraphPad Software, San Diego, CA).

Results

Baseline characteristics

Out of the 23 webpages initially reviewed, 19 met the inclusion criteria (8 procedure, 10 background, and 1 ortho-opinion), while 4 were excluded due to being video-based content with minimal written material. No hallucinations were observed during the analysis.

The baseline readability analysis of the 19 included articles revealed a mean Flesch-Kincaid Grade Level of 9.5 (SD: 1.4), exceeding the AMA’s recommended sixth-grade level, and a mean Flesch Reading Ease score of 51.1 (SD: 7.5), indicating “fairly difficult” readability (Table 1, Fig. 1). On average, the articles contained 1,269 words and included 2.8 figures (Table 1). Procedure articles were shorter, averaging 1,033 words, but were more difficult to read (Grade Level: 9.8; Reading Ease: 49) and contained more figures. (4 per article) (Fig. 1). In contrast, Background articles were longer, averaging 1,523 words, slightly easier to read (Grade Level: 8.7; Reading Ease: 54), and included fewer figures. (2 per article) (Fig. 1).

Table 1.

Baseline Descriptive Statistics for Readability Metrics

Metric Overall (Mean ± SD) Procedure (Mean ± SD) Background (Mean ± SD)
Word Count 1269 ± 493.86 1033 ± 348.24 1523 ± 474.56
2.79 ± 2.49 4.0 ± 1.28 2.0 ± 1.02
Flesch-Kincaid Grade 9.52 ± 1.36 9.8 ± 1.18 8.65 ± 1.12
Flesch Reading Ease 51.07 ± 7.48 49.0 ± 5.85 54.0 ± 8.11

Fig. 1.

Fig. 1

Baseline readability characteristics of OrthoInfo educational materials. Readability metrics are shown for overall (n = 19), procedure (n = 8), and background (n = 10) sections. AMA recommendations are indicated by the red dashed line, and error bars represent standard deviations. A Mean Flesch-Kincaid Grade Level for overall, procedure, and background sections. B Mean Flesch Reading Ease scores for overall, procedure, and background sections

Statistical tests

T-tests revealed that Procedure articles had significantly fewer words than Background articles (p = 0.03), but no significant differences were found in figures, readability scores, or grade levels. Pearson correlation analyses showed no significant relationships between readability metrics (Flesch-Kincaid Grade Level and Flesch Reading Ease) and variables such as word count, number of figures, or review age (all p > 0.05).

LLM analysis

Each LLM (ChatGPT, Gemini, CoPilot) was evaluated independently across simplification steps. Significant word count reductions were observed for all models (p < 0.0001), with the largest reduction occurring between the baseline and Step 1, followed by a plateau in subsequent steps (Table 2, Fig. 2A). Reading Ease improved significantly across steps (p < 0.0001), with the greatest improvement between the original articles and Step 1, and smaller but still significant gains from Step 1 to Step 2 (Table 3, Fig. 2B). Grade Level changes were significant (p < 0.0001) across steps, but an unexpected initial increase was observed between the original articles and Step 1, with a negative mean difference indicating movement away from the AMA’s sixth-grade recommendation. Subsequent steps showed gradual reductions, ultimately trending toward the desired grade level (Table 3, Figs. 2C, 3, 4 and 5).

Table 2.

ANOVA Results for Readability Metrics Across ChatGPT, Gemini, and CoPilot

Metric LLM F-Statistic P-value Significant?
Word Count ChatGPT 89.54  < 0.0001 ****
Gemini 95.87  < 0.0001 ****
CoPilot 86.89  < 0.0001 ****
Grade Level ChatGPT 36.26  < 0.0001 ****
Gemini 17.23  < 0.0001 ****
CoPilot 35.26  < 0.0001 ****
Reading Ease ChatGPT 45.38  < 0.0001 ****
Gemini 23.03  < 0.0001 ****
CoPilot 26.84  < 0.0001 ****

Fig. 2.

Fig. 2

Readability and word count improvements across steps for ChatGPT, Gemini, and CoPilot. Mean values are presented for Baseline (Original) and Steps 1–3, with error bars representing standard deviations. A Mean Word Count for ChatGPT, Gemini, and CoPilot across steps. B Mean Flesch-Kincaid Reading Ease scores for ChatGPT, Gemini, and CoPilot across steps. C Mean Flesch-Kincaid Grade Level for ChatGPT, Gemini, and CoPilot across steps

Table 3.

Post-Hoc Results for One-way ANOVA Readability Metrics Across ChatGPT, Gemini, and CoPilot

Metric Comparison ChatGPT: Mean Difference (P-value) Gemini: Mean Difference (P-value) CoPilot: Mean Difference (P-value)
Word Count Baseline vs. Step 1 1019 (**** < 0.0001) 1071 (**** < 0.0001) 1031 (**** < 0.0001)
Baseline vs. Step 2 1110 (**** < 0.0001) 1133 (**** < 0.0001) 1102 (**** < 0.0001)
Baseline vs. Step 3 1119 (**** < 0.0001) 1114 (**** < 0.0001) 1129 (**** < 0.0001)
Grade Level Baseline vs. Step 1 −1.595 (* 0.0140) −1.626 (** 0.0089) −1.195 (ns 0.0977)
Baseline vs. Step 2 1.484 (* 0.0255) −0.07895 (ns 0.9986) 0.6842 (ns 0.5394)
Baseline vs. Step 3 3.463 (**** < 0.001) 1.942 (** 0.0012) 2.347 (*** 0.0001)
Reading Ease Baseline vs. Step 1 18.05 (**** < 0.0001) 16.34 (**** < 0.0001) 14.41 (**** < 0.0001)
Baseline vs. Step 2 −4.474 (ns 0.4819) 5.889 (ns 0.2836) 0.9579 (ns 0.9883)
Baseline vs. Step 3 −17.90 (**** < 0.0001) −10.19 (* 0.0141) −9.321 (* 0.0128)

Fig. 3.

Fig. 3

Performance of Gemini in improving readability and reducing word count across steps. Mean and standard deviations for readability metrics (Word Count, Flesch-Kincaid Reading Ease, and Flesch-Kincaid Grade Level) of Baseline articles and across Steps 1–3 using Gemini. Significance levels are indicated within the figure. Error bars represent standard deviations. A Word Count for Baseline and Gemini Steps 1–3. B Flesch-Kincaid Reading Ease scores for Baseline and Gemini Steps 1–3, with the AMA-recommended reading ease threshold shown as a red dashed line. C Flesch-Kincaid Grade Level for Baseline and Gemini Steps 1–3, with the AMA’s recommended sixth-grade reading level shown as a red dashed line

Fig. 4.

Fig. 4

Performance of CoPilot in improving readability and reducing word count across steps. Mean and standard deviations for readability metrics (Word Count, Flesch-Kincaid Reading Ease, and Flesch-Kincaid Grade Level) of Baseline articles and across Steps 1–3 using CoPilot. Significance levels are indicated within the figure. Error bars represent standard deviations. A Word Count for Baseline and CoPilot Steps 1–3. B Flesch-Kincaid Reading Ease scores for Baseline and CoPilot Steps 1–3, with the AMA-recommended reading ease threshold shown as a red dashed line. C Flesch-Kincaid Grade Level for Baseline and CoPilot Steps 1–3, with the AMA’s recommended sixth-grade reading level shown as a red dashed line

Fig. 5.

Fig. 5

Performance of ChatGPT in improving readability and reducing word count across steps. Mean and standard deviations for readability metrics (Word Count, Flesch-Kincaid Reading Ease, and Flesch-Kincaid Grade Level) of Baseline articles and across Steps 1–3 using ChatGPT. Significance levels are indicated within the figure. Error bars represent standard deviations. A Word Count for Baseline and ChatGPT Steps 1–3. B Flesch-Kincaid Reading Ease scores for Baseline and ChatGPT Steps 1–3, with the AMA-recommended reading ease threshold shown as a red dashed line. C Flesch-Kincaid Grade Level for Baseline and ChatGPT Steps 1–3, with the AMA’s recommended sixth-grade reading level shown as a red dashed line

All LLMs performed similarly in reducing word count, improving reading ease, and lowering grade levels, with no statistically significant differences observed between models or article types. However, a two-way ANOVA revealed that Gemini provided significantly more disclaimers than ChatGPT (mean difference = 10.0, p = 0.002) and CoPilot (mean difference = 8.3, p = 0.005), while no difference was observed between ChatGPT and CoPilot (Table 4). Simplification steps did not influence disclaimer frequency (p = 0.71) (Table 4).

Table 4.

ANOVA and Post-Hoc Results for Disclaimer Frequency

Effect/Comparison F-value P-value Mean Difference Adjusted P value Significance
ANOVA: LLM (Row Factor) 40.79 0.0022 N/A N/A N/A
ANOVA: Step (Column Factor) 0.368 0.7131 N/A N/A N/A
ChatGPT vs Gemini N/A N/A −10.00 0.0024 **
Gemini vs CoPilot N/A N/A 8.333 0.0048 **
ChatGPT vs CoPilot N/A N/A −1.667 0.4206 No

LLM specific findings

The two-way ANOVA compared all models (ChatGPT, Gemini, and CoPilot) across steps to assess differences in performance. Significant reductions in Word Count were observed across steps (p < 0.0001), with no significant differences between models (p = 0.80) or interaction effects (p = 0.997) (Supplementary Table 1). Reading Ease improved significantly across steps (p < 0.0001), with minor differences between models (p = 0.03) and a significant interaction effect (p = 0.03) (Supplementary Table 1); post-hoc tests indicated ChatGPT slightly outperformed Gemini (mean difference = 4.1, p < 0.05), while CoPilot performed similarly to both (Supplementary Table 2). Grade Level reductions were driven by steps (p < 0.0001), with small differences between models (p = 0.01) but no significant interaction effects (p = 0.11) (Supplementary Table 1); Post-hoc tests revealed that ChatGPT achieved slightly lower grade levels than Gemini (mean difference = −0.8, p < 0.05), with no significant differences involving CoPilot (Supplementary Table 2). Accuracy varied slightly across steps (p = 0.03) but showed no significant differences between models (p = 0.95) or interaction effects (p = 0.17) (Supplementary Table 1, Supplementary Table 3).

Discussion

This study evaluated the ability of large language models (LLMs)—ChatGPT, Gemini, and CoPilot—to simplify orthopedic educational content and improve accessibility for patients. The findings demonstrate that LLMs are effective at significantly improving readability metrics such as word count, grade level, and reading ease, with consistent performance across models.

Baseline analysis confirmed that OrthoInfo exceeded the AMA’s recommended sixth-grade reading level, highlighting the need for simplification to make educational materials more accessible. Significant improvements in readability metrics occurred across all models, particularly between the original text and Step 1, with diminishing returns in subsequent steps. While additional iterations may offer further refinement, our data suggest a plateau effect between Steps 2 and 4. We therefore limited the process to three simplifications to balance analytical depth and practical considerations. Future studies could explore extended prompting protocols to determine optimal iteration thresholds.

Interestingly, grade level initially increased after the first simplification step. This may reflect the LLMs’ tendency to introduce polysyllabic or technical language in early summarization attempts, particularly when lacking clear context about the target audience. These changes can inadvertently inflate readability scores, even if the overall content becomes more concise. Subsequent prompts likely encouraged more targeted simplification leading to final outputs that better aligned with the sixth-grade target. Despite, these fluctuations, accuracy remained stable across steps and models, suggesting that LLMS can simplify complex medical content without compromising factual integrity.

Minimal differences were observed between models. ChatGPT slightly outperformed Gemini in both improving reading ease and lowering grade levels, while CoPilot performed similarly to both models. Although some differences were statistically significant, their practical implications appear limited. Importantly, this study was not designed to establish the superiority of any specific LLM, but rather to descriptively assess their current capabilities in simplifying orthopaedic educational content. These findings suggest that the choice of model may be less critical than the simplification process itself.

Several limitations warrant discussion. This study focused exclusively on simplify existing content and did not assess LLMs’ ability to generate new or patient-specific materials. While readability metrics are valuable, they do not fully capture comprehension, particularly for individuals with limited health literacy. Furthermore, although new chat sessions were used for each article to minimize confounding, slight variability in LLM outputs is inherent to generative models. Given our standardized approach and outcome focus, this variability likely had minimal impact.

While this study focused on quantitative outcomes such as readability, word count, and accuracy, it did not include a quantitative or thematic analysis of how content was transformed. Future studies should explore how LLMs affect tone, phrasing and message structure in ways that may influence patient perception. Additionally, while we assessed accuracy using a standardized scale, future studies may benefit from more granular error classification to better understand how LLMs might alter critical content in patient facing materials.

Practical applications of these results involve integrating LLM-based simplification pipelines into electronic health record (EHR) systems to automatically generate accessible versions of home care instructions and procedure notes. Similarly, online educational platforms such as OrthoInfo could integrate LLM-driven readability checks and iterative simplifications prior to publication, ensuring that patient-facing materials meet health literacy standards. Future research should focus on adapting these tools to the needs of specific demographic groups and incorporating direct measures of user comprehension, such as patient surveys or health literacy assessments, to evaluate real-world applicability and maximize impact.

Conclusion

This study demonstrates the potential of large language models (LLMs) like ChatGPT, Gemini, and CoPilot to improve the readability of orthopedic educational content, addressing the gap between current materials and the AMA’s recommended sixth-grade reading level. By significantly reducing grade levels and improving reading ease, these models simplify medical content without compromising accuracy. While ChatGPT showed slightly greater readability, all models achieved progressive simplifications across steps, highlighting the feasibility of iterative text refinement.

While promising, this study’s focus on simplifying existing materials leaves unexplored opportunities for generating patient-specific content and addressing broader health literacy needs. Importantly, our findings do not assess how LLM-generated content performs across diverse patient demographics or literacy levels. Future research should focus on validating these findings through qualitative studies, including user testing among patients with varying literacy levels and direct patient input to assess comprehension. Systematic efforts are needed to refine LLMs and ensure patient education materials are accessible and effective for all populations.

Acknowledgements

Not applicable.

Authors’ contributions

M.R. performed data extraction and analysis, prepared figures and tables, and wrote the manuscript. M.B. performed data extraction and edited the manuscript. M.K.P. performed LLM accuracy assessment and edited the manuscript. S.Z.H.edited the manuscript and supervised the project. M.S.K conceived the project, performed LLM accuracy assessment, and wrote the manuscript.

Funding

The authors declare there was no funding for this study.

Data availability

No datasets were generated or analysed during the current study.

Declarations

Ethics approval and consent to participate

Not applicable.

Consent for publication

Not applicable.

Competing interests

The authors declare no competing interests.

Footnotes

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

References

  • 1.Entwistle VA, Carter SM, Cribb A, McCaffery K. Supporting patient autonomy: the importance of clinician-patient relationships. J Gen Intern Med. 2010;25(7):741–5. 10.1007/s11606-010-1292-2. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 2.Kukla R. Conscientious autonomy: displacing decisions in health care. Hastings Cent Rep. 2005;35(2):34–44. [PubMed] [Google Scholar]
  • 3.Beach MC, Inui T; Relationship-Centered Care Research Network. Relationship-centered care. A constructive reframing. J Gen Intern Med. 2006;21 Suppl 1(Suppl 1):S3-S8. 10.1111/j.1525-1497.2006.00302.x. [DOI] [PMC free article] [PubMed]
  • 4.Walker RL. Medical ethics needs a new view of autonomy. J Med Philos. 2008;33(6):594–608. 10.1093/jmp/jhn033. [DOI] [PubMed] [Google Scholar]
  • 5.Davies M, Elwyn G. Advocating mandatory patient “autonomy” in healthcare: adverse reactions and side effects. Health Care Anal. 2008;16(4):315–28. 10.1007/s10728-007-0075-3. [DOI] [PubMed] [Google Scholar]
  • 6.Kelly PA, Haidet P. Physician overestimation of patient literacy: a potential source of health care disparities. Patient Educ Couns. 2007;66(1):119–22. 10.1016/j.pec.2006.10.007. [DOI] [PubMed] [Google Scholar]
  • 7.Keene Woods N, Ali U, Medina M, Reyes J, Chesser AK. Health Literacy, Health Outcomes and Equity: A Trend Analysis Based on a Population Survey. J Prim Care Community Health. 2023;14:21501319231156132. 10.1177/21501319231156132. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.Weiss B, Blanchard JS, McGee DL, Hart G, Warren B, Burgoon M, et al. Illiteracy among Medicaid recipients and its relationship to health care costs. J Health Care Poor Underserved. 1994;5(2):99–111. [DOI] [PubMed] [Google Scholar]
  • 9.Weiss BD. Removing Barriers to Better, Safer Care: Health Literacy and Patient Safety: Help Patients Understand: Manual for Clinicians (2nd ed). Chicago, IL. American Medical Association Foundation; 2007
  • 10.Daraz L, Morrow AS, Ponce OJ, Farah W, Katabi A, Majzoub A, et al. Readability of online health information: a meta-narrative systematic review. Am J Med Qual. 2018;33(5):487–92. [DOI] [PubMed] [Google Scholar]
  • 11.Badarudeen S, Sabharwal S. Readability of patient education materials from the American Academy of Orthopaedic Surgeons and Pediatric Orthopaedic Society of North America web sites. J Bone Joint Surg Am. 2008;90(1):199–204. 10.2106/JBJS.G.00347. [DOI] [PubMed] [Google Scholar]
  • 12.Albright J, de Guzman C, Acebo P, Paiva D, Faulkner M, Swanson J. Readability of patient education materials: implications for clinical practice. Appl Nurs Res. 1996;9(3):139–43. 10.1016/s0897-1897(96)80254-0. [DOI] [PubMed] [Google Scholar]
  • 13.Andalib S, Solomon SS, Picton BG, Spina AC, Scolaro JA, Nelson AM. Source Characteristics Influence AI-Enabled Orthopaedic Text Simplification: Recommendations for the Future. JB JS Open Access. 2025;10(1):e24.00007. Published 2025 Jan 8. 10.2106/JBJS.OA.24.00007. [DOI] [PMC free article] [PubMed]
  • 14.Caterson J, Ambler O, Cereceda-Monteoliva N, Horner M, Jones A, Poacher AT. Application of generative language models to orthopaedic practice. BMJ Open. 2024;14(3):e076484. Published 2024 Mar 14. 10.1136/bmjopen-2023-076484. [DOI] [PMC free article] [PubMed]
  • 15.Kirchner GJ, Kim RY, Weddle JB, Bible JE. Can Artificial Intelligence Improve the Readability of Patient Education Materials? Clin Orthop Relat Res. 2023;481(11):2260–7. 10.1097/CORR.0000000000002668. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16.Ryu JH, Yi PH. Readability of Spine-Related Patient Education Materials From Leading Orthopedic Academic Centers. Spine (Phila Pa 1976). 2016;41(9):E561-E565. 10.1097/BRS.0000000000001321. [DOI] [PubMed]
  • 17.Michel C, Dijanic C, Abdelmalek G, et al. Readability assessment of patient educational materials for pediatric spinal deformity from top academic orthopedic institutions. Spine Deform. 2022;10(6):1315–21. 10.1007/s43390-022-00545-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18.Doinn TÓ, Broderick JM, Abdelhalim MM, Quinlan JF. Readability of Patient Educational Materials in Hip and Knee Arthroplasty: Has a Decade Made a Difference? J Arthroplasty. 2020;35(11):3076–83. 10.1016/j.arth.2020.05.076. [DOI] [PubMed] [Google Scholar]
  • 19.Reaver CN, Pereira DE, Vail EV, et al. Evaluating the performance of artificial intelligence for improving readability of online English- and Spanish-language orthopaedic patient educational material: challenges in bridging the digital divide. J Bone Joint Surg Am. 2025. 10.2106/JBJS.24.01078. [DOI] [PubMed] [Google Scholar]
  • 20.Roster K, Kann RB, Farabi B, Gronbeck C, Brownstone N, Lipner SR. Readability and Health Literacy Scores for ChatGPT-Generated Dermatology Public Education Materials: Cross-Sectional Analysis of Sunscreen and Melanoma Questions. JMIR Dermatol. 2024;7:e50163. Published 2024 Mar 6. 10.2196/50163. [DOI] [PMC free article] [PubMed]
  • 21.Dihan QA, Brown AD, Chauhan MZ, et al. Leveraging large language models to improve patient education on dry eye disease. Eye (Lond). 2025;39(6):1115–22. 10.1038/s41433-024-03476-5. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22.Gomez-Cabello CA, Borna S, Pressman SM, Haider SA, Forte AJ. Large Language Models for Intraoperative Decision Support in Plastic Surgery: A Comparison between ChatGPT-4 and Gemini. Medicina. 2024;60(6):957. 10.3390/medicina60060957. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 23.Armitage RC. Implications of Large Language Models for Clinical Practice: Ethical Analysis Through the Principlism Framework. J Eval Clin Pract. 2025;31(1):e14250. 10.1111/jep.14250. [DOI] [PMC free article] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Data Availability Statement

No datasets were generated or analysed during the current study.


Articles from Journal of Orthopaedic Surgery and Research are provided here courtesy of BMC

RESOURCES