Abstract
Background:
Large language models (LLMs), an artificial intelligence tool, have become widely available and used by patients and members of the general public to answer health care questions.
Purpose:
We sought to understand whether currently available LLMs can recommend an appropriate total knee arthroplasty (TKA) surgeon when prompted.
Methods:
We used 3 widely available LLMs (ChatGPT, Gemini, and DeepSeek) to conduct searches for a TKA surgeon, with 1 of 4 home cities specified (Lynchburg, VA; Trumbull, CT; Chicago, IL; and New York, NY). The first prompt posed was (1) “I was told by my primary care provider that I need a knee replacement, I live in (city, state), who would you recommend that I see?” Once the list of surgeons was elicited, the LLM was prompted: (2) “Why did you recommend this (these) surgeon(s)?”; (3) “Where did you get your information about this (these) surgeon(s)?”; and (4) “How do I get in contact with this (these) surgeon(s)?” Recommendations were considered appropriate if the surgeon routinely performed TKA and was actively practicing in the area. Descriptive statistics and Fisher’s exact tests were used to summarize findings.
Results:
Across the 3 LLMs, 49 of the 74 (66%) recommendations were deemed appropriate, although this varied by model: Gemini (26/30, 87%), ChatGPT (14/19, 74%), and DeepSeek (9/25, 36%). Of the inappropriate responses, 6 of the surgeons were out of area, 13 were not performing TKA, and 6 were hallucinated names. When asked for rationales for the recommendations, LLMs most commonly cited hospital and practice Web sites and patient reviews, which tended to favor surgeons with longer local practice tenure. Of the 74 contact details provided, only 17 (23%) were accurate, with significant variation among models: ChatGPT (13/19, 79%), DeepSeek (2/25, 8%), and Gemini (2/30, 7%).
Conclusion:
While LLMs show potential in identifying TKA surgeons, the 3 LLMs we tested varied in their ability to validate surgeon expertise and provide reliable contact information. Further research may be necessary to elucidate the criteria by which LLMs recommend surgeons.
Keywords: artificial intelligence, total knee arthroplasty, patient participation, information seeking behavior, orthopedic surgeon
Introduction
Artificial intelligence (AI) is becoming part of health care, as large language models (LLMs) are increasingly used for information retrieval and task assistance.1-3 LLMs are being used to answer patients’ questions and provide guidance on treatment options.4-7 In orthopedics, LLMs are being considered for specialty-specific tasks, supporting both clinicians and patients.6-16 This may be particularly helpful for common, high-volume conditions such as knee osteoarthritis.
Each year in America, over 1 million total knee arthroplasties (TKAs) are performed. 17 Patients find their arthroplasty surgeon through a number of pathways, including Internet searches, provider referrals, and recommendations from family or friends.18,19 Patients who are searching for a TKA surgeon on their own may face challenges, especially when looking online.20,21 As there are numerous factors to consider when choosing a surgeon, LLMs may provide patients a new tool with which to identify providers and guide decision-making.22-24
However, the accuracy of TKA surgeon recommendations generated by LLMs remains largely unknown. Given the potential implications for both patient care and practice development, we aimed to evaluate the current capabilities of several LLMs in recommending appropriate orthopedic surgeons for TKA. Specifically, we assessed whether or not these tools could (1) identify actively practicing surgeons who routinely perform knee replacements, (2) explain the basis for their recommendations, and (3) provide correct contact information.
Methods
Three LLMs—ChatGPT (Open AI), Gemini (Google/Alphabet), and DeepSeek (Deep Seek AI)—were selected for use in this study. They were chosen due to their popularity and user-friendly interfaces. We used these models to conduct online searches for providers in 2 medium-size cities (Lynchburg, VA, and Trumbull, CT) and 2 large-size cities (Chicago, IL, and New York, NY). These cities were chosen due to our knowledge of orthopedic surgeons practicing in these areas. To emulate the typical patient experience, all queries were performed utilizing the free version of the software. Each model was queried with the same set of standardized prompts, presented verbatim. The first prompt posed was (1) “I was told by my primary care provider that I need a knee replacement, I live in (city, state), who would you recommend that I see?” We recorded the number of additional queries necessary to obtain an appropriate list of surgeons from each LLM. Once the list of surgeons was elicited, the LLMs were prompted, (2) “Why did you recommend this (these) surgeon(s)?”; (3) “Where did you get your information about this (these) surgeon(s)?”; and (4) “How do I get in contact with this (these) surgeon(s)?”
To ensure accuracy, surgeon names were validated using knowledge of practicing colleagues, supplemented by Google searches and provider Web sites. Recommendations were classified as appropriate if the surgeon routinely performed TKA and was still practicing in the area. Recommendations were classified as inappropriate if the surgeon no longer practiced in the area, did not routinely do TKAs, was not an orthopedic surgeon, or was a hallucination. Surgeon contact information was validated by calling the number provided and inquiring about setting up an appointment with the surgeon in question. Offices that confirmed the surgeon were classified as appropriate.
Statistical Analysis
Data were summarized with descriptive tables, and statistical analyses were performed using R (version 4.5.0). Fisher’s exact tests were used to compare the accuracy of provider name and phone number recommendations across the 3 LLMs. Additional Fisher’s exact tests were conducted to assess differences in performance between the 4 geographic locations. A P < .05 was considered statistically significant.
Results
Across the 3 LLMs, the percentage of appropriate recommendations varied: Gemini, 87% (26/30); ChatGPT, 74% (14/19); and DeepSeek, 36% (9/25) (Table 1). Of the inappropriate responses, 6 of the surgeons were out of area, 13 were not performing TKA, and 6 were hallucinated names. Inaccuracies in Gemini and ChatGPT responses were most commonly due to surgeons who were no longer located in a specified city or those who did not routinely perform TKAs. DeepSeek exhibited a high rate of hallucinations and performed significantly worse than Gemini (P < .001) and ChatGPT (P < .05) (Figure 1).
Table 1.
Accuracy of LLM-Generated Provider Names Across Models.
| Appropriate providers | ||
|---|---|---|
| Model | Accuracy | 95% CI a |
| ChatGPT | 14/19* | (51.2%-88.2%) |
| Gemini | 26/30** | (70.3%-94.7%) |
| DeepSeek | 9/25 | (20.2%-55.5%) |
Abbreviation: LLMs, large language models.
Compared with Deepseek.
95% Confidence intervals calculated using the Wilson method.
P < .05: calculated using the Fisher exact test. **P < .01: calculated using the Fisher exact test.
Figure 1.
Example of LLM-generated surgeon recommendations for Lynchburg, VA, categorized as inappropriate. Several listed providers and contact details were hallucinated, meaning they do not exist.
LLMs: large language models.
Accuracy of contact information varied across LLMs (Table 2). ChatGPT provided the most accurate contact information (13/19, 68%), which was significantly more accurate than that of both DeepSeek (2/25, 8%; P < .0001) and Gemini (2/30, 7%; P < .0001). Appropriate provider recommendations were similar between Lynchburg, VA (7/16, 44%); Trumbull, CT (9/20, 45%); and Chicago, IL (9/14, 64%). All LLM recommendations in New York, NY, were appropriate (24/24), significantly outperforming all other cities (P < .0001).
Table 2.
Accuracy of LLM-Generated Contact Information Across Models.
| Accurate phone numbers | ||
|---|---|---|
| Model | Accuracy | 95% CI a |
| ChatGPT | 13/19 | (46.0%-84.6%) |
| Gemini | 2/30* | (1.9%-21.3%) |
| DeepSeek | 2/25* | (2.2%-25.0%) |
Abbreviation: LLMs, large language models.
Compared with ChatGPT.
95% Confidence intervals calculated using the Wilson method.
P < .001: calculated using Fisher exact test.
When questioned why the surgeons were chosen, answers varied. The 3 most common factors included specialization in joint replacements, high volume of knee replacements (50+ was a requirement for DeepSeek), and patient satisfaction scores.
Recommendations from the 4 cities favored surgeons with the longest tenure in the area. When asked to provide sources for their recommendations, LLMs most commonly cited hospital-affiliated Web sites, practice-specific Web sites, and patient review sites. Frequently mentioned patient review sources included Healthgrades, Google Reviews, Nextdoor, local news outlets, and Castle Connolly Top Doctors. The Web site of the American Board of Orthopaedic Surgery (ABOS) was also frequently cited. The LLMs were often inaccurate in identifying orthopedic surgeons who subspecialize in arthroplasty.
Discussion
This study found that the 3 popular LLMs tested varied in providing appropriate recommendations for TKA surgeons. A majority of inappropriate recommendations were for surgeons who did not perform TKA. Others recommendations included hallucinations or TKA surgeons who practiced in a different city than the search specified.
This study was limited by a relatively small number of queries across 3 popular LLMs and by its focus on only 4 cities. Findings may not be generalizable to other regions or reflective of future LLM performance as models continue to evolve. Although standardized prompts were used to ensure consistency, LLM responses can vary with changes in wording. Future studies should evaluate how specific provider attributes—such as fellowship training, procedural volume, or online presence—are weighted in LLM-generated recommendations. Given the rapid pace of LLM development, these models are likely to evolve quickly, and ongoing evaluation will be necessary to monitor improvements, shifts in behavior, and emerging clinical implications.
When prompted for the reasoning behind their recommendations, all LLMs were able to provide sources. These sources varied, including hospital-affiliated Web sites, practice-specific pages, patient review platforms (eg, Healthgrades, Google Reviews, Nextdoor), and the ABOS Web site. Recommendations tended to emphasize provider tenure and online presence over subspecialty training and practice volume. This pattern may reflect the use of older data sources used in building LLMs.25,26 Many Web sites and other publicly available information contain historical data. These findings underscore the need for further research into LLM reasoning, to clarify how surgeons can adapt their digital presence accordingly. Regardless, LLMs are poised to shift the landscape of how patients and caregivers search for health care information. 27
The extent to which geographic region influences LLM recommendations remains unclear. In this study, 100% of LLM-generated recommendations in New York City were appropriate. This performance was significantly better than that of the LLM recommendations in Lynchburg, Trumbull, and Chicago. The reasons for this discrepancy are likely multifactorial, potentially reflecting differences in online data availability, provider density, and digital infrastructure. For surgeons practicing in cities like New York, where LLMs reliably identify appropriate providers, optimizing the quality of information on commonly cited sources may help further improve visibility. In contrast, for surgeons in regions where LLMs struggle to generate accurate recommendations, efforts may first need to focus on establishing a clear and accessible online presence.
The accuracy of contact information remains a limitation. Among the 3 models, ChatGPT performed the best, providing accurate contact details in the majority of cases. DeepSeek and Gemini both demonstrated poor performance, with accuracy rates below 20%. Although some models offered general suggestions on locating contact information, none reliably provided precise contact information directly. Future iterations of LLMs may evolve to handle this gap more seamlessly. Indeed, consumer-focused reports, such as a Wall Street Journal demonstration of using an LLM to purchase flowers, illustrate the potential trajectory of these tools toward task completion rather than simple information delivery. 28
In conclusion, while LLMs show potential in helping patients identify orthopedic surgeons, our findings suggest LLMs fall short in validating surgeon expertise and providing reliable contact information. Future research to understand the criteria with which an LLM recommendation is generated may be necessary for surgeons to effect change to their digital presence. Increasing online presence through robust practice Web sites, excellent patient reviews, and local news stories seem to be promising ways to increase the likelihood of a surgeon being reliably recommended by an LLM.
Supplemental Material
Supplemental material, sj-docx-1-hss-10.1177_15563316251412853 for Can Artificial Intelligence Models Appropriately Recommend Knee Arthroplasty Surgeons? by Colin M. Emrich, BS, Ethan C. Gazan, MS, Alexander J. Baur, Alexandra S. Gabrielli, Jenna A. Bernstein and David C. Landy in HSS Journal®
Supplemental material, sj-docx-2-hss-10.1177_15563316251412853 for Can Artificial Intelligence Models Appropriately Recommend Knee Arthroplasty Surgeons? by Colin M. Emrich, BS, Ethan C. Gazan, MS, Alexander J. Baur, Alexandra S. Gabrielli, Jenna A. Bernstein and David C. Landy in HSS Journal®
Supplemental material, sj-docx-3-hss-10.1177_15563316251412853 for Can Artificial Intelligence Models Appropriately Recommend Knee Arthroplasty Surgeons? by Colin M. Emrich, BS, Ethan C. Gazan, MS, Alexander J. Baur, Alexandra S. Gabrielli, Jenna A. Bernstein and David C. Landy in HSS Journal®
Supplemental material, sj-docx-4-hss-10.1177_15563316251412853 for Can Artificial Intelligence Models Appropriately Recommend Knee Arthroplasty Surgeons? by Colin M. Emrich, BS, Ethan C. Gazan, MS, Alexander J. Baur, Alexandra S. Gabrielli, Jenna A. Bernstein and David C. Landy in HSS Journal®
Supplemental material, sj-docx-5-hss-10.1177_15563316251412853 for Can Artificial Intelligence Models Appropriately Recommend Knee Arthroplasty Surgeons? by Colin M. Emrich, BS, Ethan C. Gazan, MS, Alexander J. Baur, Alexandra S. Gabrielli, Jenna A. Bernstein and David C. Landy in HSS Journal®
Supplemental material, sj-docx-6-hss-10.1177_15563316251412853 for Can Artificial Intelligence Models Appropriately Recommend Knee Arthroplasty Surgeons? by Colin M. Emrich, BS, Ethan C. Gazan, MS, Alexander J. Baur, Alexandra S. Gabrielli, Jenna A. Bernstein and David C. Landy in HSS Journal®
Footnotes
ORCID iD: David C. Landy
https://orcid.org/0000-0002-4024-7679
Ethical Considerations: All procedures followed were in accordance with the ethical standards of the responsible committee on human experimentation (institutional and national) and with the Helsinki declaration.
Funding: The authors received no financial support for the research, authorship, and/or publication of this article.
The authors declared the following potential conflicts of interest with respect to the research, authorship, and/or publication of this article: Jenna A. Bernstein, MD, reports relationships with Johnson and Johnson Med Tech, Smith and Nephew, Connecticut Orthopaedic Society, AAHKS, AAOS. David C. Landy, MD, PhD, is a member of the Editorial Board of HSS Journal. The author did not take part in the peer review or decision-making process for this submission. Additional potential conflicts of interest include Department of Defense, AOFAS, AAHKS, AJSM, National Institutes of Health.
Required Author Forms: Disclosure forms provided by the authors are available with the online version of this article as supplemental material.
References
- 1. Clark M, Bailey S. Chatbots in Health Care: Connecting Patients to Information: Emerging Health Technologies. Canadian Agency for Drugs and Technologies in Health; 2024. [PubMed] [Google Scholar]
- 2. Van Bulck L, Moons P. What if your patient switches from Dr. Google to Dr. ChatGPT? A vignette-based survey of the trustworthiness, value, and danger of ChatGPT-generated responses to health questions. Eur J Cardiovasc Nurs. 2024;23(1):95-98. 10.1093/eurjcn/zvad038 [DOI] [PubMed] [Google Scholar]
- 3. Wyatt JM, Booth GJ, Goldman AH. Natural language processing and its use in orthopaedic research. Curr Rev Musculoskelet Med. 2021;14(6):392-396. 10.1007/s12178-021-09734-3 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 4. Brameier DT, Alnasser AA, Carnino JM, Bhashyam AR, Von Keudell AG, Weaver MJ. Artificial intelligence in orthopaedic surgery: can a large language model “write” a believable orthopaedic journal article? J Bone Joint Surg Am. 2023;105(17):1388-1392. 10.2106/JBJS.23.00473 [DOI] [PubMed] [Google Scholar]
- 5. Li LT, Adelstein JM, Sinkler MA, Mistovich RJ. Artificial intelligence promotes the dunning Kruger effect: evaluating ChatGPT answers to frequently asked questions about adolescent idiopathic scoliosis. J Am Acad Orthop Surg. 2025;33(9):473-480. 10.5435/JAAOS-D-24-00297 [DOI] [PubMed] [Google Scholar]
- 6. Shah R, Bozic KJ, Jayakumar P. Artificial intelligence in value-based health care. HSS J. 2025;21(3):307-313. 10.1177/15563316251340074 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 7. Sparks CA, Fasulo SM, Windsor JT, et al. ChatGPT is moderately accurate in providing a general overview of orthopaedic conditions. JB JS Open Access. 2024;9(2):e23.00129. 10.2106/JBJS.OA.23.00129 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8. Farhadi F, Barnes MR, Sugito HR, Sin JM, Henderson ER, Levy JJ. Applications of artificial intelligence in orthopaedic surgery. Front Med Technol. 2022;4:995526. 10.3389/fmedt.2022.995526 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 9. Fayed AM, Mansur NSB, De Carvalho KA, Behrens A, D’Hooghe P, De Cesar Netto C. Artificial intelligence and ChatGPT in Orthopaedics and sports medicine. J Exp Orthop. 2023;10(1):74. 10.1186/s40634-023-00642-8 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 10. Federer SJ, Jones GG. Artificial intelligence in orthopaedics: a scoping review. PLoS One. 2021;16(11):e0260471. 10.1371/journal.pone.0260471 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 11. Haleem A, Vaishya R, Javaid M, Khan IH. Artificial intelligence (AI) applications in orthopaedics: an innovative technology to embrace. J Clin Orthop Trauma. 2020;11:S80-S81. 10.1016/j.jcot.2019.06.012 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 12. Kiwinda LV, Kocher SD, Bryniarski AR, Pean CA. Bioethical considerations of deploying artificial intelligence in clinical orthopedic settings: a narrative review. HSS J. 2025;21(3):274-282. 10.1177/15563316251340303 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 13. Koucheki R, Lex JR, Brock M, Goel DP. Integrating artificial intelligence and virtual reality in orthopedic surgery: a comprehensive review. HSS J. 2025;21(3):289-298. 10.1177/15563316251345479 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14. Kunze KN. Generative artificial intelligence and musculoskeletal health care. HSS J. 2025;21(3):248-256. 10.1177/15563316251335334 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15. Oettl FC, Zsidai B, Oeding JF, Samuelsson K. Artificial intelligence and musculoskeletal surgical applications. HSS J. 2025;21(3):267-273. 10.1177/15563316251339596 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 16. Pawelczyk J, Kraus M, Voigtlaender S, Siebenlist S, Rupp MC. Advancing musculoskeletal care using AI and digital health applications: a review of commercial solutions. HSS J. 2025;21(3):331-341. 10.1177/15563316251341321 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 17. Sloan M, Premkumar A, Sheth NP. Projected volume of primary total joint arthroplasty in the U.S., 2014 to 2030. J Bone Joint Surg Am. 2018;100(17):1455-1460. 10.2106/JBJS.17.01617 [DOI] [PubMed] [Google Scholar]
- 18. Fabrizio GM, Cardillo C, Egol A, Rozell JC, Schwarzkopf R, Aggarwal VK. Factors influencing patient selection of orthopaedic surgeons for total hip (THA) and total knee arthroplasty (TKA). Arch Orthop Trauma Surg. 2024;144(5):2057-2066. 10.1007/s00402-024-05314-5 [DOI] [PubMed] [Google Scholar]
- 19. Tu HT, Lauer JR. Word of mouth and physician referrals still drive health care provider choice. Res Brief. 2008;(9):1-8. [PubMed] [Google Scholar]
- 20. Bozic KJ, Kaufman D, Chan VC, Caminiti S, Lewis C. Factors that influence provider selection for elective total joint arthroplasty. Clin Orthop Relat Res. 2013;471(6):1865-1872. 10.1007/s11999-012-2640-9 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 21. Londhe SB, Shah RV, Agrawal PO, Toor RJ, Londhe SS, Parkhe M. The influence of the Internet on the patients’ choice of surgeon for their total knee replacement surgery. J Clin Orthop Trauma. 2021;17:186-190. 10.1016/j.jcot.2021.03.010 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 22. Manning BT, Bohl DD, Saltzman BM, et al. Factors influencing patient selection of an orthopaedic sports medicine physician. Orthop J Sports Med. 2017;5(8):2325967117724415. 10.1177/2325967117724415 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 23. Moser A, Korstjens I, Van Der Weijden T, Tange H. Patient’s decision making in selecting a hospital for elective orthopaedic surgery. Eval Clin Pract. 2010;16(6):1262-1268. 10.1111/j.1365-2753.2009.01311.x [DOI] [PubMed] [Google Scholar]
- 24. Pressman SM, Borna S, Gomez-Cabello CA, Haider SA, Haider CR, Forte AJ. Clinical and surgical applications of large language models: a systematic review. J Clin Med. 2024;13(11):3041. 10.3390/jcm13113041 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 25. Alber DA, Yang Z, Alyakin A, et al. Medical large language models are vulnerable to data-poisoning attacks. Nat Med. 2025;31(2):618-626. 10.1038/s41591-024-03445-1 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 26. Thirunavukarasu AJ, Ting DSJ, Elangovan K, Gutierrez L, Tan TF, Ting DSW. Large language models in medicine. Nat Med. 2023;29(8):1930-1940. 10.1038/s41591-023-02448-8 [DOI] [PubMed] [Google Scholar]
- 27. Liu L, Meng J, Yang Y. LLM technologies and information search. J Econ Technol. 2024;2:269-277. 10.1016/j.ject.2024.08.007 [DOI] [Google Scholar]
- 28. Stern J. ChatGPT vs. Claude vs. DeepSeek: the battle to be my AI work assistant. Wall Street Journal. 2025. https://www.wsj.com/tech/personal-tech/chatgpt-claude-deepseek-ai-features-compared-c5e1483c?gaa_at=eafs&gaa_n=AWEtsqeZ0YNvtdnyNlV2Donc-5M5PR9kSRpb0QzXmMXIjNXBd9vCYqnYsVVb4C–VMk%3D&gaa_ts=6955878e&gaa_sig=tBWi_vbWQxqHz4XLHM03U5LoN9tQZK08edwrU5TRb1_3WIhUFCCd2HqwWH7_5Q9_umCftysDA0KUapXM6thKhw%3D%3D
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Supplementary Materials
Supplemental material, sj-docx-1-hss-10.1177_15563316251412853 for Can Artificial Intelligence Models Appropriately Recommend Knee Arthroplasty Surgeons? by Colin M. Emrich, BS, Ethan C. Gazan, MS, Alexander J. Baur, Alexandra S. Gabrielli, Jenna A. Bernstein and David C. Landy in HSS Journal®
Supplemental material, sj-docx-2-hss-10.1177_15563316251412853 for Can Artificial Intelligence Models Appropriately Recommend Knee Arthroplasty Surgeons? by Colin M. Emrich, BS, Ethan C. Gazan, MS, Alexander J. Baur, Alexandra S. Gabrielli, Jenna A. Bernstein and David C. Landy in HSS Journal®
Supplemental material, sj-docx-3-hss-10.1177_15563316251412853 for Can Artificial Intelligence Models Appropriately Recommend Knee Arthroplasty Surgeons? by Colin M. Emrich, BS, Ethan C. Gazan, MS, Alexander J. Baur, Alexandra S. Gabrielli, Jenna A. Bernstein and David C. Landy in HSS Journal®
Supplemental material, sj-docx-4-hss-10.1177_15563316251412853 for Can Artificial Intelligence Models Appropriately Recommend Knee Arthroplasty Surgeons? by Colin M. Emrich, BS, Ethan C. Gazan, MS, Alexander J. Baur, Alexandra S. Gabrielli, Jenna A. Bernstein and David C. Landy in HSS Journal®
Supplemental material, sj-docx-5-hss-10.1177_15563316251412853 for Can Artificial Intelligence Models Appropriately Recommend Knee Arthroplasty Surgeons? by Colin M. Emrich, BS, Ethan C. Gazan, MS, Alexander J. Baur, Alexandra S. Gabrielli, Jenna A. Bernstein and David C. Landy in HSS Journal®
Supplemental material, sj-docx-6-hss-10.1177_15563316251412853 for Can Artificial Intelligence Models Appropriately Recommend Knee Arthroplasty Surgeons? by Colin M. Emrich, BS, Ethan C. Gazan, MS, Alexander J. Baur, Alexandra S. Gabrielli, Jenna A. Bernstein and David C. Landy in HSS Journal®

