Skip to main content
Plastic and Reconstructive Surgery Global Open logoLink to Plastic and Reconstructive Surgery Global Open
. 2025 Jun 2;13(6):e6825. doi: 10.1097/GOX.0000000000006825

Ethical Considerations for Generative Artificial Intelligence in Plastic Surgery

Ravi Dhawan *, Kendall Douglas Brooks *, Orr Shauly *,, Denys Shay , Albert Losken *
PMCID: PMC12133144  PMID: 40469554

Summary:

The integration of artificial intelligence (AI) into surgical care is rapidly transforming healthcare by enhancing efficiency, clinical decision-making, and patient outcomes. Generative AI (genAI), a subfield using large language models such as ChatGPT, Bard, and Midjourney, holds significant promise in automating tasks such as surgical planning and discharge summaries. However, it raises concerns about misinformation, data breaches, biases, and misuse. No genAI technology has yet received Food and Drug Administration approval for surgical use, emphasizing the need for thorough regulatory evaluation. This article proposed 5 ethical principles, adapted from World Health Organization recommendations, to guide the adoption and governance of genAI in plastic surgery. These principles include ensuring data transparency, maintaining patient autonomy, prioritizing safety and accountability, promoting equity, and investing in sustainability. Each principle is illustrated with a hypothetical case to highlight potential ethical breaches and the importance of rigorous testing, clear communication, and continuous monitoring. By adhering to these guidelines, stakeholders can ensure that genAI serves to enhance patient care and uphold the highest standards of ethical practice in surgical settings.


Takeaways

Question: How can generative AI (genAI) be ethically integrated into plastic surgery to enhance patient care while minimizing risks such as misinformation, data breaches, biases, and misuse?

Findings: This study proposed 5 ethical principles adapted from World Health Organization guidelines—data transparency, patient autonomy, safety and accountability, equity, and sustainability—to guide the use of genAI in plastic surgery. Using hypothetical cases, it illustrates potential ethical challenges and offers a code of conduct for clinical practice.

Meaning: Adhering to these ethical principles ensures that genAI in plastic surgery improves patient outcomes while maintaining trust, safety, and equity in healthcare.

INTRODUCTION

Artificial intelligence (AI) is quickly redefining surgical care by using technologies that complement human intelligence to improve healthcare efficiency; clinical decision-making; and, most importantly, patient outcomes. Generative AI (genAI) is a subfield of AI that uses large language models (LLMs) to generate realistic images, text, and videos that can ultimately assist in automating writing tasks such as discharge summaries or surgical planning.1,2 The potential applications of LLMs such as ChatGPT, Bard, Llama, Gemini, and Midjourney in the field of surgery are significant, as they possess the ability to efficiently handle complex concepts and provide multimodal responses to a wide array of inquiries and prompts.3,4 Their initial appeal is also met with legitimate concerns regarding the propagation of misinformation, clinical efficacy, patient data breaches, racial biases originating from training data, and the potential for misuse.2 To date, no genAI technology has undergone review and approval by the US Food and Drug Administration (FDA) for any surgical domain, highlighting the need for thorough regulatory evaluation in healthcare. In this article, we propose 5 ethical principles, adapted from World Health Organization (WHO) recommendations for AI integration into healthcare, to guide effective genAI adoption and governance in plastic surgery5,6 (Table 1). The WHO guidelines provide ethical principles to ensure the safe, effective, and equitable integration of AI into healthcare, addressing challenges such as bias, privacy, transparency, and accountability.5,6 Although these guidelines address ethical challenges such as bias, privacy, and misuse in AI broadly, they do not account for the unique implications of genAI, such as direct messaging with patients using chatbots or undermining patient autonomy during patient consultations.

Table 1.

Ethical Principles and Guidance for GenAI Adoption in Plastic Surgery

Ethical Clause Definition Implementation Strategies Regulatory Challenges Case Examples
Data transparency and intelligibility Ensure that the data used by AI are clear and accessible to users with clarity in the model’s decision-making process Clear labeling of clinical data sources with a commitment to regular audits and reports on performance and data integrity is essential. Independent expert review is used to validate existing algorithms with systems that allow for user feedback on AI decisions to improve model transparency
Compliance with data protection regulations and ensuring patient data anonymity An algorithm is developed to predict postoperative rhinoplasty simulations. Data to train the model include pre- and postoperative images of patients’ faces. These data are clearly labeled, and an independent plastic surgeon regularly audits the AI to validate its precision. Before using the AI, surgeons explain to patients how the system processes images and makes predictions. Patients are also allowed to provide feedback on the AI’s suggestions. All personal data are anonymized and managed in compliance with privacy regulations to ensure patient confidentiality
Patient autonomy Maintaining the patient’s right to informed decision-making in care involving AI Implement AI tools that supplement rather than replace surgeon consultations, and allow patients to opt in, rather than opt out, of AI-guided support to garner patient trust Upholding informed consent regulations. During a preoperative consultation, a patient is offered the use of a surgical planning AI tool, with the option to accept or refuse. The surgeon then presents an AI-guided recommendation for a reconstructive procedure, clearly outlining alternative options and the patient’s right to decline
Safety and efficacy Prioritizing the safety and effectiveness of AI tools for surgical application Rigorous preimplementation testing fulfilling regulatory requirements, coupled with regular updates and audits of AI systems Adherence to medical device and technology certification standards A breast augmentation clinic uses an AI system to predict surgical outcomes. This AI undergoes extensive preimplementation testing, receives regulatory approval, and is regularly audited for accuracy and safety. It adheres to medical device/technology certification standards. Surgeons use predictions to discuss potential outcomes with patients, enhancing the consultation process
Accountability and responsibility Assigning clear accountability for AI-driven decisions in clinical settings Define legal responsibilities in the use of AI and develop protocols for error reporting and resolution Legal implications of AI errors An AI system that assists in evaluating candidates for liposuction mistakenly identifies a patient as an ideal candidate despite contraindications. The clinic has clear protocols in place: the error is immediately reported, the AI’s decision is reviewed by a team of surgeons, and the patient is informed and offered alternative assessments. Legal responsibilities are predefined, ensuring the clinic remains compliant, and both the clinic and technology company are accountable for the AI’s actions
Equity and inclusivity Ensuring AI tools serve all demographic groups without reinforcing existing healthcare disparities Design AI to handle diverse datasets reflecting different demographics while monitoring AI outputs for bias Challenges in creating unbiased AI systems A nonoperative tool recommends postoperative medications for patients who have undergone reconstructive surgery. Each recommendation is followed by prior authorization support from a large dataset of diverse health insurance plans to ensure coverage. The tool submits adequate medical justifications directly to insurers, streamlining approvals and ensuring patients receive necessary care promptly and equitably
Sustainability and adaptability Focusing on the tool’s long-term viability and flexibility to adapt to evolving medical knowledge Regular updates to incorporate the latest surgical research with scalable solutions that are adaptable to various procedures Managing continuous learning of AI systems without compromising current clinical standards Preoperative breast reconstruction tools integrate the latest clinical research, data from clinical trials, and real-world evidence into their recommendations. Frequent updates ensure the system remains current with advancements in surgical techniques. Surgeons contribute to a feedback loop allowing the tool to refine its recommendations based on surgical outcomes maintaining high clinical standards and adaptability

Our framework builds on the WHO’s principles by creating specific guidelines tailored to plastic surgery, focusing on patient trust, safety, and informed consent. We provide a specialized code of conduct and use hypothetical scenarios to demonstrate how these adapted principles can guide ethical genAI adoption in clinical practice, offering a novel approach that aligns with, but extends beyond, existing WHO recommendations. We begin discussions of each ethical tenet by presenting a hypothetical case and outlining a code of conduct that plastic surgeons should follow when integrating these technologies into their practices.

Ensuring Data Transparency

Case 1

An aesthetic surgery clinic deployed a genAI tool to predict postoperative outcomes for rhinoplasty, intended to assist surgeons in clinical decision-making and to guide patient expectations.711 The AI tool generated visual simulations based on past procedures but lacked transparency about data sources, including patient demographics, surgical techniques, and outcomes. It provided no documentation on data provenance, and routine audits or validation were absent. Ethical concerns arose when AI predictions differed significantly from actual results, leading to patient dissatisfaction despite no clinical errors. The lack of transparency and systematic review hindered model improvement, and procedural risks were discussed without addressing the tool’s accuracy or reliability.

Many early genAI applications in healthcare exist in an ostensible black box, with opaque decisions and unintelligible logic. Discourse in clinical literature has called for explainable AI, where augmented clinical decision-making can be justified. Moreover, the EU General Data Protection Regulation has called for all technology companies, beyond healthcare, to be mindful of their practices and provide “meaningful information about the logic behind automated decisions using their data.”1214

The desire for explainability stems from a drive to ensure physician-patient trust and safety and inform surgeons of the steps behind AI decision-making.13 This explainability generally takes 2 forms: inherent explainability and post hoc explainability.13 Inherent explainability pertains to models with clear input–output data where relationships between independent and dependent variables can be quantified.1518 For example, the relationship between the independent and dependent variables can be teased out in a predictive linear regression that estimates the risk of postsurgical complications following breast reconstruction based on patient comorbidities.19,20

However, complex AI models capable of image or text analysis and production lack simple input–output relationships. In such cases, post hoc explainability is used, aiming to dissect the model’s decision-making process using a confluence of decision variables.13,18,21 In the scenario of an algorithm that predicts the postsurgical appearance following rhinoplasty, researchers can try to work backward to reverse-engineer the generative model’s output decision (Fig. 1).

Fig. 1.

Fig. 1.

Landscape of genAI in plastic surgery. This figure illustrates the integration of genAI within plastic surgery, from input data (clinical and surgical data, EHR, clinical endpoints) through to output data (treatment suggestions, diagnoses, postoperative images). Machine learning approaches (supervised, unsupervised, and reinforcement) lead to clustered and generative models, forming the core of genAI. The diagram highlights 3 application areas: academic and administrative support (scientific writing and generating patient notes with LLMs such as GPT-4), operative support (preoperative planning, intraoperative support with computer vision, and AI-triaging), and patient-oriented support (remote monitoring and virtual chatbots via digital health companies). DTx, digital therapeutics.

The additional goal of explainability is to detect biases; however, given that many sociological and demographic biases are encoded in AI training data that have been collected years prior, localized justifications regarding model decision-making will do little to mitigate global concerns about the model’s generalizability.13,2227 The integration of explainable AI should be complemented with commitments to collect more robust and diverse training data so that surgeons seeking genAI support are adequately informed, and disparities in plastic surgery care are not perpetuated.2832 Moreover, given the inherent complexity of many genAI technologies, there exists a trade-off between full explainability and accuracy: in achieving full explainability, researchers could inadvertently oversimplify the model, in turn, resulting in a loss of accuracy as the model’s capabilities are compromised in favor of transparency.15 Explainability is not always a prerequisite for efficacy, as evidenced by the case of acetaminophen: despite its extensive use, the precise mechanisms underlying its therapeutic effects remain unclear.33

Although explainability in healthcare AI is limited, all genAI models should undergo rigorous, plastic surgery–specific testing before clinical use. This should follow established guidelines, include simulated surgical scenarios, and use standard evaluation metrics typically assessed for other medical devices. Validation must assess the AI’s assumptions and health insurance portability and accountability act compliance. Independent review boards with multidisciplinary experts should audit these models for validity.

Maintaining Patient Autonomy

Case 2

During a preoperative breast reconstruction consultation, a patient receives an AI-assisted surgical recommendation without clear disclosure of the AI’s role or non-AI alternatives. The tool operates discreetly, refining its predictions with patient data. This opacity risks influencing surgical decisions misaligned with patient preferences. When patients later learn of the AI’s involvement and data use without explicit consent, trust and privacy concerns arise. This case underscores the need for transparency to ensure that AI supports clinical judgment while upholding patient autonomy and informed consent.

Informed consent and autonomy serve as central tenets of medical ethics.34,35 These principles require healthcare providers to collect informed, educated, and voluntary consent from patients before any surgical, research-related, or clinical procedure.3639 Integrating genAI into surgical consultations must prioritize patient autonomy, keeping human decision-making central. Surgeons should ensure informed consent and prevent overreliance on AI. Rather than classifying AI as simply autonomous or nonautonomous, a graded system would better assess the varying risks of automation in healthcare.40

Autonomous AI-based tools are capable of analyzing imaging data to provide diagnostic decisions,41 such as detecting large vessel occlusions in stroke patients and alerting specialists without the need for human radiologist interpretation.42,43 In ophthalmology, an autonomous AI system diagnoses diabetic retinopathy directly from retinal images without requiring skilled operators.44 Similarly, AI-powered systems for continuous glucose monitoring can independently adjust insulin delivery in real time based on glucose trends, minimizing the need for patient intervention and reducing the risk of hypo- or hyperglycemia.45

Popular LLMs such as ChatGPT and Bard are nonautonomous agents that are prompted to serve as assistants for surgeons. Preliminary reports show that ChatGPT is promising in oncology46 and head and neck surgical decision-making,47 and is equivalent to first-year residents on the plastic surgery in-service exam48 (Table 2). ChatGPT supported multidisciplinary tumor boards by recommending neoadjuvant therapy, aligning with treatment guidelines, when the tumor board had opted for surgery alone.46 Despite their initial appeal, ChatGPT has also been shown to produce inconsistent surgical advice54 and a diagnostic error rate of 83% in some pediatric cases.55 Especially in the nascent stages of genAI integration in surgery, where the reliability of these developing technologies is unclear, transparent communication about AI’s involvement is crucial for maintaining trust and adapting to the changing needs of both surgeons and patients.56,57 Decisions made by genAI systems should be reversible and designed to assist, not replace, surgeons.58 Surgeons should disclose the use of genAI for surgical decision-making or any other purpose to their patients and in medical notes as they would for any other clinical intervention.

Table 2.

Use Cases of GenAI and LLMs for Plastic and Aesthetic Surgeons and Patients

Use Case Primary Users Availability of Technology* Examples of Potential Applications in Plastic Surgery
Academic and administrative support Research support Academic plastic surgeons, researchers, and medical students 5 Automating data collection for systematic reviews and meta-analyses, assisting in identifying research gaps in the literature and writing academic reports, and summarizing findings from published articles
Generating, summarizing, translating, and interpreting of discharge notes Plastic surgeons, physician assistants, nurses, and patients 3 Efficiently condense postoperative care instructions, highlighting critical elements such as wound management and activity restrictions, and create standardized operative reports and patient progress notes to ensure comprehensive documentation in line with institution-specific protocols. Additionally, instantly translate postoperative instructions into multiple languages or plain language to improve patient comprehension and adherence across diverse populations
Prior authorization support Plastic surgeons, physician assistants, and nurses 2 Streamlines the compilation and submission of detailed justifications for procedures or medications to insurance companies, enhancing preauthorization efficiency
Advertising and marketing Surgeon-entrepreneurs, aesthetic surgery clinic owners, and investors 5 Leverage AI tools such as Midjourney and DALL-E to create high-quality, realistic images for marketing campaigns. These tools can showcase surgical outcomes, clinic environments, and patient testimonials, enhancing online presence and attracting clients. This approach streamlines content creation and ensures marketing materials are innovative, engaging, and visually appealing49,50
Operative support Generation of predictive images of postoperative results Plastic surgeons and patients 1 Generating photorealistic renderings of predicted postoperative results in cosmetic procedures such as rhinoplasty or reconstructive procedures such as breast reconstruction, which can be presented using AR/VR51,52
Visual assessment of postoperative aesthetic outcomes Plastic surgeons and patients 1 Objective assessment of postsurgical outcomes using predefined aesthetic criteria such as facial symmetry on an informed image-to-text model
Clinical decision support and diagnostic assistance Plastic surgeons, physician assistants, and nurses 2 Integration with institution-specific databases including longitudinal data on procedural success and complications to suggest procedural strategies
Triage of postoperative complications and emergencies Plastic surgeons, physician assistants, and nurses 1 Monitor signs of postoperative complications, such as sepsis, prioritizing cases for rapid response to prevent adverse outcomes19,20
Patient-oriented support Rehabilitation guidance Patients 2 Tailored rehabilitation plans are presented in accessible text for patients. Plans are based on recovery trajectories to optimize recovery in reconstructive patients
Postoperative remote patient monitoring Plastic surgeons, physician assistants, nurses, and patients 3 Analyze data from wearable technology and/or well-being chatbots to monitor recovery, allowing for early intervention through real-time alerts to healthcare providers
Medication reminders Patients 4 Implement automated reminders for postoperative medication schedules, ensuring compliance with pain management or antibiotic prophylaxis
Pre- and postoperative chatbot Patients 2 Provide 24/7 virtual support, delivering preoperative advice and postoperative care information to enhance patient engagement and satisfaction, and monitor well-being

The analysis presented was conducted based on the information and technological resources available up to and including May 1, 2024. It is important to acknowledge that the field of AI is rapidly evolving. Future advancements in technology may lead to modifications in the applications of these tools, potentially expanding their utility beyond the current uses outlined in this article. As such, the findings and applications discussed should be considered within the context of the state of technology at the time of analysis. Readers are encouraged to consider these dynamic aspects when applying or referencing this work in future research or clinical practice. This table was adapted from Meskó and Topol.71

*

Assessment of “availability of technology” was based on predefined criteria: Readily available (category 1): technologies fully operational and in active clinical use, such as AI-based imaging tools for preoperative planning and postoperative monitoring integrated into EHRs. Criteria: peer-reviewed clinical studies demonstrating widespread use, commercial products available to multiple healthcare facilities, and inclusion in standard guidelines. Limited availability (category 2): technologies available but limited by region, cost, or integration challenges, such as specialized imaging tools requiring specific equipment. Criteria: found in select clinical settings or regions, with some publications indicating limited adoption due to access barriers or costs. In development (category 3): technologies in active development, clinical trials, or reaching product-market fit, such as AI-driven preoperative optimization systems. Criteria: premarket studies, clinical trials, or investor-backed start-ups working toward product-market alignment. Conceptual research (category 4): Technologies proposed or undergoing initial research without clinical validation, including theoretical applications in facial reconstruction algorithms. Criteria: published in peer-reviewed journals, but with limited experimental validation or predominantly laboratory-based research studies. Speculative future (category 5): technologies still in the preresearch phase, not actively developed or tested, such as advanced generative models for automated robotic surgery. Criteria: speculative proposals in review papers, academic talks, or grant proposals without ongoing research or prototypes.

AR, augmented reality; VR, virtual reality.

Unlike traditional healthcare technologies, genAI evolves from new data, which can lead to the reuse of patient information in ways not originally consented to or expected.59 This evolving nature of AI complicates how informed consent is managed, as patients may not be fully aware of how their data could be used as AI capabilities expand.60,61 To uphold patient autonomy, healthcare providers must prioritize transparency and regular communication about AI’s role and ensure patients understand how their data might be (re)used over time.

Finally, respecting patient autonomy extends to safeguarding privacy, confidentiality, and informed consent for data usage. One recent study demonstrated the potential for users to circumvent security measures in ChatGPT and access confidential, personally identifiable information using a technique known as the Janus Interface, which involves a modification of the LLM.6264 The integration of personal identifiable information into chatbots poses privacy risks: healthcare professionals may unintentionally expose patient data, whereas malicious users may exploit weak cybersecurity. Despite safeguards, trained cybercriminals could circumvent security measures. Sharing surgical data for AI model refinement without explicit consent heightens reidentification risks, violating patient trust and informed consent principles.65 Existing government legal frameworks, which ensure patient control over their data, including access and sharing options6669 while a good start, must be continuously updated for the imminent and specific risks.

During preoperative consultations, it is crucial to maintain patient autonomy by providing clear information about AI use in surgical planning. Patients should be given AI-generated recommendations and alternative options, ensuring they understand their right to accept or decline AI tools. Surgeons must transparently discuss the AI’s benefits and limitations, emphasizing that the final decision is always the patients’.

Prioritizing Safety

Case 3

An outpatient clinic’s AI tool summarizes electronic health record (EHR) data to assess liposuction eligibility but overlooks critical comorbidities. Relying on this incomplete report, the surgeon proceeds without fully reviewing the medical chart, resulting in severe complications. Investigations reveal outdated AI guidelines and reactive error protocols. The patient sues for negligence and lack of informed consent regarding AI involvement.

The introduction of novel technologies in healthcare is often met with initial technological difficulties and complications.70 For example, the adoption of robotic-assisted tools for minimally invasive procedures, though seemingly infallible, has a nonzero number of adverse events71 and was initially met with considerable skepticism and underwent decades of review before approval.51,72,73

Both surgeons and healthcare administrators adopting genAI must prioritize safety, accountability, and responsibility to mitigate the initial risk of complications.74 This involves ensuring that genAI technologies align with regulatory standards for safety, accuracy, and efficacy, particularly for specific, well-defined surgical use cases.53,75 Before deployment, these technologies must undergo rigorous testing, similar to any medical device, to ensure patient safety. Unlike static medical tools, genAI models evolve over time, requiring continuous monitoring by developers, investors, and surgeons to maintain quality, implement improvements, and assess potential adverse effects.53 Preventing harm is crucial; genAI technologies, especially those that answer patient questions or provide diagnoses, must be carefully managed to avoid scenarios where patients are misinformed, or where these models perpetuate unrealistic surgical expectations.51,49

The FDA regulates all medical products, including AI-based technologies, under its broader framework for ensuring safety and quality.53,76,77 For AI and machine learning technologies, which fall under the category of software as a medical device, the FDA has developed a tailored regulatory strategy that focuses on continuous monitoring and evaluation of real-world performance.53,78 To date, 6 AI medical products, labeled by the FDA as plastic surgery devices, have been approved by the FDA.77 Despite these developments, no genAI device or LLM in any specialty has been approved, and challenges remain in regulating adaptive AI algorithms that evolve over time, which will require refinement of regulatory standards.

In initial studies, genAI interfaces have been shown to enhance informed consent documentation for surgery,79 generate accurate information to answer medical queries,80 and provide comprehensive responses to commonly asked plastic surgery questions.81 Although these initial use cases are exciting, they have been met with limitations including fluctuating clinical accuracy and impersonalized medical advice,8183 which can lead to dangerous consequences if used without oversight. Currently, there is little transparency regarding the quality and robustness of the surgical data that are fed into these preliminary models.

One consideration includes direct-to-patient AI communications, such as preoperative consultations or answering postoperative questions via chatbots, which offer exciting potential but also come with significant risks. Early studies suggest that AI-generated advice can save clinicians time, is often accurate and concise,84 and is sometimes even preferred over surgeon-drafted responses.85,86 However, these AI systems often lack the nuance and personalization a clinician provides, failing to consider unique patient needs, varying health literacy, or unexpected complications.50 This can result in patients misunderstanding instructions, following incorrect advice, or missing essential recovery steps, especially if they are unaware that their guidance is AI-driven rather than from a surgeon. To prevent potential harm, it is crucial to have strict oversight, ensuring that AI communications are accurate, clear, and supplemented by human supervision.86

Another well-documented philosophical consideration is AI hallucination,87 a dangerous symptom of chatbots, which involves the production of logical, well-defined, and plausible incorrect responses. LLMs such as ChatGPT sometimes produce counterfactual outputs that are articulated with confidence, which may result in inflated surgical trust in the technology leading to catastrophic consequences.53

It remains unclear who would be held liable for ill-informed genAI-augmented clinical decision-making. This is especially pertinent in the field of genAI, which uses unsupervised learning to inform their outputs; unlike other software, individual computational decisions are not coded by a team of software engineers.88,89 This may create a “responsibility gap,” placing the burden of genAI-related harm on surgeons or healthcare workers who did not develop the technology itself.2,8891 Holding surgeons solely responsible for genAI-related harm is impractical, as many LLMs are black-box models, and surgeons may not fully understand the logic behind their recommendations.89 Nonetheless, surgeons must be cautious in leveraging AI for surgical advice, recognizing the liability and ethical risks involved.2 Surgeons should not be fully exempt from AI-related errors, especially when genAI tools suggest care options rather than definitive diagnoses. AI should supplement, not replace, surgeon judgment, providing additional insights where beneficial.

Promoting Equity

Case 4

An academic hospital’s AI tool recommends postoperative medications for reconstructive surgery but overlooks social determinants of health and reflects biases from overrepresented demographic groups in its training data. As a result, underrepresented patients receive inaccessible prescriptions or face insurance delays, increasing recovery time and complication risks. In response, the hospital reassesses the tool to integrate pricing, social factors, and more inclusive data, aiming for equitable healthcare delivery.

Well-documented concerns within our healthcare system revolve around persisting health disparities, even when new healthcare technologies are introduced to enhance patient care. This situation gives rise to what can be described as a “technological divide,” where populational inequities in care are exacerbated by advancements in medical technology.92 Plastic surgery, already grappling with existing concerns regarding diversity and inclusion, must make considerations regarding the price and availability of genAI technologies such that its integration advances and does not hinder health equity.93,94

Of particular concern when using LLMs is the perpetuation of bias, which can impact clinical decision-making and surgical equity.53,95 Healthcare AI bias can arise from training data that underrepresent diverse and vulnerable groups, leading to models that perpetuate social stereotypes. Some studies have shown that text-to-generative AI models tend to underrepresent sex and ethnic minority groups when generating images of physicians96 and surgeons.97 Other studies have documented GPT-4 in exhibiting racial and gender bias in medical recommendations, patient assessments, and medical education.94,98 GPT-4 exaggerated disease prevalence disparities and amplified societal stereotypes, highlighting serious concerns about using genAI for large-scale surgical adoption.98,99

Ensuring equity in surgical genAI requires commitment throughout the so-called algorithm life cycle, starting with diverse and representative data collection.99 Development, trials, and validation should incorporate input from varied backgrounds to enhance transparency in intended use, population limits, and generalizability. AI developers must proactively address biases to prevent widening health disparities.

AI Sustainability

Case 5

A surgery center uses a preoperative AI tool for breast reconstruction. This tool is programmed to integrate the latest clinical research data from clinical trials and real-world evidence into its recommendations. Regular updates are applied to the system to ensure it remains aligned with the evolving nature of surgical techniques. Surgeons actively contribute to the model’s feedback loop, ensuring that the AI tool’s continuous learning process allows the model to adapt its recommendations in real time with contemporary surgical practice (Table 3).

Table 3.

Examples of Ethical Considerations Associated With Current GenAI Technologies

Technological Category Example of Existing Technology in Healthcare Summary of Ethical Concerns Examples of Harmful Consequences
Academic research support GPT-4 Ethical concerns include data privacy violations and bias in training datasets. GenAI models are also prone to hallucination in which they convincingly articulate a misguided or inaccurate fact. Research support tools may also inadvertently plagiarize from existing literature available in the model’s training dataset GenAI models mining patient data from EHRs to support postoperative outcomes research may inadvertently expose sensitive information or perpetuate biases if the training data lack diversity
Administrative task support Abridge healthcare Ethical issues include data privacy, accuracy in AI-generated documentation, and overreliance on automated systems. Inaccurate translations of clinical notes or prior authorization requests may result in inappropriate care or unnecessary insurance denials AI-driven clinical note-generation tools produce incomplete or incorrect prior authorization documents, leading to unjustified denials and delayed access to critical treatments
Preoperative support AEDIT Ethical concerns include reinforcing societal biases in beauty standards and creating unrealistic patient expectations. GenAI predicting postoperative outcomes may standardize results based on prevalent aesthetic norms, disadvantaging patients from diverse ethnic backgrounds. Additionally, AI tools analyzing before-and-after photographs risk compromising patient confidentiality if images are not properly anonymized or used without informed consent AI tools suggesting standardized rhinoplasty outcomes could fail to reflect a patient’s cultural or personal preferences, leading to dissatisfaction and the perpetuation of narrow beauty ideals
Clinical decision support Glass Health Clinical decision support, diagnostic assistance, and triage systems can present risks of diagnostic errors and overreliance on AI recommendations. Surgeons may overlook their clinical judgment, increasing the potential for patient harm if AI misclassifies conditions AI systems misidentifying a postoperative infection as benign could delay urgent treatment, compromising patient outcomes and increasing healthcare costs due to prolonged hospital stays or further complications
Patient-facing virtual diagnostic support Corti Virtual chatbots pose ethical challenges related to patient privacy, data security, and the potential dissemination of misinformation. Patients may disclose sensitive health information to chatbots without understanding how their data will be used or stored. Chatbots might provide incorrect medical advice or hallucinate due to limited context or inadequate training Preoperative chatbots that offer erroneous preparation guidelines can compromise patient safety and increase the risk of surgical complications
Patient-facing postoperative support EveryDose Postoperative support technologies such as medication reminders or rehabilitation guidance must ensure accurate and personalized recommendations. Errors or inadequate personalization can lead to suboptimal recovery and adverse drug events Rehabilitation chatbots providing generic exercise instructions without considering a patient’s specific surgical history may result in injury or prolonged recovery, highlighting the need for individualized, evidence-based guidance in postoperative care

Integrating genAI technologies into surgery necessitates a careful balance between enhancing efficiency and preventing the aggravation of technology burdens for surgeons. Although AI holds great promise in optimizing healthcare practices, poor integration can lead to unintended complications. In addition, the inherent nature of surgical practice necessitates a very close relationship between surgeons and AI training models to ensure proper interoperability between the operating room and the algorithms.

It is also important to highlight the importance of ensuring that genAI-driven healthcare technologies, including AI-integrated EHR systems, are designed with user-friendliness in mind to prevent additional workload and stress on healthcare providers. Incorporating genAI into surgery should prioritize usability and seamless integration, with minimal yet comprehensive tutorials and onboarding for effective large-scale adoption. Unlike EHRs, where hospitals invest heavily in technician support, initial genAI applications should provide scalable support to ensure consistency. Surgeons should collaborate with start-ups to provide iterative feedback, ensuring effective and sustainable technology uptake.

CONCLUSIONS

The integration of genAI into surgery enhances patient care and decision-making but must follow ethical principles for responsible implementation. This article outlined 5 key principles: data transparency, patient autonomy and safety, accountability, equity, and adaptability.

These principles safeguard patient safety, uphold human agency, mitigate bias, and ensure equitable AI access. As AI evolves, collaboration among providers, policymakers, developers, and patients is crucial to addressing ethical challenges. Adhering to these principles ensures that AI complements human intelligence, advancing surgical practice and healthcare delivery.

DISCLOSURE

The authors have no financial interest to declare in relation to the content of this article.

ACKNOWLEDGMENTS

The authors would like to disclose the use of genAI in the preparation of this article. The OpenAI language model, specifically ChatGPT-4.0, was used to assist in language editing, spelling corrections, and refinement of phrasing. The usage of this AI tool was strictly limited to enhancing the linguistic clarity and readability of the content that was already conceptually and substantively developed by the human authors. At no point was the AI used to generate original ideas, conduct research, or author any part of the article independently. The contributions of ChatGPT were thoroughly reviewed and, where appropriate, revised by the article’s authors to ensure they accurately reflected the intended meaning and maintained the originality of the research.

Footnotes

Published online 2 June 2025.

Disclosure statements are at the end of this article, following the correspondence information.

REFERENCES

  • 1.Wachter RM, Brynjolfsson E. Will generative artificial intelligence deliver on its promise in health care? JAMA. 2024;331:65–69. [DOI] [PubMed] [Google Scholar]
  • 2.Duffourc M, Gerke S. Generative AI in health care and liability risks for physicians and safety concerns for patients. JAMA. 2023;330:313–314. [DOI] [PubMed] [Google Scholar]
  • 3.Abdelhady AM, Davis CR. Plastic surgery and artificial intelligence: how ChatGPT improved operation note accuracy, time, and education. Mayo Clin Proc Digit Health. 2023;1:299–308. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4.Puladi B, Gsaxner C, Kleesiek J, et al. The impact and opportunities of large language models like ChatGPT in oral and maxillofacial surgery: a narrative review. Int J Oral Maxillofac Surg. 2024;53:78–88. [DOI] [PubMed] [Google Scholar]
  • 5.World Health Organization. Ethics and governance of artificial intelligence for health. Available at https://www.who.int/publications/i/item/9789240029200. Accessed July 16, 2024. [Google Scholar]
  • 6.World Health Organization. Ethics and governance of artificial intelligence for health: guidance on large multi-modal models. Available at https://www.who.int/publications/i/item/9789240084759. Accessed July 16, 2024. [Google Scholar]
  • 7.Rodler S, Ganjavi C, De Backer P, et al. Generative artificial intelligence in surgery. Surgery. 2024;175:1496–1502. [DOI] [PubMed] [Google Scholar]
  • 8.Dhawan R, Shay D. Artificial intelligence in plastic surgery: implications and limitations of text-to-image models for clinical practice. JPRAS Open. 2024;41:368–371. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9.Koljonen V. What could we make of AI in plastic surgery education. J Plast Reconstr Aesthet Surg. 2023;81:94–96. [DOI] [PubMed] [Google Scholar]
  • 10.Zhang T, Fu H, Zhao Y, et al. SkrGAN: sketching-rendering unconditional generative adversarial networks for medical image synthesis. In: Shen D, Liu T, Peters TM, et al. , eds. Medical Image Computing and Computer Assisted Intervention—MICCAI 2019. Springer International Publishing; 2019:777–785. [Google Scholar]
  • 11.Yoo TK, Choi JY, Kim HK. A generative adversarial network approach to predicting postoperative appearance after orbital decompression surgery for thyroid eye disease. Comput Biol Med. 2020;118:103628. [DOI] [PubMed] [Google Scholar]
  • 12.Oakden-Rayner L. Closing the Implementation Gap in Pre-deployment Medical AI Study Design [Thesis]. The University of Adelaide. 2021. Available at https://digital.library.adelaide.edu.au/dspace/handle/2440/136684. Accessed July 16, 2024. [Google Scholar]
  • 13.Ghassemi M, Oakden-Rayner L, Beam AL. The false hope of current approaches to explainable artificial intelligence in health care. Lancet Digit Health. 2021;3:e745–e750. [DOI] [PubMed] [Google Scholar]
  • 14.European Union. Regulation - 2016/679 - EN - gdpr - EUR-Lex. Available at https://eur-lex.europa.eu/eli/reg/2016/679/oj. Accessed July 16, 2024. [Google Scholar]
  • 15.Amann J, Blasimme A, Vayena E, et al. ; Precise4Q Consortium. Explainability for artificial intelligence in healthcare: a multidisciplinary perspective. BMC Med Inform Decis Mak. 2020;20:310. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16.Holzinger A, Langs G, Denk H, et al. Causability and explainability of artificial intelligence in medicine. Wiley Interdiscip Rev Data Min Knowl Discov. 2019;9:e1312. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17.Linardatos P, Papastefanopoulos V, Kotsiantis S. Explainable AI: a review of machine learning interpretability methods. Entropy (Basel). 2020;23:18. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18.Reddy S. Explainability and artificial intelligence in medicine. Lancet Digit Health. 2022;4:e214–e215. [DOI] [PubMed] [Google Scholar]
  • 19.Kamran F, Tjandra D, Heiler A, et al. Evaluation of sepsis prediction models before onset of treatment. NEJM AI. 2024;1:AIoa2300032. [Google Scholar]
  • 20.Wong A, Otles E, Donnelly JP, et al. External validation of a widely implemented proprietary sepsis prediction model in hospitalized patients. JAMA Intern Med. 2021;181:1065–1070. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21.Jin W, Li X, Fatehi M, et al. Generating post-hoc explanation from deep neural networks for multi-modal medical image analysis tasks. MethodsX. 2023;10:102009. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22.Gastounioti A, Kontos D. Is it time to get rid of black boxes and cultivate trust in AI? Radiol Artif Intell. 2020;2:e200088. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 23.Daneshjou R, Smith MP, Sun MD, et al. Lack of transparency and potential bias in artificial intelligence data sets and algorithms: a scoping review. JAMA Dermatol. 2021;157:1362–1369. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 24.Agarwal R, Bjarnadottir M, Rhue L, et al. Addressing algorithmic bias and the perpetuation of health inequities: an AI bias aware framework. Health Policy Technol. 2023;12:100702. [Google Scholar]
  • 25.Challen R, Denny J, Pitt M, et al. Artificial intelligence, bias and clinical safety. BMJ Qual Saf. 2019;28:231–237. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 26.DeCamp M, Lindvall C. Latent bias and the implementation of artificial intelligence in medicine. J Am Med Inform Assoc. 2020;27:2020–2023. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27.Parikh RB, Teeple S, Navathe AS. Addressing bias in artificial intelligence in health care. JAMA. 2019;322:2377–2378. [DOI] [PubMed] [Google Scholar]
  • 28.Baxter NB, Howard JC, Chung KC. A systematic review of health disparities research in plastic surgery. Plast Reconstr Surg. 2021;147:529–537. [DOI] [PubMed] [Google Scholar]
  • 29.Khetpal S, Lopez J, Redett RJ, et al. Health equity and healthcare disparities in plastic surgery: what we can do. J Plast Reconstr Aesthet Surg. 2021;74:3251–3259. [DOI] [PubMed] [Google Scholar]
  • 30.Smith BT, Egro FM, Murphy CP, et al. An evaluation of race disparities in academic plastic surgery. Plast Reconstr Surg. 2020;145:268–277. [DOI] [PubMed] [Google Scholar]
  • 31.Wang A, Kim E, Oleru O, et al. Artificial intelligence in plastic surgery: ChatGPT as a tool to address disparities in health literacy. Plast Reconstr Surg. 2024;153:1232e–1234e. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 32.Liang X, Yang X, Yin S, et al. Artificial intelligence in plastic surgery: applications and challenges. Aesthetic Plast Surg. 2021;45:784–790. [DOI] [PubMed] [Google Scholar]
  • 33.Toussaint K, Yang XC, Zielinski MA, et al. What do we (not) know about how paracetamol (acetaminophen) works? J Clin Pharm Ther. 2010;35:617–638. [DOI] [PubMed] [Google Scholar]
  • 34.Jonsen AR. The Birth of Bioethics. Oxford University Press; 2003. [Google Scholar]
  • 35.Faden RR, Beauchamp TL. A History and Theory of Informed Consent. Oxford University Press; 1986. [PubMed] [Google Scholar]
  • 36.Kapp MB. Patient autonomy in the age of consumer-driven health care: informed consent and informed choice. J Health Biomed Law. 2006;2:1–32. [DOI] [PubMed] [Google Scholar]
  • 37.O’Sullivan S, Nevejans N, Allen C, et al. Legal, regulatory, and ethical frameworks for development of standards in artificial intelligence (AI) and autonomous robotic surgery. Int J Med Robot. 2019;15:e1968. [DOI] [PubMed] [Google Scholar]
  • 38.Cohen IG. Informed consent and medical artificial intelligence: what to tell the patient? Symposium: law and the nation’s health. Georgetown Law J. 2019;108:1425–1470. [Google Scholar]
  • 39.Astromskė K, Peičius E, Astromskis P. Ethical and legal challenges of informed consent applying artificial intelligence in medical diagnostic consultations. AI Soc. 2021;36:509–520. [Google Scholar]
  • 40.Bitterman DS, Aerts HJWL, Mak RH. Approaching autonomy in medical artificial intelligence. Lancet Digit Health. 2020;2:e447–e449. [DOI] [PubMed] [Google Scholar]
  • 41.Ratner M. FDA backs clinician-free AI imaging diagnostic tools. Nat Biotechnol. 2018;36:673–674. [DOI] [PubMed] [Google Scholar]
  • 42.Karamchandani RR, Helms AM, Satyanarayana S, et al. Automated detection of intracranial large vessel occlusions using Viz.ai software: experience in a large, integrated stroke network. Brain Behav. 2023;13:e2808. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 43.Vitellas CA, Mannix NC, Nimjee SM, et al. Abstract 130: real world experience with Viz.AI automated large vessel occlusion detection. Stroke. 2022;53:A130–A130. [Google Scholar]
  • 44.Savoy M. IDx-DR for diabetic retinopathy screening. Am Fam Physician. 2020;101:307–308. [PubMed] [Google Scholar]
  • 45.Forlenza GP, Ekhlaspour L, Breton M, et al. Successful at-home use of the Tandem Control-IQ artificial pancreas system in young children during a randomized controlled trial. Diabetes Technol Ther. 2019;21:159–169. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 46.Vela Ulloa J, King Valenzuela S, Riquoir Altamirano C, et al. Artificial intelligence-based decision-making: can ChatGPT replace a multidisciplinary tumour board? Br J Surg. 2023;110:1543–1544. [DOI] [PubMed] [Google Scholar]
  • 47.Chiesa-Estomba CM, Lechien JR, Vaira LA, et al. Exploring the potential of Chat-GPT as a supportive tool for sialendoscopy clinical decision making and patient information support. Eur Arch Otorhinolaryngol. 2024;281:2081–2086. [DOI] [PubMed] [Google Scholar]
  • 48.Humar P, Asaad M, Bengur FB, et al. ChatGPT is equivalent to first-year plastic surgery residents: evaluation of ChatGPT on the plastic surgery in-service examination. Aesthet Surg J. 2023;43:NP1085–NP1089. [DOI] [PubMed] [Google Scholar]
  • 49.Dhawan R, Nair A, Shay D. Cross-sectoral strategies for managing AI-driven content in plastic surgery. J Plast Reconstr Aesthet Surg. 2024;90:21–22. [DOI] [PubMed] [Google Scholar]
  • 50.Abi-Rafeh J, Cattelan L, Xu HH, et al. Artificial intelligence–generated social media content creation and management strategies for plastic surgeons. Aesthet Surg J. 2024;44:769–778. [DOI] [PubMed] [Google Scholar]
  • 51.Dhawan R, Bikmal A, Shay D. From virtual to reality: Apple Vision Pro’s applications in plastic surgery. J Plast Reconstr Aesthet Surg. 2024;94:43–45. [DOI] [PubMed] [Google Scholar]
  • 52.Sayadi LR, Naides A, Eng M, et al. The new frontier: a review of augmented reality and virtual reality in plastic surgery. Aesthet Surg J. 2019;39:1007–1016. [DOI] [PubMed] [Google Scholar]
  • 53.Meskó B, Topol EJ. The imperative for regulatory oversight of large language models (or generative AI) in healthcare. NPJ Digit Med. 2023;6:1–6. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 54.Kuroiwa T, Sarcon A, Ibara T, et al. The potential of ChatGPT as a self-diagnostic tool in common orthopedic diseases: exploratory study. J Med Internet Res. 2023;25:e47621. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 55.Barile J, Margolis A, Cason G, et al. Diagnostic accuracy of a large language model in pediatric case studies. JAMA Pediatr. 2024;178:313–315. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 56.Lee P, Bubeck S, Petro J. Benefits, limits, and risks of GPT-4 as an AI chatbot for medicine. N Engl J Med. 2023;388:1233–1239. [DOI] [PubMed] [Google Scholar]
  • 57.Dhawan R, Brooks KD. Limitations of artificial intelligence in plastic surgery. Aesthet Surg J. 2023;44:NP323–NP324. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 58.Sezgin E. Artificial intelligence in healthcare: complementing, not replacing, doctors and healthcare providers. Digit Health. 2023;9:20552076231186520. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 59.Price WN, Cohen IG. Privacy in the age of medical big data. Nat Med. 2019;25:37–43. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 60.Grote T, Berens P. On the ethics of algorithmic decision-making in healthcare. J Med Ethics. 2020;46:205–211. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 61.Vayena E, Blasimme A, Cohen IG. Machine learning in medicine: addressing ethical challenges. PLoS Med. 2018;15:e1002689. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 62.Chen X, Tang S, Zhu R, et al. The Janus interface: how fine-tuning in large language models amplifies the privacy risks. arXiv. 2023;3:1285–1299. [Google Scholar]
  • 63.IEEE Xplore. From ChatGPT to ThreatGPT: impact of generative AI in cybersecurity and privacy. IEEE Journals & Magazine. Available at https://ieeexplore.ieee.org/abstract/document/10198233. Accessed July 16, 2024. [Google Scholar]
  • 64.Dhawan R, Nair A, Shay D. Generative artificial intelligence in surgery: balancing innovation with ethical challenges. J Plast Reconstr Aesthet Surg. 2024;90:47–48. [DOI] [PubMed] [Google Scholar]
  • 65.Reddy S, Allan S, Coghlan S, et al. A governance model for the application of AI in health care. J Am Med Inform Assoc. 2019;27:491–497. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 66.Harrer S. Attention is not all you need: the complicated case of ethically using large language models in healthcare and medicine. EBioMedicine. 2023;90:104512. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 67.Gerke S, Rezaeikhonakdar D. Privacy aspects of direct-to-consumer artificial intelligence/machine learning health apps. Intell Based Med. 2022;6:100061. [Google Scholar]
  • 68.Mohammad Amini M, Jesus M, Fanaei Sheikholeslami D, et al. Artificial intelligence ethics and challenges in healthcare applications: a comprehensive review in the context of the European GDPR mandate. Mach Learn Knowl Extr. 2023;5:1023–1035. [Google Scholar]
  • 69.Rodriguez JA, Clark CR, Bates DW. Digital health equity as a necessity in the 21st century cures act era. JAMA. 2020;323:2381–2382. [DOI] [PubMed] [Google Scholar]
  • 70.Sujan M, Scott P, Cresswell K. Digital health and patient safety: technology is not a magic wand. Health Informatics J. 2020;26:2295–2299. [DOI] [PubMed] [Google Scholar]
  • 71.Alemzadeh H, Raman J, Leveson N, et al. Adverse events in robotic surgery: a retrospective study of 14 years of FDA data. PLoS One. 2016;11:e0151470. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 72.Jara RD, Guerrón AD, Portenier D. Complications of robotic surgery. Surg Clin North Am. 2020;100:461–468. [DOI] [PubMed] [Google Scholar]
  • 73.Shah J, Vyas A, Vyas D. The history of robotics in surgical specialties. Am J Robot Surg. 2014;1:12–20. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 74.Oleck NC, Naga HI, Nichols DS, et al. Navigating the ethical landmines of ChatGPT: implications of intelligent chatbots in plastic surgery clinical practice. Plast Reconstr Surg Glob Open. 2023;11:e5290. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 75.Brookings. Generative AI in health care: opportunities, challenges, and policy. Available at https://www.brookings.edu/articles/generative-ai-in-health-care-opportunities-challenges-and-policy/. Accessed July 16, 2024. [Google Scholar]
  • 76.FDA. FDA roundup: January 16, 2024. FDA. January 16, 2024. Available at https://www.fda.gov/news-events/press-announcements/fda-roundup-january-16-2024. Accessed February 2, 2024. [Google Scholar]
  • 77.FDA. Artificial intelligence and machine learning (AI/ML)-enabled medical devices. FDA. Published August 7, 2024. Available at https://www.fda.gov/medical-devices/software-medical-device-samd/artificial-intelligence-and-machine-learning-aiml-enabled-medical-devices.Accessed August 10, 2024. [Google Scholar]
  • 78.Benjamens S, Dhunnoo P, Meskó B. The state of artificial intelligence-based FDA-approved medical devices and algorithms: an online database. NPJ Digit Med. 2020;3:1–8. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 79.Decker H, Trang K, Ramirez J, et al. Large language model−based chatbot vs surgeon-generated informed consent documentation for common procedures. JAMA Netw Open. 2023;6:e2336997. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 80.Goodman RS, Patrinely JR, Stone CA, Jr, et al. Accuracy and reliability of chatbot responses to physician questions. JAMA Netw Open. 2023;6:e2336483. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 81.Seth I, Cox A, Xie Y, et al. Evaluating chatbot efficacy for answering frequently asked questions in plastic surgery: a ChatGPT case study focused on breast augmentation. Aesthet Surg J. 2023;43:1126–1135. [DOI] [PubMed] [Google Scholar]
  • 82.Nov O, Singh N, Mann D. Putting ChatGPT’s medical advice to the (Turing) test: survey study. JMIR Med Educ. 2023;9:e46939. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 83.Mellia JA, Basta MN, Toyoda Y, et al. Natural language processing in surgery: a systematic review and meta-analysis. Ann Surg. 2021;273:900–908. [DOI] [PubMed] [Google Scholar]
  • 84.Copeland-Halperin LR, O’Brien L, Copeland M. Evaluation of artificial intelligence–generated responses to common plastic surgery questions. Plast Reconstr Surg Glob Open. 2023;11:e5226. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 85.Durairaj KK, Baker O, Bertossi D, et al. Artificial intelligence versus expert plastic surgeon: comparative study shows ChatGPT “wins” rhinoplasty consultations: should we be worried? Facial Plast Surg Aesthet Med. 2024;26:270–275. [DOI] [PubMed] [Google Scholar]
  • 86.Ayers JW, Poliak A, Dredze M, et al. Comparing physician and artificial intelligence chatbot responses to patient questions posted to a public social media forum. JAMA Intern Med. 2023;183:589–596. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 87.Huang L, Yu W, Ma W, et al. A survey on hallucination in large language models: principles, taxonomy, challenges, and open questions. arXiv. 2023;43:1–55. [Google Scholar]
  • 88.Naik N, Hameed BMZ, Shetty DK, et al. Legal and ethical consideration in artificial intelligence in healthcare: who takes responsibility? Front Surg. 2022;9:862322. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 89.Habli I, Lawton T, Porter Z. Artificial intelligence in health care: accountability and safety. Bull World Health Organ. 2020;98:251–256. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 90.Lang BH, Nyholm S, Blumenthal-Barby J. Responsibility gaps and black box healthcare AI: shared responsibilization as a solution. Digit Soc. 2023;2:52. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 91.Yeung K. A study of the implications of advanced digital technologies (including AI systems) for the concept of responsibility within a human rights framework. Published November 9, 2018. Available at https://papers.ssrn.com/abstract=3286027. Accessed July 16, 2024. [Google Scholar]
  • 92.Saeed SA, Masters RM. Disparities in health care and the digital divide. Curr Psychiatry Rep. 2021;23:61. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 93.Swanson E. Diversity, equity, and inclusion in plastic surgery: a systematic review. Ann Plast Surg. 2024;92:353–366. [DOI] [PubMed] [Google Scholar]
  • 94.Zack T, Lehman E, Suzgun M, et al. Assessing the potential of GPT-4 to perpetuate racial and gender biases in health care: a model evaluation study. Lancet Digit Health. 2024;6:e12–e22. [DOI] [PubMed] [Google Scholar]
  • 95.Omiye JA, Lester JC, Spichak S, et al. Large language models propagate race-based medicine. NPJ Digit Med. 2023;6:1–4. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 96.Lee SW, Morcos M, Lee DW, et al. Demographic representation of generative artificial intelligence images of physicians. JAMA Netw Open. 2024;7:e2425993. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 97.Ali R, Tang OY, Connolly ID, et al. Demographic representation in 3 leading artificial intelligence text-to-image generators. JAMA Surg. 2024;159:87–95. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 98.Hastings J. Preventing harm from non-conscious bias in medical generative AI. Lancet Digit Health. 2024;6:e2–e3. [DOI] [PubMed] [Google Scholar]
  • 99.Chin MH, Afsar-Manesh N, Bierman AS, et al. Guiding principles to address the impact of algorithm bias on racial and ethnic disparities in health and health care. JAMA Netw Open. 2023;6:e2345050. [DOI] [PMC free article] [PubMed] [Google Scholar]

Articles from Plastic and Reconstructive Surgery Global Open are provided here courtesy of Wolters Kluwer Health

RESOURCES