Abstract
OpenAI’s ChatGPT (Generative Pre-trained Transformer) is a chatbot that answers questions and performs writing tasks in a conversational tone. Within months of release, multiple sectors are contemplating the varied applications of this chatbot, including medicine, education, and research, all of which are involved in medical communication and scientific publishing. Medical writers and academics use several artificial intelligence (AI) tools and software for research, literature survey, data analyses, referencing, and writing. There are benefits of using different AI tools in medical writing. However, using chatbots for medical communications pose some major concerns such as potential inaccuracies, data bias, security, and ethical issues. Perceived incorrect notions also limit their use. Moreover, ChatGPT can also be challenging if used incorrectly and for irrelevant tasks. If used appropriately, ChatGPT will not only upgrade the knowledge of the medical writer but also save time and energy that could be directed toward more creative and analytical areas requiring expert skill sets. This review introduces chatbots, outlines the progress in ChatGPT research, elaborates the potential uses of ChatGPT in medical communications along with its challenges and limitations, and proposes future research perspectives. It aims to provide guidance for doctors, researchers, and medical writers on the uses of ChatGPT in medical communications.
Keywords: Artificial intelligence, chatbot, Chat Generative Pretrained Transformer, medical writing
INTRODUCTION
ChatGPT (Generative Pre-trained Transformer) is a chatbot that simulates natural human conversation – a natural language processing (NLP) model based on the GPT-3.5 architecture.[1] NLP enables machines to understand and generate human language.[2] OpenAI released ChatGPT in November 2022.[1]
Chatbots have tapped multiple sectors such as education, research, medicine, business development, coding, and arts transforming the way we communicate, illustrate, and create content.[3,4] They have potential uses in the health-care sector too, particularly in medical communications. They are already used in this sector to provide patient education, support, engagement, and personalized health care.[5,6] However, medical communication encompasses writing beyond patient education and support.[7]
Medical communications cover a large scope of writing including medical journalism, medical education, scientific communication, regulatory writing, and market access writing or health economics and outcomes research.[7] Medical writers search literature for relevant information; create medically accurate content based on this information; illustrate content in the form of infographics, figures, and images; analyze statistics; and finally, present complex medical content in different formats to various audiences.[8] The use of artificial intelligence (AI)-powered tools such as Grammarly and Elicit to perform some of these tasks is extremely common. These tools increase efficiency while saving time.[9] Will ChatGPT or any other chatbot become yet another tool in their armamentarium?
This review focuses on ChatGPT and its potential applications from the perspective of medical writing, examining the current literature evidence of ChatGPT in medical writing as well as enlisting the challenges and limitations of its use. The recently updated guidelines regarding chatbot usage in scientific publications and the future directions on the utility of ChatGPT in medical writing are also discussed herein.
THE ADVENT OF CHATBOTS
The term “chatbot” refers to a computer program designed to mimic human conversation; it facilitates human–computer interaction.[3]
ELIZA, the first “chatbot” developed in 1966 by Prof. Weizenbaum from the Massachusetts Institute of Technology, was a computer program that could simulate conversation with a human operator.[10] From ELIZA to ChatGPT and further, the evolution of chatbots has accelerated over the years with many new chatbots being introduced in a short span of time.[4] Now, we have chatbots with domain-specific knowledge such as BioGPT and PharmaGPT (currently under development) for biomedical and pharmaceutical text mining and generation.[11,12]
Today, chatbots are used for a variety of purposes and their potential applications are increasing.[4] An advanced chatbot, ChatGPT, presents a potential use in medical writing, such as ideation, content drafting, summaries, and many more.
CHATGPT
ChatGPT is a sophisticated chatbot. It is capable of declining an improper request and accepting its error.[1] ChatGPT can generate text in a wide range of styles and tones, answer questions, and translate text between languages.[13] Another feature contributing toward its superiority over its predecessors is its 175 billion parameters [Figure 1].[1]
Figure 1.
ChatGPT versions.[1,13,14,15,16,17] GPT = Generative Pre-trained Transformer
The number of parameters is analogous to the number of synapses in a human brain. In language models, these parameters are essentially the internal representations that the model uses to make predictions or generate text. The more parameters a model has, the more complex relationships and patterns it can learn from the training data, which can improve its performance.[1,13] The model is trained on a massive amount of text data. It also uses different learning approaches.[1] In addition, OpenAI continues to work on developing new versions of ChatGPT with even larger numbers of parameters, as well as versions that are more specialized for specific types of NLP tasks [Figure 1].[1,13,14,15,16,17,18]
ChatGPT usage is free. The main reason for this is to allow people to use and experiment with the technology, and to help promote and advance the field of NLP.[1] Researchers may use the information gained from ChatGPT usage to improve language models and advance the field.
In addition, ChatGPT’s development is ongoing, and researchers may use user feedback to help improve the model’s performance.[18]
This section of the review was written with the aid of ChatGPT. ChatGPT explained its development and functionality in a simple language. The authors developed this content further and rewrote this section.
POTENTIAL USES OF CHATGPT IN MEDICAL WRITING
The literature on ChatGPT usage in medical writing mostly revolves around academic manuscript writing. Several tasks are involved in writing a manuscript that are essential for clarity and brevity of the content but are time-consuming. ChatGPT can aid in several of these essential elements to provide a starting point on which the author can build a publication-worthy draft. It can assist in generating a zero (preliminary) draft of an article, editing, formatting, creating summaries, and preparing tables, or other content that requires sorting and organizing data [Table 1]. The results provided by ChatGPT may not always meet the desired quality or standards, but they can be economical on the writer’s time and energy.
Table 1.
Potential uses of ChatGPT in medical writing published in literature
Potential use as described in publications | Interpretation | References |
---|---|---|
Can make linguistically coherent text from scattered points | Like arranging a jigsaw puzzle | [19] |
Reference and citation sorting and management | Not for searching references | [19,20] |
Saves time in generating preliminary drafts | Gives the first rough draft within seconds | [21] |
Provides fresh perspective on a research topic | Relevant literature survey can then be done based on the content | [21] |
Can write conference abstracts | Based on data input by the authors | [22] |
Provides initial draft of scientific paper | Provided a starting point to the authors to write the manuscript | [23] |
Suggests title for scientific paper | It can be unoriginal but will be scientifically accurate. ChatGPT suggested the title for the present review | [23,24] |
Writes methods section from raw information | Provides detailed methods from shorter input | [23] |
Formatting and language editing | Very efficient at copyediting | [20,23,25] |
Rewriting complex sentences or paraphrasing | Can provide several options of writing style to select from | [20,23] |
Summarizing entire text for abstract | May miss on important information if the prompt is not detailed and specific | [20,23] |
Automatic generation of tables | ChatGPT tabulated the content provided by the authors for the present review | [23] |
Collaboration in academic writing | Provides assistance for writing | [24,26] |
Suggests journals based on criteria provided | Authors do not report any instance of misinformation | [27] |
Assists in brainstorming research hypothesis | Particularly helpful for entry-level researchers | [27] |
Can write grant proposals in provided format | Needs further modification but provides a starting point | [28] |
Can create case studies for student education | In our experience, ChatGPT may not provide satisfactory results for complex case reports | [29] |
Writing patient-facing educational brochures | Uses appropriate language level and tone | [30] |
Language translation | Translation capabilities need further improvement | [31] |
Writing case reports | Several case reports were written with ChatGPT, with varying degrees of success | [32,33,34,35,36,37,38,39,40] |
Potential for generating effective queries for systematic review literature surveys | This paper is not peer-reviewed | [41] |
Development of topics for systematic reviews | 65% accuracy in suggesting novel systematic review ideas | [42] |
ChatGPT=Chat Generative Pre-trained Transformer
ChatGPT works best when used for summarizing content from the data provided. It can also be used to understand complex topics. However, this strictly applies to upgrading one’s knowledge.
The responses of ChatGPT depend on prompts – the questions input by the user. It is a chatbot, and the interaction has to be in a conversational tone. Thus, prompt engineering (writing a detailed and accurate prompt) is an important facet of ChatGPT usage. A clear, detailed, and unambiguous prompt will generate better responses than short, unclear prompts. The prompt engineering methods are not released by OpenAI but are proposed as guides by writers who have used and assessed ChatGPT extensively, albeit there is no research published on prompt engineering yet.[43]
Although ChatGPT’s capability has caused quite an uproar globally, its potential uses and limitations in the medical communications sector have not been assessed objectively and comprehensively. While several reports have explored its potential as a writing tool, any claims must be evaluated before drawing conclusions.
LIMITATIONS OF CHATGPT
The capabilities of ChatGPT have the potential for both positive and negative impacts. ChatGPT is an assistant that necessitates human supervision. Inadequate prompts from a user will generate inaccurate outcomes. Table 2 summarizes the limitations of ChatGPT and was auto-generated by ChatGPT from content provided by the authors.
Table 2.
Limitations of ChatGPT for medical writing
Limitation | Interpretation | References |
---|---|---|
Artificial hallucination | References are mostly wrong and can never be trusted | [19,20,21,22,24,44] |
Ethical consideration for using ChatGPT as an author | No accountability | [21] |
Risk of bias | Inherent bias in the training data | [20,21,30] |
Inadequate inputs lead to inadequate responses | The prompts have to be detailed and specific | [23] |
Lack of nuance, style, or originality | ChatGPT-generated text can be repetitive and unoriginal | [23,45] |
Meticulous human supervision is required | Content needs to be verified by a human expert | [46] |
Knowledge cutoff of September 2021 | No access to current information | [47,48] |
ChatGPT=Chat Generative Pre-trained Transformer
Every writer has a certain style of writing. ChatGPT is not a writer but a content generator. It is unoriginal; it lacks creativity, analytical skills, and the ability to infer complex medical content. All of these qualities and more are expected in a medical writer.[8] ChatGPT cannot be used for literature survey. The references have been shown to be fabricated most of the times.[19,20,21,22,24,44] Plausible-sounding inaccurate content is termed as “hallucination” that is one of the most concerning challenges of AI.[1] ChatGPT hallucinates information, and if prompted, it confesses its mistake of providing inaccurate or misleading responses. While an expert and experienced writer will recognize hallucinations instantly, a novice may take more time to corroborate the information, or worse, they may rely on this inaccurate information for writing.
Furthermore, ChatGPT cannot provide current information because it was trained on data prior to September 2021.[1] The information may be outdated.
Thus, it is necessary to maintain strict vigilance while using ChatGPT. Medical writers should be aware of the misuse and potential ethical implications of using ChatGPT.
CHATGPT: A TALE OF TWO TRAJECTORIES
Since its advent, ChatGPT has been tried by several medical writers. Many discovered its strengths and pitfalls by trial and error. While some are deterred by the initial hiccups, others continue using ChatGPT incorrectly. We provide here two case studies of ChatGPT usage from published literature. These examples demonstrate the continued incorrect use of ChatGPT despite contrary data and the proclivity of medical writers toward or against ChatGPT.
THE “HALLUCINATION” EFFECT
Salvagno et al. published an article 3 months after ChatGPT release in which they claimed that ChatGPT can help in literature review, identify research questions, provide an overview of the current state of the field, and assist with tasks such as formatting and language review.[23] The writers used a prompt to summarize a research article published in 2023. ChatGPT hallucinated and provided plausible-sounding unrealistic content because the knowledge cutoff of ChatGPT is September 2021. To this publication, Azamfirei et al. replied that the response generated by ChatGPT was inaccurate and the premise of the prompt was incorrect.[47]
However, 2 months later, Beutel et al.[49] replied to the original article. Here, they repeated the prompt given originally. This time, ChatGPT did not provide a response; it explained that it is trained on data up to 2021 and has no access to the 2023 paper. This was a correct response by ChatGPT! However, the premise of the author was incorrect again. This same mistake of asking ChatGPT to create content based on information from a later date by other authors also yielded no results.[50]
If ChatGPT is used for incorrect purposes, the results will be unsatisfactory and misleading. And authors may dismiss the usage of ChatGPT for academic writing.[32]
POTENTIAL BIAS AMONG MEDICAL WRITERS
Altmäe et al. used ChatGPT for writing a complete manuscript based on fictional data. According to their assessment, ChatGPT can provide methods for analysis and suggest important factors to be considered in the analysis, generate an acceptable abstract, suggest a title for the paper, and write a good methods section.[24] The introduction, results, and discussion sections needed a lot of improvement. However, the draft provided by ChatGPT helped guide the authors to improve the content. The references were completely wrong and unreliable. When the manuscript was checked for plagiarism, the similarity index was 19%, which is acceptable for most journals. Thus, according to these authors, ChatGPT has the potential to design a study, suggest analyses, and draft the paper.[24]
On the other hand, some authors outrightly rejected the use of ChatGPT for academic writing because of incorrect references, inaccurate information (hallucinations), potential harm to the article’s originality, and inability to write a complete error-free manuscript.[32,51]
If ChatGPT could write complete manuscripts, we would not need medical writers. ChatGPT cannot replace medical writers or the complete writing exercise. It can merely assist the writer to ease and improve their writing.
GUIDELINES ON CHATGPT IN SCIENTIFIC PUBLISHING
Several journals do not accept ChatGPT as an author.[30,52,53,54,55,56] Journal editors ask for a declaration in acknowledgments (or other sections) specifying the assistance of ChatGPT.[20,54,55,56,57] An international journal explicitly indicates that papers written using ChatGPT will have low priority for publication.[58]
On the other hand, Cureus journal had run a contest for case reports written with the assistance of ChatGPT.[59] This exercise may help the publishing industry evaluate the ability of ChatGPT to assist in manuscript writing.[32,33,34,35,36,37,38,39,40]
The World Association of Medical Editors (WAME) has clarified that “Chatbots cannot be authors.” In addition, they also recommend transparency in reporting the use of chatbots. Authors should take full responsibility for the content provided by a chatbot. WAME also insisted that editors should also have access to tools to detect AI-generated text.[60] GPTZero, GPT-2 Output Detector, AI Detector by Content at Scale, and many other freely available tools report the likelihood of portions of text being human generated or AI generated; however, these tools can give flawed results and be easily “fooled” by simple paraphrasing of AI-generated text.[61,62]
The International Committee of Medical Journal Editors (ICMJE) has released a statement regarding not listing chatbots as authors. They recommend describing the use of AI-assisted technology in the cover letter and the manuscript.[63] Several manuscripts have mentioned ChatGPT usage in the “Acknowledgments” section. As per ICMJE criteria, “acknowledgment may imply endorsement by acknowledged individuals of a study’s data and conclusions.”[63] ChatGPT is incapable of endorsing content and, therefore, cannot be included in acknowledgments.
We suggest that the “Methods” section would be ideal for the description of ChatGPT usage. It would be equivalent to mentioning statistics software or image creation tools.
Providing guidelines is a decision of journal editors and experts. However, practicing high ethical publishing standards is expected from everyone involved in medical writing. Moreover, this includes disclosing the use of ChatGPT, and other AI tools designed or adapted for writer’s assistance.
MEDICAL COMMUNICATIONS
Apart from scientific publications, medical writers also write plain language summaries, books, continued medical education slide decks, e-learning modules, patient brochures, promotional literature, blogs, advisory board meeting reports, regulatory documents, etc.[8] In one instance, ChatGPT was used to write patient-facing educational brochures.[30] Besides this, there are no instances of the use of ChatGPT for writing other medical communication documents in published literature. However, the principles learned from scientific writing can be applied for writing other documents too.
In medical marketing, market insights, competitor intelligence, and key opinion leaders’ perspectives direct the flow and content of promotional literature. ChatGPT cannot capture the nuances in writing required for creating content for advisory board reports, slide decks, and other medical marketing documents.
Medical writers should exercise utmost caution while relying on information obtained from ChatGPT, rather it would be prudent to use ChatGPT for writing assistance only. It may be used to garner fresh perspectives on a particular topic or as a sounding board to spark creative ideas.[21]
Orchestrating meetings and communicating with clients is a major part of a medical writer’s daily tasks. ChatGPT can help with rote tasks of drafting E-mails and minutes of meetings from pointers provided by the user. It can be a time-saving tool when used for these “nonmedical writing” tasks.
FUTURE DIRECTIONS
Advanced research in the field of large language models is expected to provide better chatbots with improved performance in the future. OpenAI introduced ChatGPT Plus, the paid version of ChatGPT with access to the GPT-4 model, in February 2023, which aims to provide a better user experience in terms of speed and quality of responses than ChatGPT.[14,15]
It is essential to evaluate the appropriate use of ChatGPT in medical communications. The papers in literature are either for or against ChatGPT. There are several reports that have encouraged medical writers to use ChatGPT but with caution. Currently, 85% of published papers concur that ChatGPT improves scientific writing and enhances research equity and versatility.[64] However, we need insights into ChatGPT use in medical writing for drawing conclusions. If we adapt and adopt ChatGPT for medical writing, a calculated and regulated approach would reduce the liability associated with its use.
Apart from writing prowess, medical writers need to have several other competencies including in-depth knowledge of a drug, and therapeutic area; ability to understand, analyze, and present complex data; understanding of ethics in clinical research and publishing; and awareness of tools used for efficient medical writing.[65] Interestingly, NLP tools were predicted to aid medical writers in authoring structured content long before the release of ChatGPT.[9] They can help medical writers save time spent on mundane, tedious, and time-consuming tasks, which will allow them to focus on areas requiring creativity and medical expertise.
CONCLUSIONS
AI and chatbots are rapidly evolving and will continue to evolve. ChatGPT can be used for saving time and writing assistance. Medical writers should bear in mind the limitations of using chatbots, and meticulously supervise and revise the content provided by ChatGPT. Moreover, they should take full responsibility for the written document.
ChatGPT is not a writer but a tool to aid the writer. Medical writers should conscientiously follow guidelines and recommendations laid down by journals and experts.
So, is ChatGPT a game-changer or a gimmick? We say it is neither!
Financial support and sponsorship
Nil.
Conflicts of interest
There are no conflicts of interest.
Acknowledgments
We acknowledge initial support provided by Dr. Neha Tanwar and creative support provided by D. Daniel Solomon Raj.
REFERENCES
- 1.OpenAI. Introducing ChatGPT. [[Last accessed on 2023 Jun 08]]. Available from: https://www.openai.com/blog/chatgpt .
- 2.Juluru K, Shih HH, Keshava Murthy KN, Elnajjar P. Bag-of-words technique in natural language processing: A primer for radiologists. Radiographics. 2021;41:1420–6. doi: 10.1148/rg.2021210025. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 3.Shawar BA, Atwell E. Chatbots: Are they really useful? J Lang Technol Comput Linguist. 2007;22:29–49. [Google Scholar]
- 4.Adamopoulou E, Moussiades L. Chatbots: History, technology, and applications. Mach Learn Appl. 2020;2:100006. [Google Scholar]
- 5.Wilson L, Marasoiu M. The development and use of chatbots in public health: Scoping review. JMIR Hum Factors. 2022;9:e35882. doi: 10.2196/35882. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 6.Tudor Car L, Dhinagaran DA, Kyaw BM, Kowatsch T, Joty S, Theng YL, et al. Conversational agents in health care: Scoping review and conceptual analysis. J Med Internet Res. 2020;22:e17158. doi: 10.2196/17158. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 7.Sharma S. Development of medical writing in India: Past, present and future. Perspect Clin Res. 2017;8:45–50. doi: 10.4103/2229-3485.198556. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8.Sharma S. How to become a competent medical writer? Perspect Clin Res. 2010;1:33–7. [PMC free article] [PubMed] [Google Scholar]
- 9.Parisis N. Medical writing in the era of artificial intelligence. Med Writ. 2019;28:4–9. [Google Scholar]
- 10.Weizenbaum J. ELIZA –A computer program for the study of natural language communication between man and machine. Commun ACM. 1966;9:36–45. [Google Scholar]
- 11.Luo R, Sun L, Xia Y, Qin T, Zhang S, Poon H, et al. BioGPT: Generative pre-trained transformer for biomedical text generation and mining. Brief Bioinform. 2022;23:bbac409. doi: 10.1093/bib/bbac409. [DOI] [PubMed] [Google Scholar]
- 12.PharmaGPT Ltd. PharmaGPT. Transforming Medicines Information. [[Last accessed on 2023 May 16]]. Available from: https://www.pharmadvisor.org.uk/pharmaGPT.html .
- 13.Brown TB, Mann B, Ryder N, Subbiah M, Kaplan J, Dhariwal P, et al. Language models are few-shot learners. arXiv. 2020 2005.14165. [Google Scholar]
- 14.OpenAI. GPT-4 Technical Report. arXiv. 2023 2303.08774. [Google Scholar]
- 15.OpenAI. Introducing ChatGPT Plus. [[Last accessed on 2023 Jun 09]]. Available from: https://www.openai.com/blog/chatgpt-plus .
- 16.Solaiman I, Brundage M, Clark J, Askell A, Herbert-Voss A, Wu J, et al. Release strategies and the social impacts of language models. arXiv. 2019 1908.09203. [Google Scholar]
- 17.Radford A, Narasimhan K, Salimans T, Sutskever I. Improving Language Understanding by Generative Pre-Training. [[Last accessed on 2023 May 17]]. Available from: https://www.cdn.openai.com/research-covers/language-unsupervised/language_understanding_paper.pdf .
- 18.OpenAI. Research Overview. [[Last accessed on 2023 May 16]]. Available from: https://www.openai.com/research/overview .
- 19.Alkaissi H, McFarlane SI. Artificial hallucinations in ChatGPT: Implications in scientific writing. Cureus. 2023;15:e35179. doi: 10.7759/cureus.35179. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 20.Koo M. The importance of proper use of ChatGPT in medical writing. Radiology. 2023;307:e230312. doi: 10.1148/radiol.230312. [DOI] [PubMed] [Google Scholar]
- 21.Sharma P. Chatbots in medical research: Advantages and limitations of artificial intelligence-enabled writing with a focus on ChatGPT as an author. Clin Nucl Med. 2023;48:838–9. doi: 10.1097/RLU.0000000000004665. [DOI] [PubMed] [Google Scholar]
- 22.Babl FE, Babl MP. Generative artificial intelligence: Can ChatGPT write a quality abstract? Emerg Med Australas. 2023;35:809–11. doi: 10.1111/1742-6723.14233. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 23.Salvagno M, Taccone FS, Gerli AG. Can artificial intelligence help for scientific writing? Crit Care. 2023;27:75. doi: 10.1186/s13054-023-04380-2. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 24.Altmäe S, Sola-Leyva A, Salumets A. Artificial intelligence in scientific writing: A friend or a foe? Reprod Biomed Online. 2023;47:3–9. doi: 10.1016/j.rbmo.2023.04.009. [DOI] [PubMed] [Google Scholar]
- 25.Kim SG. Using ChatGPT for language editing in scientific articles. Maxillofac Plast Reconstr Surg. 2023;45:13. doi: 10.1186/s40902-023-00381-x. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 26.Ang TL, Choolani M, See KC, Poh KK. The rise of artificial intelligence: Addressing the impact of large language models such as ChatGPT on scientific publications. Singapore Med J. 2023;64:219–21. doi: 10.4103/singaporemedj.SMJ-2023-055. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 27.Haman M, Školník M. Exploring the capabilities of ChatGPT in academic research recommendation. Resuscitation. 2023;187:109795. doi: 10.1016/j.resuscitation.2023.109795. [DOI] [PubMed] [Google Scholar]
- 28.Najafali D, Hinson C, Camacho JM, Galbraith LG, Gupta R, Reid CM. Can chatbots assist with grant writing in plastic surgery?Utilizing ChatGPT to start an R01 grant. Aesthet Surg J. 2023;43:P663–5. doi: 10.1093/asj/sjad116. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 29.Evans J. Working smarter using ChatGPT. Nurse Educ. 2023;10 doi: 10.1097/NNE.0000000000001424. 1097/NNE.0000000000001424. [DOI] [PubMed] [Google Scholar]
- 30.Solomon DH, Allen KD, Katz P, Sawalha AH, Yelin E, ChatGPT et al. …Artificial intelligence, authorship, and medical publishing. ACR Open Rheumatol. 2023;5:288–9. doi: 10.1002/acr2.11538. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 31.Chen TJ. ChatGPT and other artificial intelligence applications speed up scientific writing. J Chin Med Assoc. 2023;86:351–3. doi: 10.1097/JCMA.0000000000000900. [DOI] [PubMed] [Google Scholar]
- 32.Manohar N, Prasad SS. Use of ChatGPT in academic publishing: A rare case of seronegative systemic lupus erythematosus in a patient with HIV infection. Cureus. 2023;15:e34616. doi: 10.7759/cureus.34616. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 33.Jansz J, Manansala MJ, Sweiss NJ. Treatment of periorbital edema in a patient with systemic lupus erythematosus during pregnancy: A case report written with the assistance of ChatGPT. Cureus. 2023;15:e36302. doi: 10.7759/cureus.36302. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 34.Schussler JM, Tomson C, Dresselhouse MP. Extreme hyperthermia due to methamphetamine toxicity presenting as ST-elevation myocardial infarction on EKG: A case report written with ChatGPT assistance. Cureus. 2023;15:e36101. doi: 10.7759/cureus.36101. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 35.Schuppe K, Burke S, Cohoe B, Chang K, Lance RS, Mroch H. Atypical Nelson syndrome following right partial and left total nephrectomy with incidental bilateral total adrenalectomy of renal cell carcinoma: A chat generative pre-trained transformer (ChatGPT)-assisted case report and literature review. Cureus. 2023;15:e36042. doi: 10.7759/cureus.36042. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 36.Raxwal B, Baisla P, Nath J. A collaborative case report utilizing ChatGPT AI technology of traumatic right coronary artery dissection resulting in inferior wall ST-elevation myocardial infarction. Cureus. 2023;15:e35894. doi: 10.7759/cureus.35894. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 37.Hegde A, Srinivasan S, Menon G. Extraventricular neurocytoma of the posterior fossa: A case report written by ChatGPT. Cureus. 2023;15:e35850. doi: 10.7759/cureus.35850. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 38.Lantz R. Toxic epidermal necrolysis in a critically Ill African American woman: A case report written with ChatGPT assistance. Cureus. 2023;15:e35742. doi: 10.7759/cureus.35742. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 39.Nachshon A, Batzofin B, Beil M, van Heerden PV. When palliative care may be the only option in the management of severe burns: A case report written with the help of ChatGPT. Cureus. 2023;15:e35649. doi: 10.7759/cureus.35649. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 40.Akhter HM, Cooper JS. Acute pulmonary edema after hyperbaric oxygen treatment: A case report written with ChatGPT assistance. Cureus. 2023;15:e34752. doi: 10.7759/cureus.34752. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 41.Wang S, Scells H, Koopman B, Zuccon G. Can ChatGPT write a good Boolean query for systematic review literature search? arXiv. 2023 2302.03495. [Google Scholar]
- 42.Gupta R, Pande P, Herzog I, Weisberger J, Chao J, Chaiyasate K, et al. Application of ChatGPT in cosmetic plastic surgery: Ally or antagonist? Aesthet Surg J. 2023;43:P587–90. doi: 10.1093/asj/sjad042. [DOI] [PubMed] [Google Scholar]
- 43.Martin K, Llewellyn P. Enhancing Your Workflow with ChatGPT: Effective Prompt Engineering and Practical Applications. [[Last accessed on 2023 May 16]]. Available from: https://www.youtube.com/watch?v=thXimZxXuyc .
- 44.Sanchez-Ramos L, Lin L, Romero R. Beware of references when using ChatGPT as a source of information to write scientific articles. Am J Obstet Gynecol. 2023;229:356–7. doi: 10.1016/j.ajog.2023.04.004. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 45.Kaplan MH. I haven't been replaced by ChatGPT. Immunohorizons. 2023;7:286–7. doi: 10.4049/immunohorizons.2300024. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 46.Gilat R, Cole BJ. How will artificial intelligence affect scientific writing, reviewing and editing?The future is here …. Arthroscopy. 2023;39:1119–20. doi: 10.1016/j.arthro.2023.01.014. [DOI] [PubMed] [Google Scholar]
- 47.Azamfirei R, Kudchadkar SR, Fackler J. Large language models and the perils of their hallucinations. Crit Care. 2023;27:120. doi: 10.1186/s13054-023-04393-x. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 48.Goto A, Katanoda K. Should we acknowledge ChatGPT as an author? J Epidemiol. 2023;33:333–4. doi: 10.2188/jea.JE20230078. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 49.Beutel G, Geerits E, Kielstein JT. Artificial hallucination: GPT on LSD? Crit Care. 2023;27:148. doi: 10.1186/s13054-023-04425-6. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 50.Zheng H, Zhan H. ChatGPT in scientific writing: A cautionary tale. Am J Med. 2023;136:725–6.e6. doi: 10.1016/j.amjmed.2023.02.011. [DOI] [PubMed] [Google Scholar]
- 51.Wittmann J. Science fact versus science fiction: A ChatGPT immunological review experiment gone awry. Immunol Lett. 2023;256-257:42–7. doi: 10.1016/j.imlet.2023.04.002. [DOI] [PubMed] [Google Scholar]
- 52.Lee JY. Can an artificial intelligence chatbot be the author of a scholarly article? J Educ Eval Health Prof. 2023;20:6. doi: 10.3352/jeehp.2023.20.6. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 53.Fulton JS. Authorship and ChatGPT. Clin Nurse Spec. 2023;37:109–10. doi: 10.1097/NUR.0000000000000750. [DOI] [PubMed] [Google Scholar]
- 54.Thorp HH. ChatGPT is fun, but not an author. Science. 2023;379:313. doi: 10.1126/science.adg7879. [DOI] [PubMed] [Google Scholar]
- 55.Tools such as ChatGPT threaten transparent science;here are our ground rules for their use. Nature. 2023;613:612. doi: 10.1038/d41586-023-00191-1. [DOI] [PubMed] [Google Scholar]
- 56.Ide K, Hawke P, Nakayama T. Can ChatGPT Be considered an author of a medical article? J Epidemiol. 2023;33:381–2. doi: 10.2188/jea.JE20230030. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 57.Norris C. Large language models like ChatGPT in ABME: Author guidelines. Ann Biomed Eng. 2023;51:1121–2. doi: 10.1007/s10439-023-03212-2. [DOI] [PubMed] [Google Scholar]
- 58.Gurha P, Ishaq N, Marian AJ. ChatGPT and other artificial intelligence chatbots and biomedical writing. J Cardiovasc Aging. 2023;3:20. doi: 10.20517/jca.2023.13. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 59.Cureus. [[Last accessed on 2023 May 16]]. Available from: https://www.cureus.com/newsroom/news/164.%C2%A0Accessed%20January%2020 .
- 60.Zielinski C, Winker MA, Aggarwal R, Ferris LE, Heinemann M, Lapeña JF, et al. Chatbots, ChatGPT, and Scholarly Manuscripts: WAME Recommendations on ChatGPT and Chatbots in Relation to Scholarly Publications. WAME. [[Last accessed on 2023 May 09]]. Available from: https://www.wame.org/page3.php?id=106 . [DOI] [PubMed]
- 61.Májovský M, Černý M, Kasal M, Komarc M, Netuka D. Artificial intelligence can generate fraudulent but authentic-looking scientific medical articles: Pandora's box has been opened. J Med Internet Res. 2023;25:e46924. doi: 10.2196/46924. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 62.Anderson N, Belavy DL, Perle SM, Hendricks S, Hespanhol L, Verhagen E, et al. AI did not write this manuscript, or did it?Can we trick the AI text detector into generated texts?The potential future of ChatGPT and AI in sports and exercise medicine manuscript generation. BMJ Open Sport Exerc Med. 2023;9:e001568. doi: 10.1136/bmjsem-2023-001568. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 63.Defining the Role of Authors and Contributors. ICMJE. [[Last accessed on 2023 May 09]]. Available from: https://www.icmje.org/recommendations/browse/rolesand-responsibilities/defining-the-role-of-authors-and-contributors.html .
- 64.Sallam M. ChatGPT utility in healthcare education, research, and practice: Systematic review on the promising perspectives and valid concerns. Healthcare (Basel) 2023;11:887. doi: 10.3390/healthcare11060887. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 65.Shirke S. Medical writing on an accelerated path in India. Perspect Clin Res. 2015;6:125–8. doi: 10.4103/2229-3485.159934. [DOI] [PMC free article] [PubMed] [Google Scholar]