Abstract
ChatGPT is a generative artificial intelligence chatbot which may have a role in medicine and science. We investigated if the freely available version of ChatGPT can produce a quality conference abstract using a fictitious but accurately calculated data table as applied by a non‐medically trained person. The resulting abstract was well written without obvious errors and followed the abstract instructions. One of the references was fictitious, known as ‘hallucination’. ChatGPT or similar programmes, with careful review of the product by authors, may become a valuable scientific writing tool. The scientific and medical use of generative artificial intelligence, however, raises many questions.
Keywords: artificial intelligence, ChatGPT, generative artificial intelligence
Key findings.
We investigated if ChatGPT can create a quality conference abstract using a fictitious data set.
The resulting abstract was well written without obvious errors. One of the references was made up.
Generative artificial intelligence like ChatGPT may become a valuable writing tool but also raises many questions.
Generative artificial intelligence (AI) is capturing global attention and multiple uses in science and medicine have been suggested. 1 , 2 , 3 , 4 A readily available system is the recently launched ChatGPT (Chat Generative Pre‐trained Transformer). It produces human‐like text trained on data from the internet originally written by humans, created by OpenAI, San Francisco. ChatGPT uses a typed or pasted ‘prompt’ which the chatbot responds to within seconds using natural language. This exchange can be continued and refined with additional ‘prompts’. 2 Since its release in November 2022, ChatGPT has been used by more than 100 million people. ChatGPT‐3 is free and can be optimised for speed with a subscription; a new more powerful version GPT‐4 was released in March 2023. 2
We investigated if the freely available version of ChatGPT can produce a grammatically and structurally correct conference abstract of reasonable quality, based on a standard set of instructions with a data set provided (Box 1). To avoid ChatGPT extracting from previously published abstracts or data, we used an adult disease with no data in children and created a fictitious, but accurately calculated data table. ChatGPT was applied by a non‐medically trained person (MPB) familiar with ChatGPT, based on the initial prompt and was not further corrected.
BOX 1. Abstract prompt and ChatGPT response.
Prompt
Write a 250 word abstract on plantar fasciitis in children with title, introduction, methods, results and discussion section using the data below. Add 2 references.
Double blind RCT comparing oral prednisolone and placebo for 10 days. Intention to treat. 89 in placebo group, 88 in steroid group. Children aged 8–12 years old. Outcome: return of normal foot function at 1, 3 and 6 months. No adverse events.
Group | Time (months) | n | n | % | (95%CI) |
---|---|---|---|---|---|
Placebo | 1 | 40 | 89 | 44.94 | (34.38–55.86) |
Placebo | 3 | 45 | 85 | 52.94 | (41.81–63.87) |
Placebo | 6 | 50 | 82 | 60.98 | (49.57–71.56) |
Prednisolone | 1 | 60 | 90 | 66.67 | (55.95–76.26) |
Prednisolone | 3 | 72 | 88 | 81.82 | (72.16–89.24) |
Prednisolone | 6 | 84 | 85 | 98.82 | (93.62–99.97) |
Comparison
1 month OR = 2.45 (1.34–4.49), P = 0.003.
3 months OR = 4.00 (2.01–7.97), P < 0.001.
6 months OR = 53.76 (7.13–405.63), P < 0.001.
Response
Title: Efficacy of oral prednisolone for plantar fasciitis in children: A double‐blind randomized controlled trial
Introduction: Plantar fasciitis is a common cause of heel pain in adults, but it is rarely reported in children. However, recent studies have shown that plantar fasciitis is a common cause of heel pain in children aged 8–12 years old. Although there is no standard treatment for plantar fasciitis in children, non‐steroidal anti‐inflammatory drugs and rest are commonly used. The aim of this study is to investigate the efficacy of oral prednisolone in treating plantar fasciitis in children.
Methods: A double‐blind randomized controlled trial was conducted on 177 children aged 8–12 years old, who were randomly assigned to receive either oral prednisolone or placebo for 10 days. The outcome measure was the return of normal foot function at 1, 3 and 6 months after treatment. The study was conducted with intention to treat, and no adverse events were reported.
Results: At 1, 3 and 6 months after treatment, the percentage of children who had returned to normal foot function was significantly higher in the prednisolone group than in the placebo group (66.67% vs 44.94%, 81.82% vs 52.94% and 98.82% vs 60.98%, respectively). The odds ratios for the return of normal foot function at 1, 3 and 6 months were 2.45 (95% CI 1.34–4.49, P = 0.003), 4.00 (95% CI 2.01–7.97, P < 0.001) and 53.76 (95% CI 7.13–405.63, P < 0.001), respectively.
Discussion: Our study suggests that oral prednisolone may be an effective treatment for plantar fasciitis in children. However, further studies are needed to confirm these findings and to investigate the long‐term effects of oral prednisolone on foot function in children with plantar fasciitis.
References:
Yan S, Li B, Hu Y et al. A systematic review and meta‐analysis of risk factors for pediatric plantar fasciitis. J. Foot Ankle Res. 2020; 13: 16. doi: 10.1186/s13047‐020‐0397‐1
Radford JA, Landorf KB, Buchbinder R, Cook C. Effectiveness of low‐dye taping for the short‐term treatment of plantar heel pain: a randomised trial. BMC Musculoskelet. Disord. 2006; 7: 64. doi: 10.1186/1471‐2474‐7‐64
The resulting abstract is impressive: it is well written with headings and word length as requested. The title is appropriate, the introductory sentences are consistent with the state of knowledge on plantar fasciitis, the results including the statistical tests were correctly extracted from the data table, and the conclusions are a reasonable interpretation of the results. There are no obvious errors. The references seem appropriate but only the second reference is real with the first reference fictitious; this has been noted before for ChatGPT and has been termed ‘hallucinations’. 2
The impact of generative AI in medicine is potentially revolutionary with proposed areas of use shown in Box 2. 1 , 2 , 4 As this small experiment has shown, ChatGPT allows a non‐scientist to produce a satisfactory conference abstract. We doubt even the current, uncorrected version would be rejected by an abstract assessment committee. Editorials and letters using ChatGPT have already been published. 5 , 6
BOX 2. Some current and possible future uses of artificial intelligence in medicine 1 , 2 , 3 , 4 .
Clinical medicine
Analysis of medical images
Detection of drug interactions
Identifying high‐risk patients
Coding of medical notes
Medical note taking
Medical consultation
Retrieval of medical information
Electronic medical record decision support
Translation of foreign languages
Explanation of laboratory and imaging tests
Research
Improved trial performance, for example via decision support, improved recruitment of participants, outcome monitoring
Analysis and interpretation of large research databases
Writing, revision and summarising of scientific papers and grants
Other uses
Role in public health, for example identify outbreaks
Assessor and teacher in medical education
Hospital business operations
ChatGPT or similar programmes when used as a writing tool with careful review by scientific authors may become a valuable tools to speed a first draft of papers and grants, summarise or simplify long articles, reduce duplication, polish and streamline existing documents or assist researchers with limited English skills (although it also works in other languages). It does not seem obvious why standard methodology or even results sections of a paper could not be written with the assistance of generative AI when large sections of research papers are proscribed in some detail – often with suggested wording by various EQUATOR reporting guidelines. 7 EQUATOR now includes guidance for studies including AI and machine learning.
Advantages of AI in medicine will need to be balanced with possible legal, ethical and patient safety challenges. 1 , 3 , 4 In its scientific use, there are many unanswered questions and issues: what are the ethical limits of its use and does its use need to be disclosed? Can or should AI like ChatGPT be listed as an author and how can its use be acknowledged? Recently published guidance by the World Association of Medical Editors states that chatbots cannot be authors and that authors be transparent when they are used. 8 AI may plagiarise the work of others; even designated tools to detect AI may be limited in their ability to detect AI written text and may falsely attribute human written text to AI. 4 Who is responsible for mistakes or harmful advice created by AI? Are data entered into ChatGPT and its outputs accessible to others as it is drawing on existing information and learning from its users? Could it lead to the amplification of biases, such as racial, sex and other stereotypes, or fuel the spread of misinformation? 3 , 6 , 8
It seems likely that AI, which is constantly evolving and improving, will be widely incorporated into medicine in general and scientific writing in particular. Researchers, journals and funding agencies will need to rapidly come to terms with it.
Acknowledgements
FEB's time was in part funded by an Investigator grant from the National Health and Medical Research Council (GNT2017605), Canberra, Australia, and the Royal Children's Hospital Foundation, Parkville, Australia. Open access publishing facilitated by The University of Melbourne, as part of the Wiley ‐ The University of Melbourne agreement via the Council of Australian University Librarians.
Competing interests
FEB is a section editor for Emergency Medicine Australasia. FEB and MPB are related.
Data availability statement
No data are available.
References
- 1. Haug CJ, Drazen JM. Artificial intelligence and machine learning in clinical medicine, 2023. N. Engl. J. Med. 2023; 388: 1201–8. [DOI] [PubMed] [Google Scholar]
- 2. Lee P, Bubeck S, Petro J. Benefits, limits, and risks of GPT‐4 as an AI chatbot for medicine. N. Engl. J. Med. 2023; 388: 1233–9. [DOI] [PubMed] [Google Scholar]
- 3. The Lancet Digital Health . ChatGPT: friend or foe? Lancet Digit. Health 2023; 5: e102. [DOI] [PubMed] [Google Scholar]
- 4. Stokel‐Walker C, Van Noorden R. The promise and peril of generative AI. Nature 2023; 614: 214–6. [DOI] [PubMed] [Google Scholar]
- 5. Curtis N, ChatGPT . To ChatGPT or not to ChatGPT? The impact of artificial intelligence on academic publishing. Pediatr. Infect. Dis. J. 2023; 42: 275. [DOI] [PubMed] [Google Scholar]
- 6. D'Amico RS, White TG, Shah HA, Langer DJ. I asked a ChatGPT to write an editorial about how we can incorporate chatbots into neurosurgical research and patient care…. Neurosurgery 2023; 92: 663–4. [DOI] [PubMed] [Google Scholar]
- 7. The EQUATOR (Enhancing the QUAlity and Transparency Of health Research) . [Cited 8 Apr 2023.] Available from URL: www.equator-network.org/
- 8. Zielinski C, Winker M, Aggarwal R et al. Chatbots, ChatGPT, and Scholarly Manuscripts. WAME Recommendations on ChatGPT and Chatbots in Relation to Scholarly Publications. World Association of Medical Editors, 2023. [Cited 22 Feb 2023.] Available from URL: https://wame.org/page3.php?id=106 [DOI] [PubMed]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Data Availability Statement
No data are available.