Skip to main content
Arquivos Brasileiros de Cirurgia Digestiva : ABCD logoLink to Arquivos Brasileiros de Cirurgia Digestiva : ABCD
. 2023 May 8;36:e1727. doi: 10.1590/0102-672020230002e1727

FUTURE OF THE LANGUAGE MODELS IN HEALTHCARE: THE ROLE OF CHATGPT

FUTURO DOS MODELOS DE LINGUAGEM NOS CUIDADOS EM SAÚDE: O PAPEL DO CHATGPT

Francisco Tustumi 1, Nelson Adami Andreollo 2, José Eduardo de Aguilar-Nascimento 3
PMCID: PMC10168663  PMID: 37162073

ABSTRACT

The field of medicine has always been at the forefront of technological innovation, constantly seeking new strategies to diagnose, treat, and prevent diseases. Guidelines for clinical practice to orientate medical teams regarding diagnosis, treatment, and prevention measures have increased over the years. The purpose is to gather the most medical knowledge to construct an orientation for practice. Evidence-based guidelines follow several main characteristics of a systematic review, including systematic and unbiased search, selection, and extraction of the source of evidence. In recent years, the rapid advancement of artificial intelligence has provided clinicians and patients with access to personalized, data-driven insights, support and new opportunities for healthcare professionals to improve patient outcomes, increase efficiency, and reduce costs. One of the most exciting developments in Artificial Intelligence has been the emergence of chatbots. A chatbot is a computer program used to simulate conversations with human users. Recently, OpenAI, a research organization focused on machine learning, developed ChatGPT, a large language model that generates human-like text. ChatGPT uses a type of AI known as a deep learning model. ChatGPT can quickly search and select pieces of evidence through numerous databases to provide answers to complex questions, reducing the time and effort required to research a particular topic manually. Consequently, language models can accelerate the creation of clinical practice guidelines. While there is no doubt that ChatGPT has the potential to revolutionize the way healthcare is delivered, it is essential to note that it should not be used as a substitute for human healthcare professionals. Instead, ChatGPT should be considered a tool that can be used to augment and support the work of healthcare professionals, helping them to provide better care to their patients.

HEADINGS: Guidelines as topic, Artificial intelligence, Diagnosis, Costs and cost analysis, Delivery of health care

INTRODUCTION

The field of medicine has always been at the forefront of technological innovation, constantly seeking new strategies to diagnose, treat, and prevent diseases. The number of publications in the main medical databases, such as PubMed and Embase, has been increasing steadily over the years due to the contributions from authors worldwide, the expansion of research areas, the rise of open-access publishing, and advancements in technology 2 . This growth in biomedical literature provides a wealth of information that can be used to advance research and improve patient care 7 . However, this fast science production in the medical field has boosted a new problem: how can clinical practice follow the constantly updated improvement of science?

Guidelines for clinical practice to orientate medical teams regarding diagnosis, treatment, and prevention measures have increased over the years 6 . The purpose is to gather the most medical knowledge to construct an orientation for practice. Guidelines help standardize care, improve patient outcomes, promote efficient use of resources, and reduce the risk of adverse events. In this sense, evidence-based reviews and guidelines are located at the top of the pyramid in the level of evidence (Figure 1) 14 and in the last stages of translational medicine (Figure 2) 18 .

Figure 1. Pyramid of the level of evidence. Systematic reviews and evidence-based guidelines are usually considered the top evidence, gathering all the current evidence for clinical practice 14 .

Figure 1

Figure 2. Translational medicine stages. Evidence-based guidelines and healthcare policies are the endpoints of any research line. These guidelines and policies orientate healthcare professionals for patient management 18 .

Figure 2

Evidence-based guidelines follow several main characteristics of a systematic review, including systematic and unbiased search, selection, and extraction of the source of evidence 21,24 . However, there are numerous obstacles to constructing evidence-based guidelines. Search, selection, and extraction take many efforts and a long time 5 . Consequently, evidence-based guidelines are usually published with a long-time delay. An evidenced-based guideline can take anywhere from several months to more than a year to complete, depending on the scope and complexity of the review. Conducting a comprehensive search for relevant studies can take several months, depending on the literature database's size and the search terms’ complexity. Screening studies for inclusion based on pre-specified criteria can take another couple of months, depending on the number of studies identified and the number of reviewers involved. Updates of these guidelines may take years or even never be performed, and healthcare professionals may take their diagnosis, treatment, and prevention measures based on guidelines out of step with the current technological and scientific level.

Another relevant issue in evidence-based guidelines and systematic reviews refers to the risk of selection bias 10,22,23 . Some authors may make a poor selection of the pieces of evidence for supporting guidelines for some reasons which may be due to quite restricted eligibility criteria, such as period, language, or databases searched, or even due to human failure during the process of search, selection, and extraction of the source of evidence 17 . Besides, poor selection can also be due to improper manipulation of outcomes, influenced by personal beliefs or opinions 8 .

In this context, in recent years, the rapid advancement of artificial intelligence (AI) has provided new opportunities for healthcare professionals to improve patient outcomes, increase efficiency, and reduce costs 25 . In the medical field, AI has the potential to provide clinicians and patients with access to personalized, data-driven insights and support.

One of the most exciting developments in AI has been the emergence of chatbots, which have the potential to revolutionize the way healthcare is delivered 1 . A chatbot is a computer program used to simulate conversations with human users. Recently, OpenAI, a research organization focused on machine learning, developed ChatGPT, a large language model that generates human-like text. The initial version of Generative Pre-Trained Transformer (GPT) was first introduced in 2018. Since then, several improved versions of GPT have been released. ChatGPT is a variant of the GPT models that have been specifically fine-tuned and optimized for chat-based applications 13 .

ChatGPT uses a type of AI known as a deep learning model (Figure 3). Specifically, it is a type of transformer-based language model that uses a neural network architecture known as the Transformer. This architecture was first introduced in 2017 and has since become a popular choice for natural language processing tasks, such as language translation, language generation, and text classification. The Transformer architecture uses multiple layers of neurons, or “transformer blocks,” that allow it to process and extract features from input data hierarchically. This architecture has proven to be very effective for natural language processing tasks, and it forms the basis of many of the most advanced language models, including ChatGPT. ChatGPT is designed to understand and respond to text-based inputs from users in order to provide helpful and informative responses to their questions and comments 4,11,12 .

Figure 3. Modern language models use artificial intelligence to create a human-like text.

Figure 3

There are different categories of chatbots, and a chatbot can belong to more than one category: Knowledge Domain (generic, open domain, and closed domain); Service Provided (interpersonal, intrapersonal, and inter-agent); Goals (informative, chat-based/conversational, and task-based); Response Generation Method (rule based, retrieval based, and generative); Human-Aid (human-mediated and autonomous); Permissions (open-source and commercial); and Communication Channel (text, voice, and image) 1 . Therefore, there are many applications of chatbots, like education environments, customer service, medicine and health, robotics, industrial, and others 1 .

The chatbots, AI, and telemedicine are increasingly being used in healthcare services with good acceptance, such as education, diagnostic imaging and genetic diagnosis, as well as clinical laboratory, screening, and health communications 9,15,25 .

One of the key benefits of using ChatGPT in medicine is the ability to provide fast, accurate, and up-to-date information to healthcare professionals and patients. ChatGPT can quickly search and select pieces of evidence through numerous databases to provide answers to complex questions, reducing the time and effort required to research a particular topic manually. Consequently, language models can accelerate the creation of clinical practice guidelines. AI may help screen numerous databases quickly, saving time and accelerating the finishing of the guidelines 11,20 .

In addition to providing information, ChatGPT can also assist healthcare professionals with making diagnoses by analyzing symptoms and recommending tests or treatments. This can help to improve the accuracy of diagnoses and reduce the number of misdiagnoses. ChatGPT can also be used to help manage patients with chronic conditions by providing information about medications, lifestyle changes, and treatment options 3,15 .

Besides, the use of AI might reduce the risk of selection bias. The dataset used to train ChatGPT is typically chosen to be as inclusive as possible, including a broad range of sources and perspectives. ChatGPT is not influenced by individual cognitive biases, which can affect human decision-making. As an AI, ChatGPT uses a purely algorithmic approach to generate responses without being influenced by personal beliefs or opinions. This can reduce the risk of selection bias, as the responses are based solely on the data used to train the algorithm 16 .

However, ChatGPT is trained on large amounts of data, which may contain biases or inaccuracies that the system could inadvertently propagate. This information bias could result in inaccurate or discriminatory recommendations or treatment options 15,16 . ChatGPT is trained based on web data, and several web information sources may be easily wrong. Besides, ChatGPT may not always have a full understanding of the context of the data in which a question or prompt is being asked. This could lead to inaccurate or inappropriate responses, particularly in complex medical situations or situations comprehending patients’ preferences or feelings.

For evidence-based guidelines with quantitative synthesis (meta-analysis), despite AI potentially detecting statistical heterogeneity, clinical heterogeneity depends on the critical analysis of the included articles 19 . Without the human critical analysis capacity, language models may gather information in a review based on articles whose data cannot be pooled due to their methodological differences!

While there is no doubt that ChatGPT has the potential to revolutionize the way healthcare is delivered, it is essential to note that it should not be used as a substitute for human healthcare professionals. Instead, ChatGPT should be considered a tool that can be used to augment and support the work of healthcare professionals, helping them to provide better care to their patients (Figure 4).

Figure 4. Language models in medicine are promising and have the potential to significantly improve patient outcomes and increase the efficiency of healthcare delivery. However, it is crucial to consider the limitations of this technology carefully. AI: Artificial intelligence.

Figure 4

CONCLUSION

The use of ChatGPT in medicine is a promising development that has the potential to significantly improve patient outcomes and increase the efficiency of healthcare delivery. However, it is important to consider this technology's limitations carefully and ensure that it is used responsibly and in conjunction with human healthcare professionals.

Footnotes

Financial Source: None

Editorial Support: National Council for Scientific and Technological Development (CNPq).

Central Message

In recent years, the rapid advancement of artificial intelligence (AI) has provided clinicians and patients with access to personalized, data-driven insights, support and new opportunities for healthcare professionals to improve patient outcomes, increase efficiency, and reduce costs. One of the most exciting developments in AI has been the emergence of chatbots. A chatbot is a computer program used to simulate conversations with human users.

Perspectives

The use of ChatGPT in medicine is a promising development that has the potential to significantly improve patient outcomes and increase the efficiency of healthcare delivery. However, it is important to consider this technology's limitations carefully and ensure that it is used responsibly and in conjunction with human healthcare professionals.

REFERENCES

  • 1.Adamopoulou E, Moussiades L. Chatbots: history, technology, and applications. Mach Learn Appl. 2020;2:100006–100006. doi: 10.1016/j.mlwa.2020.100006. [DOI] [Google Scholar]
  • 2.AlRyalat SAS, Malkawi LW, Momani SM. Comparing bibliometric analysis using PubMed, scopus, and web of science databases. J Vis Exp. 2019;(152) doi: 10.3791/58494. [DOI] [PubMed] [Google Scholar]
  • 3.Athota L, Shukla VK, Pandey N, Rana A. Chatbot for healthcare system using artificial intelligence. IEEE. 2020:619–622. doi: 10.1109/ICRITO48877.2020.9197833. [DOI] [Google Scholar]
  • 4.Aydın Ö, Karaarslan E. In: Emerging computer technologies. Aydın Ö, editor. Tire: İzmir Akademi Dernegi; 2022. OpenAI ChatGPT generated literature review: digital twin in healthcare; pp. 22–31. [Google Scholar]
  • 5.Boland A, Dickson R, Cherry G. Doing a systematic review. 2nd edition. SAGE Publications; 2017. Doing a systematic review: a student's guide; pp. 1–304. [Google Scholar]
  • 6.Brouwers MC, Florez ID, McNair SA, Vella ET, Yao X. Clinical practice guidelines: tools to support high quality patient care. Semin Nucl Med. 2019;49(2):145–152. doi: 10.1053/j.semnuclmed.2018.11.001. [DOI] [PubMed] [Google Scholar]
  • 7.Engle RL, Mohr DC, Holmes SK, Seibert MN, Afable M, Leyson J, Meterko M. Evidence-based practice and patient-centered care: doing both well. Health Care Manage Rev. 2021;46(3):174–184. doi: 10.1097/HMR.0000000000000254. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.Hernán MA, Hernández-Díaz S, Robins JM. A structural approach to selection bias. Epidemiology. 2004;15(5):615–625. doi: 10.1097/01.ede.0000135174.63482.43. [DOI] [PubMed] [Google Scholar]
  • 9.Isolan G, Malafaia O. How does telemedicine fit into healthcare today? Arq Bras Cir Dig. 2022;34(3):e1584–e1584. doi: 10.1590/0102-672020210003e1584. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10.Kho ME, Duffett M, Willison DJ, Cook DJ, Brouwers MC. Written informed consent and selection bias in observational studies using medical records: systematic review. BMJ. 2009;338:b866–b866. doi: 10.1136/bmj.b866. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11.Kung TH, Cheatham M, Medenilla A, Sillos C, De Leon L, Elepaño C, et al. Performance of ChatGPT on USMLE: potential for AI-assisted medical education using large language models. PLOS Digital Health. 2023;2(2):e0000198. doi: 10.1371/journal.pdig.0000198. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12.Li N, Liu S, Liu Y, Zhao S, Liu M. Neural speech synthesis with transformer network. Arxiv. 2019;33(1):6706–6713. doi: 10.1609/aaai.v33i01.33016706.. [DOI] [Google Scholar]
  • 13.Lund BD, Wang T. Chatting about ChatGPT: how may AI and GPT impact academia and libraries? Library Hi Tech News. 2023 doi: 10.1108/LHTN-01-2023-0009. [DOI] [Google Scholar]
  • 14.Murad MH, Asi N, Alsawas M, Alahdab F. New evidence pyramid. Evid Based Med. 2016;21(4):125–127. doi: 10.1136/ebmed-2016-110401. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15.Nadarzynski T, Miles O, Cowie A, Ridge D. Acceptability of artificial intelligence (AI)-led chatbot services in healthcare: a mixed-methods study. Digit Health. 2019;5:2055207619871808–2055207619871808. doi: 10.1177/2055207619871808.. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16.Ntoutsi E, Fafalios P, Gadiraju U, Iosifidis V, Nejdl W, Vidal ME, et al. Bias in data-driven artificial intelligence systems—An introductory survey. WIREs Data Mining Knowl Discov. 2020;10(3):e1356–e1356. doi: 10.1002/widm.1356. [DOI] [Google Scholar]
  • 17.Rother ET. Systematic literature review X narrative review. Acta Paul Enferm. 2007;20:v–i. doi: 10.1590/S0103-21002007000200001. [DOI] [Google Scholar]
  • 18.Seyhan AA. Lost in translation: the valley of death across preclinical and clinical divide–identification of problems and overcoming obstacles. Transl Med Commun. 2019;4:18–18. doi: 10.1186/s41231-019-0050-7. [DOI] [Google Scholar]
  • 19.Thompson SG. Why sources of heterogeneity in meta-analysis should be investigated. BMJ. 1994;309(6965):1351–1355. doi: 10.1136/bmj.309.6965.1351. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 20.Thorp HH. ChatGPT is fun, but not an author. Science. 2023;379(6630):313–313. doi: 10.1126/science.adg7879. [DOI] [PubMed] [Google Scholar]
  • 21.Urra Medina E, Barría Pailaquilén RM. Systematic review and its relationship with evidence-based practice in health. Rev Latino-am Enfermagem. 2010;18:824–831. doi: 10.1590/S0104-11692010000400023. [DOI] [PubMed] [Google Scholar]
  • 22.Williamson PR, Gamble C, Altman DG, Hutton JL. Outcome selection bias in meta-analysis. Stat Methods Med Res. 2005;14(5):515–524. doi: 10.1191/0962280205sm415oa. [DOI] [PubMed] [Google Scholar]
  • 23.Williamson PR, Gamble C. Identification and impact of outcome selection bias in meta-analysis. Stat Med. 2005;24(10):1547–1561. doi: 10.1002/sim.2025. [DOI] [PubMed] [Google Scholar]
  • 24.Yao X, Vella ET, Sussman J. More thoughts than answers: what distinguishes evidence-based clinical practice guidelines from non-evidence-based clinical practice guidelines? J Gen Intern Med. 2021;36(1):207–208. doi: 10.1007/s11606-020-05825-y. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25.Yu KH, Beam AL, Kohane IS. Artificial intelligence in healthcare. Nat Biomed Eng. 2018;2(10):719–731. doi: 10.1038/s41551-018-0305-z. [DOI] [PubMed] [Google Scholar]

Articles from Arquivos Brasileiros de Cirurgia Digestiva : ABCD are provided here courtesy of Colégio Brasileiro de Cirurgia Digestiva

RESOURCES