Abstract
The utilization of artificial intelligence (AI) in clinical practice has increased and is evidently contributing to improved diagnostic accuracy, optimized treatment planning, and improved patient outcomes. The rapid evolution of AI, especially generative AI and large language models (LLMs), have reignited the discussions about their potential impact on the healthcare industry, particularly regarding the role of healthcare providers. Concerning questions, “can AI replace doctors?” and “will doctors who are using AI replace those who are not using it?” have been echoed. To shed light on this debate, this article focuses on emphasizing the augmentative role of AI in healthcare, underlining that AI is aimed to complement, rather than replace, doctors and healthcare providers. The fundamental solution emerges with the human–AI collaboration, which combines the cognitive strengths of healthcare providers with the analytical capabilities of AI. A human-in-the-loop (HITL) approach ensures that the AI systems are guided, communicated, and supervised by human expertise, thereby maintaining safety and quality in healthcare services. Finally, the adoption can be forged further by the organizational process informed by the HITL approach to improve multidisciplinary teams in the loop. AI can create a paradigm shift in healthcare by complementing and enhancing the skills of healthcare providers, ultimately leading to improved service quality, patient outcomes, and a more efficient healthcare system.
Keywords: Artificial intelligence, large language models, generative AI, doctor, healthcare provider, medicine, implementation
Introduction
The advancements in artificial intelligence (AI) have provided a wealth of opportunities for clinical practice and healthcare. Large language models (LLMs), such as BERT, GPT, and LaMDA have experienced exponential growth, with some now containing over a trillion parameters. 1 This growth in AI capabilities allows for seamless integration between different types of data and has led to multimodal applications in various domains, including medicine. 2 Evidence shows that AI has the potential to improve healthcare delivery by enhancing diagnostic accuracy, optimizing treatment planning, and improving patient outcomes.3–5 With the recent developments in AI, specifically LLM and generative AI (e.g. DALL-E, GPT-4 via ChatGPT), we reassess benefits and opportunities presented by AI toward being one step closer to an artificial general intelligence (AGI, AI with human cognitive abilities).6,7 Current evidence showed LLM capabilities with medical knowledge and support. The University of Florida's GatorTron, an 8.9 billion parameter LLM, is one of the first medical foundation models developed by an academic health system and medical data. 8 It is designed to improve five clinical natural language processing tasks, such as medical question answering and medical relation extraction. LLM further showed capability of medical knowledge, as an AI model achieved a 79.5% accuracy rate on the U.K. Royal College of Radiology examination compared to 84.8% for human radiologists. 9 Recently, the LLMs (PaLM, GPT) demonstrated their capabilities for the United States Medical Licensing Examination and several other medical question-answering tasks, showcasing the potential of AI in medicine.10,11
The increased capabilities also spread concerning conversations. These include AI and the alignment problem, 12 and ethical and unbiased implementations. 13 Recently, there has been a movement towards urging a “pause” in AI development, 14 to address these concerns, as well as to investigate societal implications, to build robust frameworks, governance, and control mechanisms. Among all, the notion of “AI taking over human jobs,” as it achieves highly accurate results and performance in completing human tasks, 15 has been one of the emerging concerns. In line with that, in the healthcare domain, the question has echoed: Can AI replace doctors? (“doctors” is referring to healthcare providers in this article, including physicians, nurses, and other healthcare providers), or will they serve as invaluable tools that complement and support them?16,17
AI to replace doctors
Even though the idea is intriguing, fundamentally, AI is not meant (designed and developed) to replace doctors, but able to repurpose roles and improve efficiency, as demonstrated by LLM-powered digital scribes and conversation summarization tools. 18 If we step back and look at current applications in clinical practice, AI has already been an integral part of health services, without replacing doctors. For example, AI-aided decision support systems with ultrasound or MRI machines to assist diagnosis, 19 or improving voice recognition in dictation devices to keep radiology notes. 20 However, recent developments in AI are highly complex, rapidly evolving, and overwhelmingly positive—as seen in the increased accuracy of LLMs in completing tasks, high language comprehension, and human-like conversational responses—leading us to question their value and contribution to practice.
AI to collaborate with doctors
By repurposing (not replacing) the roles, AI can contribute to a more efficient and streamlined healthcare system. Does that mean doctors using AI replace those not using it? This statement has been highly popular recently against the “AI replacing doctors” argument. The statement is true at its core. A decision by a doctor aided by AI could be more accurate (and timely) than without AI, 21 as minimizing risk for patients and improving decision-making process, the quality of service, and efficiency. 22 However, one feasible way to enable the “use” of AI goes through “collaboration.” Following human-in-the-loop and human-AI collaboration principles, a structure for utilizing the potential of AI can be established.23,24 Human-in-the-Loop (HITL) approach emphasizes a collaborative partnership between AI and human expertise to optimize outcomes. Through collaborative decision-making, AI offers insights, and individuals leverage their knowledge for final judgment, establishing oversight and quality control to validate AI predictions, reducing potential errors or biases. The collaboration nurtures continuous learning and improvement as both parties learn from each other. This further contributes to trust and acceptance. Ethical practices are important ensuring transparency, accountability and explainability in AI decisions. The human feedback improves AI adaptability, enabling AI to handle complex cases beyond training data.
In healthcare, HITL could be implemented as the trained doctors could collaborate with AI, monitor, validate and guide the process, interpret AI outputs, and provide feedback to improve the capability and accuracy of AI. A recent study showed that AI could enhance the accuracy of diagnosis and clinical decisions when combined with expert human evaluation, emphasizing the collaborative nature of AI and doctors. 25 This collaboration can further contribute to an often overseen value proposition toward addressing the disparities. AI has a higher value to act as a complementary tool, or knowledge augmentation mechanism, to fill the gaps, particularly in low-resource settings, such as rural areas or underdeveloped countries toward improving diagnosis, patient communication and education, and reducing language barriers. 26
However, such AI collaboration and decision support mechanisms are available via the adoption of organizations rather than solely based on personal choices in healthcare settings.
AI to be adopted by healthcare organizations
The adoption of AI is driven by organizational decisions, necessity, and readiness. 27 Therefore, the ultimate question is: will healthcare organizations successfully adopt AI?
Healthcare organizations (e.g., hospitals and clinics) are responsible for providing AI tools that have undergone rigorous evaluation and validation to ensure safety and effectiveness for clinical practice (e.g., FDA clearance and FTC guidelines).28–30 In addition to that, legal, infrastructure, privacy, and security teams need to revisit organizational policies and protocols to ensure compliance, including state and federal laws and regulations, with a specific focus on personal health information exchange protocols, accountability, liability, service reimbursement, and clinical workflows.31,32 In parallel, there is a need to develop curricula and educational methods to train doctors on the fundamentals of AI, effective use in practice, and AI-supported healthcare delivery. 33
The organizational process can be informed by the HITL approach to improve the multidisciplinary team in the loop. 34 Enabling human–AI collaboration and inclusion of human feedback and control can forge the partnership, diminishing the false perception of “AI as replacement” (Figure 1). The specific considerations for collaborative AI adoption in a healthcare organization to enable human–AI collaboration are summarized in Textbox 1.
Figure 1.
AI adoption to enable doctor–AI collaboration and considerations.
AI, artificial intelligence.
Textbox 1.
Considerations for AI adoption in a healthcare organization
1. Establish multidisciplinary teams (clinic, research, information systems,
operation, management and administration, patient and community advocate) to
explore and evaluate cost-effective and impactful collaborative AI solutions and
establish HITL protocols. This requires collaboration and knowledge sharing among
team members to anticipate utility and expected outcomes.35,36 2. Prioritize clinical processes, operational workflows, and practices for AI support which need improvements and the leverage from AI collaboration, such as tasks and processes contributing to burnout, tasks leading to inefficiency or low performance, tasks to address patient needs, service quality, and satisfaction.37–39 3. Involve multistakeholder groups, including doctors, nurses, administrators, and patients, in identifying essential inclusive training, education needs, and cultural transformation, and in testing AI tools for effective collaboration before large-scale deployment.34,40 This approach ensures that the perspectives of all included parties are communicated and considered in the development and implementation of AI solutions. 4. Establish rigorous evaluation methods and assessment frameworks for AI, focusing on validation, verification, utility, and adoption.41,42 It requires monitoring and testing AI in a controlled environment and a real-world setting with a range of capabilities before expanding to broader applications. 43 This will enable organizations to identify potential issues and lower the risks, refine AI systems, train human collaborators, and measure their impact on patient care before extensive implementation. 5. Revise organizational policies and protocols to facilitate AI adoption and address ethical and legal concerns, ensuring compliance, privacy, and security, and strategizing to build organizational trust, access, and governance.36,44,45 This includes developing guidelines for AI collaboration and practice transparency, accountability, and explainability to minimize risks and ensure patient safety.43,44,46 6. Commit to ethical, inclusive, equitable, responsible, and fair AI practices.34,47 This requires focusing on reducing the digital divide among healthcare organizations. 48 We need to ensure equitable access to collaborative AI tools, training, and resources among hospitals, practitioners, and communities by developing partnerships and initiatives that promote inclusive access to technology and skills development. |
Conclusion
The advancements in AI are reassuring, showing promise in creating a paradigm shift in healthcare by complementing and enhancing the skills of doctors and healthcare providers rather than replacing them. To successfully harness the power of AI, healthcare organizations must be proactive, especially now, where generative AI and LLMs are highly accessible but still in need of control and guidance. As AI becomes an essential component of modern healthcare, it is vital for organizations to invest in the necessary infrastructure, training, resources, and partnerships to support its successful adoption and ensure equitable access for all.
Acknowledgements
Figure 1 is created with BioRender.com.
Footnotes
The author declared no potential conflicts of interest with respect to the research, authorship, and/or publication of this article.
Funding: The author received no financial support for the research, authorship, and/or publication of this article.
ORCID iD: Emre Sezgin https://orcid.org/0000-0001-8798-9605
References
- 1.Mitchell M, Krakauer DC. The debate over understanding in AI’s large language models. Proc Natl Acad Sci USA 2023; 120: e2215907120. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 2.Acosta JN, Falcone GJ, Rajpurkar Pet al. et al. Multimodal biomedical AI. Nat Med 2022; 28: 1773–1784. [DOI] [PubMed] [Google Scholar]
- 3.Gadaleta M, Radin JM, Baca-Motes K, et al. Passive detection of COVID-19 with wearable sensors and explainable machine learning algorithms. NPJ Digit Med 2021; 4: 66. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 4.Jiang F, Jiang Y, Zhi H, et al. Artificial intelligence in healthcare: past, present and future. Stroke Vasc Neurol 2017; 2: 230–243. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 5.Filipp FV. Opportunities for artificial intelligence in advancing precision medicine. Curr Genet Med Rep 2019; 7: 208–213. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 6.Bubeck S, Chandrasekaran V, Eldan R, et al. Sparks of artificial general intelligence: Early experiments with GPT-4. arXiv [csCL], http://arxiv.org/abs/2303.12712 (2023).
- 7.Fei N, Lu Z, Gao Y, et al. Towards artificial general intelligence via a multimodal foundation model. Nat Commun 2022; 13: 3094. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8.Yang X, Chen A, PourNejatian N, et al. A large language model for electronic health records. NPJ Digit Med 2022; 5: 94. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 9.Shelmerdine SC, Martin H, Shirodkar Ket al. et al. FRCR-AI Study Collaborators. Can artificial intelligence pass the Fellowship of the Royal College of Radiologists examination? Multi-reader diagnostic accuracy study. Br Med J. 2022; 379: e072826. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 10.Kung TH, Cheatham M, Medenilla A, et al. Performance of ChatGPT on USMLE: potential for AI-assisted medical education using large language models. PLOS Digit Health 2023; 2: e0000198. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 11.Singhal K, Azizi S, Tu T, et al. Large language models encode clinical knowledge. arXiv [csCL], http://arxiv.org/abs/2212.13138 (2022).
- 12.Harrer S. Attention is not all you need: the complicated case of ethically using large language models in healthcare and medicine. EBioMedicine 2023; 90: 104512. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 13.Illia L, Colleoni E, Zyglidopoulos S. Ethical implications of text generation in the age of artificial intelligence. Business Ethics Env Resp 2023; 32: 201–210. [Google Scholar]
- 14.Pause giant AI experiments: An open letter. Future of Life Institute, https://futureoflife.org/open-letter/pause-giant-ai-experiments/ (2023, Accessed 11 May 2023).
- 15.DePillis L, Lohr S. Tinkering with ChatGPT, workers wonder: will this take my job? The New York Times. https://www.nytimes.com/2023/03/28/business/economy/jobs-ai-artificial-intelligence-chatgpt.html (2023, accessed 3 April 2023).
- 16.Satariano A, Metz C. For AS. Using A.I. to detect breast cancer that doctors miss. The New York Times, https://www.nytimes.com/2023/03/05/technology/artificial-intelligence-breast-cancer-detection.html (2023, accessed 3 April 2023).
- 17.Goldhahn J, Rampton V, Spinas GA. Could artificial intelligence make doctors obsolete? Br Med J 2018; 363: k4563. [DOI] [PubMed] [Google Scholar]
- 18.van Buchem MM, Boosman H, Bauer MP, et al. The digital scribe in clinical practice: a scoping review and research agenda. NPJ Digit Med 2021; 4: 57. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 19.IntelliSpace Portal 9.0. Philips, https://www.usa.philips.com/healthcare/product/HC881072/intellispace-portal-advanced-visualization-solution (accessed 3 April 2023).
- 20.Sistrom CL, Langlotz CP. A framework for improving radiology reporting. J Am Coll Radiol 2005; 2: 159–167. [DOI] [PubMed] [Google Scholar]
- 21.Haupt CE, Marks M. AI-Generated Medical advice-GPT and beyond. JAMA 2023; 329: 1349. [DOI] [PubMed] [Google Scholar]
- 22.Miller DD, Brown EW. Artificial intelligence in medical practice: the question to the answer? Am J Med 2018; 131: 129–133. [DOI] [PubMed] [Google Scholar]
- 23.Wu X, Xiao L, Sun Y, et al. A survey of human-in-the-loop for machine learning. Future Gener Comput Syst 2022; 135: 364–381. [Google Scholar]
- 24.Cai CJ, Winter S, Steiner D, et al. “Hello AI”: uncovering the onboarding needs of medical practitioners for human-AI collaborative decision-making. Proc ACM Hum-Comput Interact 2019; 3: 1–24.34322658 [Google Scholar]
- 25.Schaffter T, Buist DSM, Lee CI, et al. Evaluation of combined artificial intelligence and radiologist assessment to interpret screening mammograms. JAMA Netw Open 2020; 3: e200265. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 26.Wahl B, Cossy-Gantner A, Germann Set al. et al. Artificial intelligence (AI) and global health: how can AI contribute to health in resource-poor settings? BMJ Glob Health 2018; 3: e000798. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 27.Yang J, Luo B, Zhao Cet al. et al. Artificial intelligence healthcare service resources adoption by medical institutions based on TOE framework. Digit Health 2022; 8: 20552076221126034. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 28.Center for Devices, Radiological Health. Artificial intelligence and machine learning in software as a medical device. U.S. Food and Drug Administration, https://www.fda.gov/medical-devices/software-medical-device-samd/artificial-intelligence-and-machine-learning-software-medical-device (2021, accessed 3 April 2023).
- 29.Using artificial intelligence and algorithms. Federal Trade Commission, https://www.ftc.gov/business-guidance/blog/2020/04/using-artificial-intelligence-and-algorithms (2020, accessed 3 April 2023).
- 30.Nicholson Price W II. Health care AI: Law, regulation, and policy, https://repository.law.umich.edu/book_chapters/357/ (2019, accessed 3 April 2023).
- 31.Wolff J, Pauling J, Keck Aet al. et al. Success factors of artificial intelligence implementation in healthcare. Front Digit Health 2021; 3(594971. doi: 10.3389/fdgth.2021.594971 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 32.Abràmoff MD, Roehrenbeck C, Trujillo S, et al. A reimbursement framework for artificial intelligence in healthcare. NPJ Digit Med 2022; 5(1): 72. doi: 10.1038/s41746-022-00621-w [DOI] [PMC free article] [PubMed] [Google Scholar]
- 33.Wartman SA, Combs CD. Reimagining medical education in the age of AI. AMA J Ethics 2019; 21: E146–E152. [DOI] [PubMed] [Google Scholar]
- 34.Boch S, Sezgin E, Lin Linwood S. Ethical artificial intelligence in paediatrics. Lancet Child Adolesc Health 2022; 6: 833–835. [DOI] [PubMed] [Google Scholar]
- 35.Hogg HDJ, Al-Zubaidy M, Technology Enhanced Macular Services Study Reference Group , et al. Stakeholder perspectives of clinical artificial intelligence implementation: systematic review of qualitative evidence . J Med Internet Res. 2023;25:e39742. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 36.Liao F, Adelaine S, Afshar Met al. et al. Governance of clinical AI applications to facilitate safe and equitable deployment in a large health system: key elements and early successes. Front Digit Health 2022; 4: 931439. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 37.CHAI. https://coalitionforhealthai.org/insights (2023, accessed 11 May 2023).
- 38.Smart use of artificial intelligence in health care. Deloitte Insights, https://www2.deloitte.com/us/en/insights/industry/health-care/artificial-intelligence-in-health-care.html (2020, accessed 11 May 2023).
- 39.Ash M, Petro J, Rab S. How AI in the exam room could reduce physician burnout. Harvard Business Review, https://hbr.org/2019/11/how-ai-in-the-exam-room-could-reduce-physician-burnout (2019, accessed 11 2023).
- 40.Reverberi C, Rigon T, Solari A, et al. Experimental evidence of effective human-AI collaboration in medical decision-making. Sci Rep 2022; 12: 14952. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 41.U.S. Government Accountability Office. Artificial intelligence in health care: Benefits and challenges of machine learning technologies for medical diagnostics, https://www.gao.gov/products/gao-22-104629 (2022, accessed 11 May 2023).
- 42.Coravos A, Doerr M, Goldsack J, et al. Modernizing and designing evaluation frameworks for connected sensor technologies in medicine. NPJ Digit Med 2020; 3: 37. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 43.AI Risk Management Framework | NIST, https://www.nist.gov/itl/ai-risk-management-framework (2021, accessed 11 May 2023).
- 44.Sezgin E, Sirrianni J, Linwood SL. Operationalizing and implementing pretrained, large artificial intelligence linguistic models in the US health care system: outlook of generative pretrained transformer 3 (GPT-3) as a service model. JMIR Med Inform 2022; 10: e32875. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 45.Blueprint for an AI bill of rights. The White House, https://www.whitehouse.gov/ostp/ai-bill-of-rights/ (2022, accessed 14 May 2023).
- 46.The OECD Artificial Intelligence (AI) Principles, https://oecd.ai/en/ai-principles (2019, accessed 11 May 2023).
- 47.Barredo Arrieta A, Díaz-Rodríguez N, Del Ser J, et al. Explainable artificial intelligence (XAI): concepts, taxonomies, opportunities and challenges toward responsible AI. Inf Fusion 2020; 58: 82–115. [Google Scholar]
- 48.Riggins FJ, Dewan S. The digital divide: current and future research directions. J Assoc Inf Sys 2005; 6: 13. [Google Scholar]