Dear Editor,
Recently, International Journal of Surgery has published clinical trials on generative artificial intelligence (AI).1 Clinical trials represent public events with direct ramifications on medical practice and patient well-being. However, a dearth of transparency and information disclosure could precipitate misdirection and hazards. In pursuit of upholding integrity and equity in medical research, the International Committee of Medical Journal Editors (ICMJE) issued a declaration in 2004 mandating that member journals, effective 1 July 2005, exclusively accept trial results registered with public clinical trial registries and require the registration number upon publication. This edict’s implementation propelled clinical trial registration into a requisite procedure. Presently, numerous clinical trial registration platforms exist, including the largest clinicaltrials.gov. Clinical trial registration empowers the public to independently access pertinent information such as ongoing trials, recruitment status, research designs, and anticipated outcomes, thereby enhancing research transparency and traceability. Simultaneously, registration protocols aid in thwarting selective reporting and outcome manipulation, thereby safeguarding the objectivity and credibility of research findings. Consequently, the establishment and enforcement of clinical trial registration systems hold paramount significance in upholding the ethical and scientific standards of medical research and safeguarding patient rights.
The advent of generative AI, employing diverse machine learning methods to learn patterns from data and create entirely original text, images, and even videos, marks a pivotal moment in artificial intelligence. The year 2022 was hailed as the ‘Year of Generative AI,’ with models like OpenAI’s ChatGPT representing a focal point in the AI landscape. In recent years, an increasing number of studies have incorporated generative AI into clinical research. Notably, a study published in Science showcased results from a clinical trial involving 453 highly educated professionals, demonstrating that the use of ChatGPT significantly enhanced productivity, improved task quality, and reduced task completion time2. To our knowledge, this is the first ChatGPT-related clinical trial with the complete registration number.
Since then, an array of clinical trials evaluating ChatGPT’s applications in oncology, radiology, surgery, and critical care medicine has emerged, contributing to a burgeoning body of research3-5. However, amidst these published clinical studies, a significant proportion has overlooked a crucial aspect of clinical trials: registration. A search on clinicaltrials.gov using ‘ChatGPT’ as a keyword yields only a dozen registered clinical studies, a stark mismatch compared to the number of published ChatGPT-related clinical trials. This discrepancy underscores the prevalent neglect of clinical registration in the current landscape of published trials.
Given the nascent nature of generative AI, substantial controversy persists within the scientific community regarding its principles and applications. Consequently, delineating which generative AI clinical trials necessitate registration proves challenging. In many of the currently published clinical studies, researchers utilize generative AI, such as ChatGPT, as a control group, comparing its efficacy with that of clinical professionals in diagnosis, treatment, and even prognosis prediction. The classification of these studies as quality improvement endeavors or clinical trials remains subject to further delineation. However, based on our interpretation of the ICMJE guidelines, all generative AI clinical trials involving human data should undergo prospective registration to ensure trial transparency and safety.
In light of the transformative potential of generative AI in healthcare, proactive measures are warranted to align research practices with ethical and scientific standards. As such, concerted efforts are needed to foster a culture of responsible conduct and accountability within the research community. By advocating for the registration of generative AI clinical trials, we could collectively advance the field toward greater transparency, rigor, and, ultimately, improved patient outcomes.
Ethical approval
This study does not include any individual-level data and thus does not require any ethical approval.
Sources of funding
This study is supported by China Postdoctoral Science Foundation (2022M720385) and Beijing JST Research Funding (YGQ-202313).
Author contribution
H.W.: conceptualization, methodology, data curation, formal analysis, resources, investigation, and writing – original draft; Z.S. and Q.G.: conceptualization, methodology, data curation, formal analysis, resources, and investigation; X.L. and K.C.: conceptualization and methodology; C.L.: methodology, data curation, formal analysis, resources, investigation, and writing – review and editing.
Conflicts of interest disclosure
The authors declare no conflicts of interest.
Research registration unique identifying number (UIN)
Name of the registry: not applicable.
Unique identifying number or registration ID: not applicable.
Hyperlink to your specific registration (must be publicly accessible and will be checked): not applicable.
Guarantor
Haiyang Wu and Cheng Li.
Data availability statement
The data underlying this article will be shared by the corresponding author on reasonable request.
Footnotes
Haiyang Wu, Zaijie Sun, and Qiang Guo contributed equally to the study.
Sponsorships or competing interests that may be relevant to content are disclosed at the end of this article.
Published online 22 May 2024
Contributor Information
Haiyang Wu, Email: wuhaiyang2021@tmu.edu.cn.
Zaijie Sun, Email: gukeszj@163.com.
Qiang Guo, Email: duduqiang@126.com.
Xiaojun Liu, Email: xiaojunliucn@163.com.
Kunming Cheng, Email: chengkm2013@163.com.
Cheng Li, Email: lichengcharite@gmail.com.
References
- 1. Deng L, Wang T, Yangzhang, et al. Evaluation of large language models in breast cancer clinical scenarios: a comparative analysis based on ChatGPT-3.5, ChatGPT-4.0, and Claude2. Int J Surg 2024;110:1941–1950. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 2. Noy S, Zhang W. Experimental evidence on the productivity effects of generative artificial intelligence. Science 2023;381:187–192. [DOI] [PubMed] [Google Scholar]
- 3. Zhu L, Lai Y, Mou W, et al. ChatGPT’s ability to generate realistic experimental images poses a new challenge to academic integrity. J Hematol Oncol 2024;17:27. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 4. Huang J, Yang DM, Rong R, et al. A critical assessment of using ChatGPT for extracting structured data from clinical notes. NPJ Digit Med 2024;7:106. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 5. Gakuba C, Le Barbey C, Sar A, et al. Evaluation of ChatGPT in predicting 6-month outcomes after traumatic brain injury. Crit Care Med 2024;52:942–950. [DOI] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Data Availability Statement
The data underlying this article will be shared by the corresponding author on reasonable request.
