Recent advances in artificial intelligence (AI) and related technologies now surpass human capabilities in areas once thought to be uniquely human. AI has already outdone humans in complex reasoning tasks like chess and Go1,2). It has been reported that some AIs have qualified for examinations for highly professional jobs3-5). More recently, AI applications have moved beyond superficial judgments and deep learning, expanding into previously human-only domains of art and creativity.
There is no doubt regarding the expectation that scientific progress will eventually benefit humans in everyday life. However, serious concerns also exist about the easy accessibility to AI. One of the biggest concerns could be writing a scientific article. Publishing a paper in a peer-reviewed journal requires originality, creativity, logical thinking, and vital ethical considerations. It seems, however, that these criteria are no longer of the utmost importance. Generally called LLM (large language models), the current version of generative AIs can generate texts based on user prompts through the ‘natural language processing’ of ‘machine-learned’ knowledge6). This helpful technology is just one click away from every author. Ease of accessibility also escalates concerns surrounding potential misuse. Suppose that a research group with little prior publication history suddenly submits multiple papers in a very short period. This sudden productivity may result from their extensive hard work. However, from the current perspective of an editor’s office, we must first examine whether the AI has been maliciously used. As several scientific journals have recently shared such experiences, a new set of contemporary concerns surrounding the use of AI have arisen.
Currently, most generative AI can develop research ideas from a simple user prompt, provide a theory from the concept, and write comprehensive papers with full references. At the very least, AI can provide crucial assistance in nearly all aspects of preparing a research paper. For those involved in scientific publishing, this reality poses several significant dilemmas. First, a generative AI may produce inaccurate or erroneous descriptions, known as “hallucinations.”7) While the providers of the most current AI version claim that this phenomenon is no longer an issue, authors should be aware of the possibility. Without meticulous checks and verification by the authors, these “hallucinations” may slip through the publication process and mislead readers with incorrect information. As AI continues to improve at simulating human language, it becomes increasingly difficult to distinguish between fact and fiction and between AI and human writing8). It raises fears that even initially flawed text could become indistinguishable from original work. Another concern is the ambiguity of applying the strict ethical standards expected of human researchers, authors, and publishers to AI tools. To illustrate this point, consider a scenario where a researcher, maliciously or inadvertently, uploads a part of their research data to an AI-powered company or program. Following the researcher’s intentions, the AI could fabricate parts of the material and create a draft manuscript. After some time, a co-worker, not recognizing where they originated, might write a scientific paper with the assistance of another AI using this dataset and the draft. This process could then lead to the fraudulent creation of a scientific article without extensive and arduous research. Worse still, if malicious intent was involved, recently popularized deepfake technology could even be employed to create or modify graphs and images in research papers9,10). The final concern would be the ambiguity of the definition of plagiarism11). AI-generated text is likely to originate from the imitated content from its training database. Thus, its originality should be questioned if authors or reviewers fail to find a proper citation in every sentence. These issues compound editorial challenges. Even with specialized tools, the possibility of definitively identifying AI-generated content may diminish as technology advances. Consequently, researchers’ adherence to ethical guidelines and transparency in the research process becomes increasingly crucial. Due to these complexities, most editorial organizations hesitate to establish definitive regulations. Instead of strict rules, they state the individual stances or seek consensus.
There is widespread agreement that, rather than being a simple yes/no proposition, AI is currently used somewhere along the entire spectrum from idea generation to publication. If this is true, what is the maximum extent to which AI can be used and still define a manuscript as written by humans? How can we ascertain the originality, creativity, and integrity of scientific research?
Hip & Pelvis acknowledge that the era of AI in scientific writing has already arrived and urge all researchers to fulfill their corresponding ethical responsibilities. To make our faith manifest, we added AI-related items to the author checklist, effective August 2024. The editorial office always trusts authors’ declarations and respects their decisions. However, to accomplish our mission, we will use all available methods to verify that generative AI has not been excessively used in creating a manuscript, image, or graph. If discrepancies arise between an author’s declarations and our determinations, every effort will be made to clarify the situation. Also, the authors should take full responsibility for their statements. The current stance about the AI of the Editorial Board of Hip & Pelvis is as follows: AI cannot be recognized as an author under any circumstances. It should be used as a tool for refining papers published in Hip & Pelvis, not a magic wand for creating data or documents and realizing ideas.
Hip & Pelvis has undergone a long and winding journey since 1989. The perseverance of many researchers and former editorial members has compelled us to move forward against unexpected challenges. I believe that this journal is now on its third starting line. The first was the inauguration, and the second was the decision to expand our publication scope from domestic to international research. Now, the journal is on the line of embracing AI. Definitively, we are entering a new ocean with a new companion, not knowing what will happen after the gate closes behind us. So before you set sail, I ask you this: Who will hold the rudder of your boat, you or the AI?
Funding Statement
Funding No funding to declare.
Footnotes
Conflict of Interest
Kee Hyung Rhyu has been an Editor-in-Chief since January 2023, but had no role in the decision to publish this article. No other potential conflict of interest relevant to this article was reported.
References
- 1.Silver D, Schrittwieser J, Simonyan K, et al. Mastering the game of Go without human knowledge. Nature. 2017;550:354–9. doi: 10.1038/nature24270. https://doi.org/10.1038/nature24270. [DOI] [PubMed] [Google Scholar]
- 2.Schrittwieser J, Antonoglou I, Hubert T, et al. Mastering Atari, Go, chess and shogi by planning with a learned model. Nature. 2020;588:604–9. doi: 10.1038/s41586-020-03051-4. https://doi.org/10.1038/s41586-020-03051-4. [DOI] [PubMed] [Google Scholar]
- 3.Lin SY, Hsu YY, Ju SW, Yeh PC, Hsu WH, Kao CH. Assessing AI efficacy in medical knowledge tests: a study using Taiwan's internal medicine exam questions from 2020 to 2023. Digit Health. 2024;10:20552076241291404. doi: 10.1177/20552076241291404. https://doi.org/10.1177/20552076241291404. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 4.Meyer A, Riese J, Streichert T. Comparison of the performance of GPT-3.5 and GPT-4 with that of medical students on the written German medical licensing examination: observational study. JMIR Med Educ. 2024;10:e50965. doi: 10.2196/50965. https://doi.org/10.2196/50965. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 5.Kelly SM. ChatGPT passes exams from law and business schools [Internet] CNN Business; 2023. Jan 26, [cited 2024 Nov 3]. Available from: https://edition.cnn.com/2023/01/26/tech/chatgpt-passes-exams/index.html . [Google Scholar]
- 6.Dergaa I, Chamari K, Zmijewski P, Ben Saad H. From human writing to artificial intelligence generated text: examining the prospects and potential threats of ChatGPT in academic writing. Biol Sport. 2023;40:615–22. doi: 10.5114/biolsport.2023.125623. https://doi.org/10.5114/biolsport.2023.125623. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 7.Alkaissi H, McFarlane SI. Artificial hallucinations in ChatGPT: implications in scientific writing. Cureus. 2023;15:e35179. doi: 10.7759/cureus.35179. https://doi.org/10.7759/cureus.35179. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8.Ozkara BB, Boutet A, Comstock BA, et al. Artificial intelligence-generated editorials in radiology: can expert editors detect them? AJNR Am J Neuroradiol. doi: 10.3174/ajnr.A8505. Published online September 17, 2024; https://doi.org/10.3174/ajnr.a8505 . [DOI] [PMC free article] [PubMed] [Google Scholar]
- 9.Prezja F, Paloneva J, Pölönen I, Niinimäki E, Äyrämö S. DeepFake knee osteoarthritis X-rays from generative adversarial neural networks deceive medical experts and offer augmentation potential to automatic classification. Sci Rep. 2022;12:18573. doi: 10.1038/s41598-022-23081-4. https://doi.org/10.1038/s41598-022-23081-4. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 10.Gu J, Wang X, Li C, et al. AI-enabled image fraud in scientific publications. Patterns (N Y) 2022;3:100511. doi: 10.1016/j.patter.2022.100511. https://doi.org/10.1016/j.patter.2022.100511. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 11.Anders BA. Is using ChatGPT cheating, plagiarism, both, neither, or forward thinking? Patterns (N Y) 2023;4:100694. doi: 10.1016/j.patter.2023.100694. https://doi.org/10.1016/j.patter.2023.100694. [DOI] [PMC free article] [PubMed] [Google Scholar]
