Skip to main content
Iranian Journal of Basic Medical Sciences logoLink to Iranian Journal of Basic Medical Sciences
editorial
. 2026;29(1):1–2. doi: 10.22038/ijbms.2026.27130

Artificial intelligence (AI) in academic publishing: Legitimate use, plagiarism detection, and ethical challenges

Leila Arabi 1, Ali Roohbakhsh 2, Bizhan Malaekeh-Nikouei 1, Bibi Sedigheh Fazly Bazzaz 3,*
PMCID: PMC12867115  PMID: 41641141

Abstract

Objective(s):

This study aimed to determine the effect of 8-week high-intensity interval training (HIIT) on oxidative stress and apoptosis in the hippocampus of male rats with type 2 diabetes (T2D). The study focused on examining the role of proliferator-activated receptor gamma co-activator 1α (PGC1α)/Kelch-like ECH-associated protein Keap1/nuclear factor erythroid 2-related factor 2 (Nrf2) signaling pathway.

Materials and Methods:

Twenty-eight 8-week-old Wistar rats were randomly assigned to one of four groups (n=7): control (Con), type 2 diabetes (T2D), exercise (Ex), and exercise + type 2 diabetes (Ex+T2D). The Ex and Ex+T2D groups completed an 8-week exercise program consisting of 80-100% Vmax and 4–10 intervals. The homeostasis model assessment of insulin resistance (HOMA-IR) index was used to assess insulin resistance. The levels of Bcl2, BAX, musculoaponeurotic fibrosarcoma (Maf), Nrf2, Keap1, and PGC1α in the hippocampus were assessed using the western blot method. Additionally, the levels of antioxidant enzymes in the hippocampus were measured using ELISA.

Results:

The findings indicated that the T2D group had lower levels of antioxidant enzymes, Maf, Bcl2, PGC1α, and Nrf2, and higher levels of BAX and Keap1 in the hippocampus. Conversely, the HIIT group exhibited increased levels of antioxidant enzymes, Maf, Bcl2, Nrf2, and PGC1α, along with decreased levels of BAX and Keap1 in the hippocampus.

Conclusion:

The study demonstrated that 8-week HIIT was effective in reducing hippocampal apoptosis and oxidative stress induced by T2D by activating the PGC1α-Keap1-Nrf2 signaling pathway. The metabolic changes induced by exercise may lead to an increase in PGC1 expression, which is the primary stimulator of the Keap1-Nrf2 signaling pathway.

Key Words: Anti-oxidant enzymes, Apoptosis, Hippocampus, PGC1α, Type 2 diabetes

Introduction

The rapid evolution of artificial intelligence (AI) continues to reshape the landscape of academic publishing. As machine learning and large language models (LLMs) become more sophisticated, their integration into scientific writing, peer review, and editorial processes promises to improve efficiency, accessibility, and research integrity. However, it also raises critical questions about authorship, authenticity, and ethical responsibility.

In this editorial of Iranian Journal of Basic Medical Sciences (IJBMS), we explore how AI can be legitimately used by authors and reviewers, how it assists in detecting plagiarism and data fabrication, and how emerging patterns of AI use among Generation Z researchers require new ethical awareness.

Legitimate use of AI by authors

AI tools have become valuable aids for researchers seeking clarity, conciseness, and coherence in their manuscripts. Writing assistants such as ChatGPT or Grammarly can help improve grammar, style, and fluency, particularly for non-native English speakers. AI-based summarization tools can also extract large bodies of literature, helping authors identify relevant research gaps and trends (1). However, legitimate use requires transparency and accountability. According to the Committee on Publication Ethics (COPE) and several leading publishers, authors must disclose the use of AI tools and ensure that all intellectual contributions remain human. Importantly, AI-generated text should never be used without verification. As COPE emphasizes, authors are fully responsible for checking all factual accuracy, references, and potential bias introduced by automated systems (2). AI can support expression but not idea creation. Scientific reasoning, data interpretation, and hypothesis formulation must originate from human authors.

AI in peer review: Potential and caution

Peer review is the foundation of academic quality control, yet it remains time-consuming, subjective and often inconsistent. AI offers promising support tools that can assist in pre-screening submissions for clarity, structure, or methodological completeness. Additionally, AI can help reviewers by suggesting relevant literature, verifying citation accuracy, or summarizing sections for faster evaluation. Nevertheless, AI should improve human reviewers. However, relying too much on algorithms may reinforce existing biases, particularly if the models were trained on unbalanced or outdated data (3). In addition, reviewers are advised against uploading any portion of a submitted manuscript into generative AI tools, as this action may compromise the authors’ confidentiality and proprietary rights.

Therefore, journals should encourage a hybrid model of peer review, where AI performs mechanical checks, and human reviewers focus on conceptual depth and scientific integrity. Reviewers and editors must use AI tools only within secure, confidential frameworks and are prohibited from uploading manuscripts to external platforms that may compromise privacy or intellectual property.

AI for plagiarism, data fabrication, and image manipulation detection

Perhaps the most impactful contribution of AI in academic publishing is in research integrity verification. Traditional plagiarism detection software is evolving into more advanced, AI-driven systems capable of identifying semantic or conceptual plagiarism (3).

AI also plays an increasing role in detecting data fabrication and image manipulation. Deep-learning models can analyze image metadata, detect duplicated microscopy figures, or notice unnatural statistical patterns. These tools have proven invaluable in large-scale editorial screening and even in post-publication investigations (4).

However, editors and reviewers must remember that AI tools provide signals, not a final judgement. Human judgment remains essential in determining whether an identified similarity or irregularity constitutes ethical misconduct. AI should therefore be seen as a guardian of integrity, supporting editors but never replacing editorial expertise and ethical judgment.

Generation Z: New challenges and responsibilities

Generation Z researchers bring both innovation and risk to scholarly communication. Their comfort with AI-based writing and image-generation tools can accelerate research dissemination and creativity. However, it also introduces new ethical dilemmas.

Many young scholars may unintentionally cross ethical boundaries by relying too much on generative models for drafting, paraphrasing, or even producing figures without appropriate acknowledgment. In addition, some AI models fabricate fake references and citations and researchers should pay attention to this important issue. Therefore, journals and academic institutions must educate emerging researchers on the ethical boundaries of AI use. Workshops, author guidelines, and editorial statements could clearly define what constitutes legitimate versus unethical AI assistance.

The IJBMS perspective and policy

As we mentioned in our previous editorial, IJBMS policy aligns with COPE’s position statement on the use of AI in manuscript preparation (5). We emphasized that authors are required to declare any AI tool used during writing, data analysis, or figure generation.

AI can help authors refine language, assist reviewers in quality assurance, and enable editors to detect misconduct more effectively than ever before. We hope that employment of the AI-powered plagiarism and image-detection tools could uphold research integrity. We believe that when AI used responsibly, it enhances clarity, fairness, and precision in research dissemination. However, these benefits come with ethical boundaries. Transparency, disclosure, and accountability must guide every use of AI. In fact, Human creativity, critical thinking, and moral responsibility remain irreplaceable principles in science.

We wish our authors, reviewers, and readers a year of progress guided by both human and artificial intelligence, working together for the advancement of science.

References

  • 1.Gao C, Howard F, Markov N, Dyer E, Ramesh S, Luo Y, Pearson A. Comparing scientific abstracts generated by ChatGPT to real abstracts with detectors and blinded human reviewers. NPJ Digit Med. 2023;6:75–79. doi: 10.1038/s41746-023-00819-6. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 2.Arabi L, Roohbakhsh A, Malaekeh-Nikouei B, Fazly Bazzaz BS. Joining COPE: Opportunities and benefits for the Iranian Journal of Basic Medical Sciences. Iran J Basic Med Sci. 2024;27:1–2. doi: 10.22038/ijbms.2023.76381.16530. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3.Giray L. Benefits and challenges of using AI for peer review: A study on researchers’ perceptions. Ser Libr. 2024;85:144–154. [Google Scholar]
  • 4.Pellegrina D, Helmy M. AI for scientific integrity: Detecting ethical breaches, errors, and misconduct in manuscripts. Front Artif Intell. 2025;8:1644098. doi: 10.3389/frai.2025.1644098. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5.Sadhin I, Hassan T, Nayim A. Plagiarism detection using artificial intelligence. Int J Comput Inf Syst. 2023;5:102–108. [Google Scholar]
  • 6.Arabi L, Roohbakhsh A, Malaekeh-Nikouei B, Fazly Bazzaz BS. The impact of artificial intelligence (AI) in academic writing and publication: Iranian Journal of Basic Medical Sciences (IJBMS) policy. Iran J Basic Med Sci. 2025;28:1–2. doi: 10.22038/ijbms.2025.25229. [DOI] [PMC free article] [PubMed] [Google Scholar]

Articles from Iranian Journal of Basic Medical Sciences are provided here courtesy of Mashhad University of Medical Sciences

RESOURCES