Skip to main content
Journal of Medical Internet Research logoLink to Journal of Medical Internet Research
letter
. 2023 Aug 31;25:e50591. doi: 10.2196/50591

AI Increases the Pressure to Overhaul the Scientific Peer Review Process. Comment on “Artificial Intelligence Can Generate Fraudulent but Authentic-Looking Scientific Medical Articles: Pandora’s Box Has Been Opened”

Nicholas Liu 1,, Amy Brown 2
Editor: Tiffany Leung
PMCID: PMC10502600  PMID: 37651167

Májovský and colleagues’ [1] concern regarding OpenAI’s ChatGPT is valid. Seven months ago, the release of ChatGPT was quickly tempered by warnings of perpetuating biases and spreading misinformation. Artificial intelligence (AI) tools threaten to amplify preexisting issues in academic publishing, particularly the scientific peer review process. This outdated system is overwhelmed by the volume of new journals and papers produced by a growing global academic community [2], a problem that AI is fit to accentuate. Here are three areas in need of modification:

The scientific peer review process

  • The body of qualified reviewers is drowning in a rising sea of writers that includes top researchers, undergraduate students, and academics at all levels in between [2].

  • Peer review lacks formal standards or guidelines, as well as training, particularly in statistics [3], creating a restricted and top-heavy pool of qualified reviewers [2].

  • Reviewers are declining to perform reviews more often since the notion of reviewing as a professional obligation fails to sufficiently recognize or reward the burden it imposes [2].

  • Peer review fraud, which involves conflicts of interest, influence, and false identities, has evolved out of a need for reviews.

Publication pressure

  • The dearth of reviewers is compounded by the proliferation of plagiarized, fraudulent, and otherwise low-quality work [4].

  • Pressure to “publish or perish” has led to high-profile cases of academic fraud and likewise feeds “paper mills” that churn out questionable research for academics who are desperate to progress in their careers [4].

  • The proliferation of “for-profit” journals subverts respectful publishing through financialization that exploits and alienates scientists [2].

AI integration

  • Májovský et al [1] displayed the effectiveness of ChatGPT as an open access ghostwriter, capable of fabricating a complete and convincing article in just one hour.

  • In one study, only 63% of ChatGPT-generated abstracts were caught by reviewers as fakes [5]. In response to such findings, Science is updating its license and editorial policies to prohibit AI-generated text, figures, or graphics [5].

  • Such staunch resistance is misguided; AI may not be an author per se, but its utility in all stages of research, from generating topics and compiling information to writing text, cannot be ignored. If an AI-generated, human-reviewed paper communicates quality research, why should it be disallowed? Moreover, how would we tell?

  • Although AI-generated text detection software can help [1], detection bypass tools are similarly available online.

AI makes the need for high-quality peer reviews greater and more pressing than ever before. The cornerstone of scientific integrity is on the path to obsoletion without a viable successor. As academic pursuits become increasingly inseparable from industry, conceptualizing peer review as a duty to science will no longer suffice. Respecting and empowering the peer review system will involve considering reviewers as expert consultants, performing reviews as productive work, and creating system-wide guidelines that integrate (rather than resist) AI technologies. This problem, emerging from an imperative for success, needs a peer review system and publication process that has more teeth than trust, a commodity that served us well in the past, but whose restoration bears reinvention.

Abbreviations

AI

artificial intelligence

Footnotes

Conflicts of Interest: None declared.

References

  • 1.Májovský M, Černý M, Kasal M, Komarc M, Netuka D. Artificial intelligence can generate fraudulent but authentic-looking scientific medical articles: Pandora’s box has been opened. J Med Internet Res. 2023 May 31;25:e46924. doi: 10.2196/46924. https://www.jmir.org/2023//e46924/ v25i1e46924 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 2.Dance A. Stop the peer-review treadmill. I want to get off. Nature. 2023 Feb 13;614(7948):581–583. doi: 10.1038/d41586-023-00403-8.10.1038/d41586-023-00403-8 [DOI] [PubMed] [Google Scholar]
  • 3.Ballester PL. Open science and software assistance: commentary on “Artificial intelligence can generate fraudulent but authentic-looking scientific medical articles: Pandora’s box has been opened”. J Med Internet Res. 2023 May 31;25:e49323. doi: 10.2196/49323. https://www.jmir.org/2023//e49323/ v25i1e49323 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4.Else H, Van Noorden R. The fight against fake-paper factories that churn out sham science. Nature. 2021 Mar 23;591(7851):516–519. doi: 10.1038/d41586-021-00733-5.10.1038/d41586-021-00733-5 [DOI] [PubMed] [Google Scholar]
  • 5.Thorp HH. ChatGPT is fun, but not an author. Science. 2023 Jan 27;379(6630):313–313. doi: 10.1126/science.adg7879. [DOI] [PubMed] [Google Scholar]

Articles from Journal of Medical Internet Research are provided here courtesy of JMIR Publications Inc.

RESOURCES