Abstract
Artificial intelligence (AI) text detection tools are considered a means of preserving the integrity of scholarly publication by identifying whether a text is written by humans or generated by AI. This study evaluates three popular tools (GPTZero, ZeroGPT, and DetectGPT) through two experiments: first, distinguishing human-written abstracts from those generated by ChatGPT o1 and Gemini 2.0 Pro Experimental; second, evaluating AI-assisted abstracts where the original text has been enhanced by these large language models (LLMs) to improve readability. Results reveal notable trade-offs in accuracy and bias, disproportionately affecting non-native speakers and certain disciplines. This study highlights the limitations of detection-focused approaches and advocates a shift toward ethical, responsible, and transparent use of LLMs in scholarly publication.
Keywords: AI text detection tools, Large language models (LLMs), Non-native authors, ChatGPT, Gemini, GPTZero, ZeroGPT, DetectGPT, Fairness in scholarly publication, Accuracy-bias trade-off
Introduction
In recent years, there has been an unprecedented rapid development in the field of artificial intelligence (AI), particularly with the surge in popularity of large language models (LLMs) across various domains. Shortly after OpenAI released ChatGPT in late 2022, several other LLMs became available to the general public, including Google’s Bard, which was rebranded as Gemini one year after its launch. Within a brief period, LLMs have become the next big thing, and their use has become widespread and mainstream, disrupting numerous aspects across multiple domains. LLMs have also significantly impacted academic and scholarly publication. Researchers increasingly use LLMs for information seeking, editing, ideation, framing, direct writing, data cleaning, data analysis, and even data and content generation (Liao et al., 2024).
This widespread adoption, however, raises some ethical concerns. In some cases, authors have been caught directly copying and pasting content generated by LLMs into publications without disclosure (Kendall & Teixeira da Silva, 2024), some of which have led to retractions (Lei et al., 2024). Other cases involve the controversial listing of LLMs as co-authors in scholarly works (Nazarovets & Teixeira da Silva, 2024; Stokel-Walker, 2023; Yeo-Teh & Tang, 2024), which has sparked debates regarding accountability and authorship. Eventually, this led to a consensus that LLMs cannot be listed as authors of scholarly articles, but their use as assistive tools to improve the readability of writing is acceptable (Lund & Naheem, 2024). Furthermore, some journals and publishers have modified their guidelines to address the use of generative AI and LLMs (Cheng & Wu, 2024; Ganjavi et al., 2024; Flanagin et al., 2023). However, not all researchers acknowledge LLM involvement in their scholarly articles (Pesante, Mauffrey & Parry, 2024). Some argue that it is unnecessary to disclose LLM involvement when it is solely used as a tool to improve readability, either by correcting spelling and grammar or other general editing (Chemaya & Martin, 2024), while others hesitate due to fear of stigma or misconceptions about their work’s originality and integrity (Giray, 2024a). This further complicates the ethical landscape of LLM adoption in academia and scholarly publication.
Historically, non-native English-speaking scholars have faced significant challenges in navigating the scientific communication realm, where English is the international language of scholarly publication (Amano et al., 2023; Cho, 2004; Ferguson, Pérez-Llantada & Plo, 2011; Flowerdew, 2001; Horn, 2017; Kojima & Popiel, 2022; Raitskaya & Tikhonova, 2020). Nearly all reputable journals require that manuscripts conform to elevated linguistic standards, often necessitating professional proofreading services that can be prohibitively expensive (Amano et al., 2023; Mumin, 2022; Ramírez-Castañeda, 2020; Van Noorden, 2013). This is where LLMs can play a key role by providing a transformative and unprecedented opportunity to equalize access, assisting researchers in improving the language, style, and clarity of their work (Liao et al., 2024). For these researchers, whose first language is not English, LLMs can offer a cost-effective alternative to traditional editing services, leveling the playing field and promoting fairness in scholarly publication.
Meanwhile, it is well-known that different disciplines have distinct writing styles when it comes to scholarly publication, reflecting the varied conventions, methodologies, and epistemologies unique to each field (Alluqmani & Shamir, 2018; Dong et al., 2024; Dong, Mao & Pei, 2023). This diversity, however, raises the possibility of bias in AI-generated text produced by LLMs, which may perform better in disciplines with simpler, more standardized language structures while struggling with the nuanced and often interpretive styles found in disciplines like the humanities, social sciences, and interdisciplinary journals. Furthermore, concerns have been raised about the potential for LLMs to drive both language use and knowledge construction toward homogeneity and uniformity (Kuteeva & Andersson, 2024). By reproducing existing patterns in training data, LLMs risk amplifying pre-existing biases and structural inequalities embedded within scholarly publication. This could inadvertently reinforce dominant paradigms while marginalizing alternative voices and perspectives, undermining the diversity that is critical to advancing knowledge across disciplines.
On the other hand, AI text detection tools have emerged as a means to address concerns about undisclosed use of AI, especially in academic settings (Perkins et al., 2024; Mitchell et al., 2023). These tools claim to be able to differentiate between human-written and AI-generated text with a very high accuracy, thereby ensuring the integrity of academic work. However, the use of these tools raised additional concerns, not only about their effectiveness (Elkhatat, Elsaid & Almeer, 2023; Weber-Wulff et al., 2023) but also their impact on equity (Giray, 2024b; Liang et al., 2023). AI text detection tools differ in their design, precision, and underlying approaches. Similar to LLMs, these AI text detection tools operate as black box systems, lacking transparent explanations for their classifications of text as human-written or AI-generated, as they rely solely on pattern recognition. Moreover, even if these tools demonstrate high accuracy in detecting AI-generated text, none of them claim to be free of bias against certain groups of authors or writing styles.
Things become even more complicated since the involvement of LLMs in writing is not as straightforward as when the text is either purely human-written or entirely AI-generated. Increasingly, texts are hybrid or a mix of the two. For instance, humans may manually edit AI-generated text to varying degrees. Conversely, some researchers may write the first draft themselves and then run it through LLMs to improve readability. It is also possible for texts to undergo multiple iterations of AI and human editing, regardless of who authored the original draft. In these more nuanced cases, the final results can vary significantly, depending on several factors, including the quality of the original draft, the LLMs used, the prompt given to improve the text, and the extent of manual editing involved. This kind of hybrid text, also referred to as AI-assisted text, may cause detection tools to become less accurate, more biased, or both. This makes it more difficult to see the practical application of these tools in real-world scenarios without significant compromise.
To position this work within existing literature, previous studies have examined the performance of AI text detection tools primarily by comparing human-written and AI-generated texts, though most have not specifically addressed fairness or potential biases related to author background, such as native vs. non-native English speakers (Perkins et al., 2024; Weber-Wulff et al., 2023). Among the studies that do consider fairness, some are largely theoretical (Giray, 2024a, 2024b), while another provides empirical evidence but is limited only to the comparison between human and AI-generated texts (Liang et al., 2023). Notably, none of them have yet empirically investigated fairness in detecting AI-assisted texts, a nuanced category situated between purely human-written and purely AI-generated content. This study addresses these gaps by first analyzing human vs. AI-generated text detection before extending the evaluation to include AI-assisted texts, with explicit consideration of potential biases against non-native authors as well as differences across academic disciplines. Additionally, this study employs the latest and most advanced large language models at the time (i.e., ChatGPT o1 and Gemini 2.0 Pro Experimental) to ensure the findings reflect state-of-the-art capabilities and contemporary relevance.
Given the gaps identified above, this study aims to answer the following research questions:
- 
RQ1.How accurate are AI text detection tools in identifying human-written, AI-generated, and AI-assisted texts in scholarly articles? 
- 
RQ2.Is there any accuracy-bias trade-off in AI text detection tools? 
- 
RQ3.Do certain groups of researchers face disadvantages when their work is evaluated by AI text detection tools? 
By providing empirical evidence to answer these three questions, this study sheds light on the use and limitations of AI text detection tools, as well as their impact on fairness in scholarly communication. In doing so, it contributes to a more nuanced understanding of both the opportunities and challenges presented by LLM involvement in scholarly publication.
The remainder of this article is organized as follows: the Materials & Methods section describes the experimental design, dataset selection, and methodologies used to evaluate AI text detection tools. The Results section presents findings from the two experimental scenarios. The Discussion section explores the accuracy-bias trade-offs, implications for fairness, practical recommendations, as well as study limitations and directions for future research. Finally, the Conclusion section summarizes the key insights from this study.
Materials and Methods
Research design
This study is designed to evaluate the performance of AI text detection tools when applied to texts from scholarly journal articles. To provide a comprehensive analysis and to ensure a logical progression of findings, they will be examined for accuracy and fairness under two different experimental scenarios:
- 
(1)Differentiating between original human-written texts and AI-generated texts, and 
- 
(2)Evaluating AI-assisted text, where the original human-written texts were run through LLMs to improve readability. 
Considering factors such as length, accessibility, and content standardization, only the abstract from each scholarly article was used for evaluation.
Dataset selection
The dataset for this research is a compilation of abstracts from peer-reviewed journal articles published by 2021, which is at least one year before the release of ChatGPT and other publicly available LLMs and their widespread use by researchers. To ensure balance in representation, the dataset was stratified based on disciplines (i.e., technology & engineering, social sciences, and interdisciplinary) and authorship (i.e., native and non-native English). Each discipline is represented by three journals: ACM Computing Surveys, IEEE Access, and PeerJ Computer Science for technology & engineering; Sociology, International Sociology, and SAGE Open for social sciences; and British Journal of Educational Technology, Computers & Education, and Education and Information Technologies for interdisciplinary.
Authors affiliated with institutions in the five Anglosphere countries (i.e., Australia, Canada, New Zealand, United Kingdom, and United States) were chosen to represent the native English speakers. Meanwhile, to represent the non-native English speakers, only authors affiliated with institutions from countries where English is neither an official language nor spoken by more than half of the population were included. Furthermore, to avoid confounding variables, any articles with multiple authors from both native and non-native categories were excluded. The final dataset, as summarized in Table 1, comprises 72 articles with an equal distribution across disciplines and author categories.
Table 1. Summary of dataset by discipline, journal, and country of affiliation.
AI-generated and AI-assisted texts
For each original abstract in the dataset, its AI-generated version was produced using two of the most widely used LLMs available to the general public: OpenAI’s ChatGPT and Google’s Gemini. More specifically, only the most advanced versions available at the time of the experiment in late December 2024 were used: ChatGPT o1 and Gemini 2.0 Pro Experimental (Preview gemini-exp-1206). Note that for these AI-generated abstracts, only the article title and the journal name were provided to the LLMs as input; nothing from the original abstract was included. Subsequently, in a separate chat session, an AI-assisted abstract for each article was produced. To ensure consistency, a standardized prompt, as summarized in Table 2, was used to generate both AI-generated and AI-assisted abstracts, once for each journal.
Table 2. Standardized prompts used to produce AI-generated and AI-assisted abstracts.
| Type | Prompt used | 
|---|---|
| AI-Generated | Generate an abstract for each of these journal articles prepared for submission to the scientific journal “journal name”: 
 | 
| AI-Assisted | Below are several research articles in preparation for submission to the “journal name” journal. Enhance the clarity and readability of the abstract but keep the title unchanged: 
 | 
This approach simulates a realistic and legitimate scenario where researchers use LLMs to refine their original writing, thus preserving the substance of the content while improving its readability, rather than generating content from scratch. Table 3 provides an example of these abstracts in their original, AI-generated, and AI-assisted formats. These variations of abstracts represent the different types of text to be evaluated by AI text detection tools in the next step of the experiment. The complete dataset, along with the experimental results and Python code used to analyze and visualize the findings, is publicly available for download at the author’s GitHub repository (https://github.com/ahmadrpratama/ai-text-detection-bias).
Table 3. Examples of the original, AI-generated, and AI-assisted abstracts from the same article.
| Type | Text | 
|---|---|
| Original Abstract (Hwang, Chien & Li, 2021) | Digital game-based learning (DGBL) is a frequently adopted mode for facilitating teaching and learning, where learners are immersed in game contexts to engage in gameplay and to construct meaningful knowledge. However, without guidance, students might have difficulties organizing what they have experienced in gaming contexts. Thus, in this research, a multidimensional repertory grid (MDRG) approach was proposed, and a digital game was implemented accordingly. Moreover, an experiment was executed in a geography course to examine the effectiveness of the strategy. A total of 83 high school students (15.5 years old on average) participated in the experiment. The experimental group learned with the multidimensional repertory grid-based educational game, while the control group learned with the conventional gaming mode. The analysis results revealed that the experimental group had better learning achievement as well as higher learning motivation, self-efficacy and metacognition awareness. In addition, the behavioral analysis and interview results revealed that those learning with the proposed strategy had a greater tendency to promote higher order thinking. Accordingly, the limitations and suggestions for future research are provided. | 
| AI-Generated Chat GPT o1 | This study explores the use of a multidimensional repertory grid as a graphic organizer to systematically integrate digital games into classroom instruction, with the aim of enhancing students’ learning performances and behaviors. The repertory grid framework aligns game elements, pedagogical strategies, and learning objectives, enabling teachers to develop coherent, learner-centered game-based activities. Employing a mixed-methods approach that includes classroom observations, student feedback, and performance assessments, the research examines the extent to which this visual tool supports effective instructional design. Findings show that the repertory grid can improve knowledge acquisition, foster positive learning attitudes, and sustain student motivation. Students engaged more deeply with content, demonstrated improved problem-solving skills, and exhibited greater collaboration. The organizer also guided teachers in selecting suitable games, scaffolding learning tasks, and making informed adjustments to instruction over time. These results highlight the potential of a multidimensional repertory grid to serve as a valuable resource for educators seeking to optimize the educational benefits of digital games, ultimately shaping more dynamic, impactful, and student-centered learning experiences. | 
| AI-Generated Gemini 2.0 Pro Experimental | This study explores the efficacy of a multidimensional repertory grid (MDRG) as a graphic organizer in facilitating the implementation of digital games for enhanced student learning. MDRGs, based on Personal Construct Theory, enable the elicitation and representation of individuals’ perceptions and understandings. We argue that MDRGs can help educators navigate the complexities of game-based learning by identifying key constructs related to learning outcomes and student behaviors. This research employs a mixed-methods approach, combining quantitative analysis of student performance data with qualitative analysis of MDRG constructs generated by both teachers and students. Findings indicate that using MDRGs as a planning and reflection tool promotes more intentional and targeted game integration. Furthermore, the process facilitated a shared understanding between teachers and students about the learning objectives and expected behaviors within the game-based environment. | 
| AI-Assisted Chat GPT o1 | Digital game-based learning (DGBL) is widely adopted to facilitate teaching and learning by immersing students in game environments, allowing them to engage in gameplay and construct meaningful knowledge. However, without proper guidance, students may find it challenging to organize what they have learned in such contexts. To address this issue, this study proposes a multidimensional repertory grid (MDRG) approach and implements it within a digital game. An experiment was conducted in a geography course with 83 high school students (mean age = 15.5 years). The experimental group learned using the MDRG-based educational game, while the control group followed a conventional game-based approach. The results indicate that students in the experimental group outperformed their counterparts in learning achievement, motivation, self-efficacy and metacognitive awareness. Behavioral observations and interview data further reveal that students using the MDRG approach engaged in higher order thinking to a greater extent. Limitations and suggestions for future studies are also discussed. | 
| AI-Assisted Gemini 2.0 Pro Experimental | Digital game-based learning (DGBL) offers an engaging approach to education, immersing learners in game contexts to foster meaningful knowledge construction. However, students may struggle to organize their experiences without adequate guidance. This research proposes a multidimensional repertory grid (MDRG) approach, integrated into a digital game, to address this challenge. An experiment was conducted with 83 high school students (average age 15.5 years) in a geography course to evaluate the effectiveness of this strategy. The experimental group used the MDRG-based educational game, while the control group engaged with a conventional gaming mode. Results showed that the experimental group achieved significantly better learning outcomes, alongside higher levels of learning motivation, self-efficacy, and metacognitive awareness. Furthermore, behavioral analysis and interviews revealed that students using the MDRG strategy exhibited a greater tendency towards higher-order thinking. The article concludes with limitations and suggestions for future research in this area. | 
AI text detection tools
In this study, three AI text detection tools that rank among the top results in Google search queries for “AI text detection tool” were included in the experiment: GPTZero, ZeroGPT, and DetectGPT. All three tools are freely available for public use, with optional premium subscriptions that unlock additional features such as longer word detection limits and detailed report generation. Each tool provides user with both the qualitative results (i.e., categorical labels, such as “human,” “AI,” “mixed,” or “uncertain”) and quantitative results (i.e., percentage scores representing the likelihood that a text is AI-generated), as compiled in Figs. 1–3.
Figure 1. Examples of AI text detection results by GPTZero.
Figure 3. Examples of AI text detection results by DetectGPT.
Figure 2. Examples of AI text detection results by ZeroGPT.
Evaluation metrics
In the first scenario, the task was to evaluate the tools’ ability to distinguish between original human-written and AI-generated abstracts. Ground truth is clearly established in this scenario, as the origin of each text (human or AI-generated) is explicitly known. This allows for precise evaluation of the tools’ performance using the following metrics:
- 
1.Accuracy: The percentage of correctly classified abstracts. 
- 
2.False positive rate (FPR): The percentage of original abstracts where the tool failed to classify as human written. 
- 
3.False negative rate (FNR): The percentage of AI-generated abstracts that the tool failed to classify as AI-generated. 
In addition, two complementary metrics were used to capture broader patterns of false positives across the dataset, in which original abstracts were misclassified as AI-generated. These two metrics offer a better picture of how such tools could potentially harm authors:
- 
4.False accusation rate (FAR): The percentage of original abstracts with at least one false positive result. 
- 
5.Majority false accusation rate (MFAR): The percentage of original abstracts with more false positive results than true negative results. 
The second scenario is more nuanced, as the task involved evaluating AI-assisted abstracts, which are human-authored texts enhanced by LLMs. Unlike the first scenario, determining ground truth for AI-assisted texts is much more complex because these texts blend both human and AI contributions. In this case, the focus shifted to quantitative analysis, evaluating the percentage scores provided by the tools. The analysis involved:
- 
6.Summary statistics of scores, which include min, max, quartiles, median, mean, and standard deviation. 
- 
7.Statistical tests to identify significant differences in detection scores across disciplines (i.e., technology & engineering, social sciences, interdisciplinary), author categories (i.e., native, non-native), and LLMs (i.e., ChatGPT o1, Gemini 2.0 Pro Experimental). 
Significant differences in scores would indicate potential biases against certain groups of authors, such as non-native speakers or researchers in specific disciplines. The purpose of this scenario was to highlight the challenge of accurately evaluating partially AI-generated content, where no clear-cut ground truth exists.
Nevertheless, considering that in many cases, these AI text detection tools do make extreme classifications (i.e., 0% or 100%), even for AI-assisted texts, their performance was further evaluated with two additional metrics:
- 
8.Under-Detection Rate (UDR): The percentage of AI-assisted abstracts labeled as 0% AI by the tool, indicating a failure to detect any AI contribution. 
- 
9.Over-Detection Rate (ODR): The percentage of AI-assisted abstracts labeled as 100% AI by the tool, incorrectly attributing the entire text to AI and disregarding human contribution. 
Results
Scenario 1: Original vs. AI-generated abstracts
The performance of all three AI text detection tools in classifying abstracts, where the ground truth is clearly established as either human or AI-generated, is summarized in Table 4. The five metrics used in this scenario were analyzed across author categories and disciplines to identify patterns of reliability and potential bias.
Table 4. Overall performance metrics from Scenario 1: Original vs. AI-generated abstracts.
| Metric | GPTZero (%) | ZeroGPT (%) | DetectGPT (%) | 
|---|---|---|---|
| Accuracy | 97.22 | 64.35 | 54.63 | 
| - By author categories | |||
| - Native | 99.07 | 64.81 | 58.33 | 
| - Non-Native | 97.22 | 63.89 | 50.93 | 
| - By disciplines | |||
| - Technology & Engineering | 100.00 | 61.11 | 56.94 | 
| - Social Sciences | 98.61 | 66.67 | 63.89 | 
| - Interdisciplinary | 95.83 | 65.28 | 43.06 | 
| False Positive Rate (FPR) | 0.00 | 16.67 | 31.94 | 
| - By author categories | |||
| - Native | 0.00 | 19.44 | 27.78 | 
| - Non-Native | 0.00 | 13.89 | 36.11 | 
| - By disciplines | |||
| - Technology & Engineering | 0.00 | 12.50 | 41.67 | 
| - Social Sciences | 0.00 | 12.50 | 12.50 | 
| - Interdisciplinary | 0.00 | 25.00 | 41.67 | 
| False Negative Rate (FNR) | 2.78 | 45.15 | 52.08 | 
| - By author categories | |||
| - Native | 2.78 | 43.06 | 48.61 | 
| - Non-Native | 8.33 | 47.22 | 55.56 | 
| - By disciplines | |||
| - Technology & Engineering | 2.08 | 52.08 | 43.75 | 
| - Social Sciences | 4.17 | 43.75 | 47.92 | 
| - Interdisciplinary | 2.08 | 39.58 | 64.58 | 
| False Accusation Rate (FAR) | 44.44 | ||
| - By author categories | |||
| - Native | 44.44 | ||
| - Non-Native | 44.44 | ||
| - By disciplines | |||
| - Technology & Engineering | 45.83 | ||
| - Social Sciences | 25.00 | ||
| - Interdisciplinary | 62.50 | ||
| Majority False Accusation Rate (MFAR) | 4.17 | ||
| - By author categories | |||
| - Native | 2.78 | ||
| - Non-Native | 5.56 | ||
| - By disciplines | |||
| - Technology & Engineering | 8.33 | ||
| - Social Sciences | 0.00 | ||
| - Interdisciplinary | 4.17 | 
As shown in Table 4, out of the three AI text detection tools in this study, GPTZero achieved the highest accuracy at 97.22% with a 0% FPR and a very low FNR at 2.78%. These numbers suggest strong reliability of GPTZero in differentiating between human-written and AI-generated texts in this first scenario, where the ground truth is clearly established. ZeroGPT performed much worse, with an overall accuracy of just 64.35%, a relatively high FPR at 16.67%, and even higher FNR at 45.14%. DetectGPT, on the other hand, despite claiming a 99% accuracy rate, performed the worst in practice, achieving merely 54.63% accuracy. This makes it virtually no better than random guessing. It also exhibited much higher FPR and FNR values at 31.94% and 52.08%, respectively.
In addition to the three tool-specific metrics above, the other two dataset-wide metrics (i.e., FAR and MFAR) further highlight the broader risks of misclassification that could potentially harm researchers by falsely accusing them of using AI for their original writings. The FAR of 44.44% means nearly half of the original abstracts in the dataset were misclassified as AI-generated by at least one AI text detection tool, which is alarming. The MFAR, while notably much lower at 4.17%, still indicates the potential risk of consensus false accusations by multiple AI text detection tools, albeit less frequently.
When examining variability across author categories and disciplines, slight differences in accuracy were observed. All three tools performed slightly more accurately for native authors compared to non-native authors. However, performance consistency for FPR and FNR across tools and categories was lacking. For instance, ZeroGPT exhibited a higher FPR for original abstracts written by native authors compared to those written by non-native authors, whereas DetectGPT demonstrated the opposite trend. Regarding discipline-based variability, FNR values for ZeroGPT and DetectGPT were lowest for social sciences abstracts, but their highest FNRs occurred in different categories: ZeroGPT struggled the most with AI-generated abstracts in technology & engineering disciplines, while DetectGPT performed worst with AI-generated abstracts in interdisciplinary disciplines.
For the dataset-wide metrics, FAR showed no difference between native and non-native authors, remaining constant at 44.44%. However, MFAR was slightly higher for non-native authors (5.56%) compared to native authors (2.78%). Across disciplines, social sciences abstracts consistently exhibited the lowest FAR (25%) and MFAR (0%), whereas interdisciplinary abstracts had the highest FAR (62.50%), and technology & engineering abstracts showed the highest MFAR (8.33%). These findings underscore the tools’ struggles with texts in interdisciplinary and technology & engineering disciplines, particularly those authored by non-native English speakers.
While the results from this first scenario offer some insight into how these tools perform, they only address cases where the ground truth is clear. In real-world contexts these days, texts often blend human-written texts with AI-generated texts. The second scenario fills this gap by examining how these tools handle AI-assisted abstracts and whether biases emerge across author statuses or disciplines.
Scenario 2: AI-assisted abstracts
In this scenario, quantitative measures (i.e., the probability scores of AI-generated text) are used to assess the performance of each AI text detection tool, and the summary statistics are presented in Table 5. This approach is more suitable given that the evaluated texts (i.e., AI-assisted abstracts) blend human-written content with AI-generated enhancements, creating a more nuanced category where clear-cut labels are no longer applicable.
Table 5. Summary statistics of detection scores from Scenario 2: AI-assisted abstracts.
| Statistic | GPTZero (%) | ZeroGPT (%) | DetectGPT (%) | 
|---|---|---|---|
| Mean (SD) | 37.65 (39.99) | 20.92 (29.24) | 52.36 (47.40) | 
| • By author categories | |||
| • Native | 30.68 (35.29) | 21.91 (27.79) | 54.40 (46.79) | 
| • Non-native | 44.61 (43.33) | 19.94 (30.79) | 50.32 (48.24) | 
| • By disciplines | |||
| • Tech & Engineering | 41.04 (40.21) | 10.99 (21.30) | 53.90 (48.41) | 
| • Social Sciences | 35.21 (38.12) | 20.06 (33.15) | 52.10 (46.92) | 
| • Interdisciplinary | 36.69 (42.16) | 31.73 (28.77) | 51.08 (47.81) | 
| • By LLMs | |||
| • ChatGPT o1 | 19.79 (30.51) | 10.04 (21.50) | 29.47 (21.50) | 
| • Gemini 2.0 Pro | 55.50 (40.55) | 31.80 (31.93) | 75.25 (31.93) | 
| Q1, Median, Q3 | 2.00, 12.50, 80.25 | 0.00, 0.00, 40.51 | 0.00, 80.50, 100.00 | 
| • By author categories | |||
| • Native | 3.00, 9.50, 58.00 | 0.00, 0.00, 44.84 | 0.00, 81.50, 100.00 | 
| • Non-native | 2.00, 22.50, 99.25 | 0.00, 0.00, 36.38 | 0.00, 80.50, 100.00 | 
| • By disciplines | |||
| • Tech & Engineering | 2.00, 26.50, 80.25 | 0.00, 0.00, 5.85 | 0.00, 80.50, 100.00 | 
| • Social Sciences | 2.00, 14.00, 74.25 | 0.00, 0.00, 34.72 | 0.00, 79.50, 100.00 | 
| • Interdisciplinary | 3.00, 9.00, 90.25 | 0.00, 30.48, 52.96 | 0.00, 83.50, 100.00 | 
| • By LLMs | |||
| • ChatGPT o1 | 1.00, 5.00, 25.75 | 0.00, 0.00, 0.00 | 0.00, 0.00, 0.00 | 
| • Gemini 2.0 Pro | 9.00, 58.00, 100.00 | 0.00, 29.34, 53.06 | 0.00, 78.50, 91.00 | 
| Min, Max | 0.00, 100.00 | 0.00, 100.00 | 0.00, 100.00 | 
| • By author categories | |||
| • Native | 0.00, 100.00 | 0.00, 100.00 | 0.00, 100.00 | 
| • Non-native | 0.00, 100.00 | 0.00, 100.00 | 0.00, 100.00 | 
| • By disciplines | |||
| • Tech & Engineering | 0.00, 100.00 | 0.00, 79.43 | 0.00, 100.00 | 
| • Social Sciences | 0.00, 100.00 | 0.00, 100.00 | 0.00, 100.00 | 
| • Interdisciplinary | 0.00, 100.00 | 0.00, 96.64 | 0.00, 100.00 | 
| • By LLMs | |||
| • ChatGPT o1 | 0.00, 100.00 | 0.00, 100.00 | 0.00, 100.00 | 
| • Gemini 2.0 Pro | 0.00, 100.00 | 0.00, 100.00 | 0.00, 100.00 | 
The findings reveal a clear bias in GPTZero, which tends to assign lower AI-generated probabilities to AI-assisted abstracts written by native authors (median = 9.50%, mean = 30.68%, SD = 35.29%) and higher probabilities to those written by non-native authors (median = 22.50%, mean = 44.61%, SD = 43.33%). Welch’s t-test confirms this disparity, showing a statistically significant difference (t = −2.115, p = 0.036). This pattern, visualized in the density plot in Fig. 4, highlights a concerning imbalance that potentially places non-native authors at a serious disadvantage, amplifying existing inequities in scholarly publishing. In contrast, the other two tools, ZeroGPT and DetectGPT, exhibit more balanced performance, as evidenced by their Welch’s t-test results, which show no statistically significant differences (p = 0.687 and p = 0.607, respectively). However, given GPTZero’s high accuracy in the previous scenario, these findings raise significant concerns about the unintended consequences of its use.
Figure 4. Density plots of AI-generated probability scores for AI-assisted abstracts from each AI text detection tool by author status.
The findings also reveal that ZeroGPT struggles significantly in detecting AI-assisted abstracts from technology & engineering disciplines (median = 0%, mean = 10.99%, SD = 21.30%), compared to social sciences (median = 0%, mean = 20.06%, SD = 33.15%) and even more so interdisciplinary abstracts (median = 30.48%, mean = 31.73%, SD = 28.77%). Welch’s ANOVA confirms this disparity, with a statistically significant difference (F = 7.94, p < 0.001). This pattern is clearly illustrated in the density plot in Fig. 5. GPTZero (F = 0.28, p = 0.759) and DetectGPT (F = 0.04, p = 0.960), on the other hand, exhibit relatively more consistent performance across disciplines, as reflected in their non-significant Welch’s ANOVA results. These findings highlight the challenges AI text detection tools face when applied to different disciplines, with ZeroGPT showing particular weaknesses in more technical fields.
Figure 5. Density plots of AI-generated probability scores for AI-assisted abstracts from each AI text detection tool by discipline.
In terms of LLMs, the findings reveal that abstracts enhanced by Gemini 2.0 Pro Experimental are significantly more likely to be detected as AI-generated than those enhanced by ChatGPT o1, across all three detector tools. GPTZero assigns a notably higher mean probability score to Gemini-enhanced abstracts (mean = 55.50%, SD = 40.55%) compared to those enhanced by ChatGPT (mean = 19.79%, SD = 30.51%). Welch’s t-test indicates this difference is statistically significant (t = −5.97, p < 0.001). Similarly, ZeroGPT assigns higher mean probability scores to Gemini-enhanced abstracts (mean = 31.80%, SD = 31.93%) compared to ChatGPT-enhanced abstracts (mean = 10.04%, SD = 21.50%), with Welch’s t-test confirming significance (t = −4.80, p < 0.001). DetectGPT also shows the same pattern, assigning much higher mean probability scores to Gemini-enhanced abstracts (mean = 75.25%, SD = 31.93%) compared to ChatGPT-enhanced abstracts (mean = 29.47%, SD = 21.50%), a huge difference that once again confirmed to be statistically significant with Welch’s t-test (t = −6.60, p < 0.001). These findings indicate that using either LLM to enhance text increases the AI-generated probability scores assigned by detection tools, even when the underlying content is human-authored. However, among the two models used to enhance the original text, the risk of elevated scores appears to be more pronounced with Gemini-enhanced text, as depicted in the density plot in Fig. 6. This may reflect distinct stylistic or linguistic patterns detectable by current tools. Consequently, authors relying on Gemini Pro 2.0 Experimental are at greater risk of their work being unfairly flagged as AI-generated by detection tools than those relying on ChatGPT o1.
Figure 6. Density plots of AI-generated probability scores for AI-assisted abstracts from each AI text detection tool by LLM.
Finally, Table 6 presents UDR and ODR for AI-assisted abstracts, broken down by author status, discipline, and LLM. Across all tools, the ODR for non-native authors is consistently higher than for native authors, with GPTZero showing the most pronounced disparity: its ODR for non-native authors (25%) is more than double that of native authors (11%). In other words, one in four non-native authors who use AI to help refine their text is at risk of being accused of having submitted an entirely AI-generated text by GPTZero, while the risk for native authors is closer to one in ten. This pattern, once again, underscores the disproportionate risks faced by non-native authors, who ironically are the ones most likely to benefit from using LLMs to enhance their English writing, yet are also the most likely to have their contributions dismissed entirely by detection tools.
Table 6. Overall performance metrics from Scenario 2: AI-assisted abstracts.
| Metric | GPTZero (%) | ZeroGPT (%) | DetectGPT (%) | 
|---|---|---|---|
| Under-Detection Rate (UDR) | 3.47 | 58.33 | 44.44 | 
| • By author categories | |||
| • Native | 4.17 | 52.78 | 41.67 | 
| • Non-Native | 2.78 | 63.89 | 47.22 | 
| • By disciplines | |||
| • Technology & Engineering | 4.17 | 75.00 | 43.75 | 
| • Social Sciences | 4.17 | 66.67 | 43.75 | 
| • Interdisciplinary | 2.08 | 33.33 | 45.83 | 
| • By LLMs | |||
| • ChatGPT o1 | 5.56 | 76.39 | 69.44 | 
| • Gemini 2.0 Pro | 1.39 | 40.28 | 19.44 | 
| Over-Detection Rate (ODR) | 18.06 | 2.08 | 34.03 | 
| • By author categories | |||
| • Native | 11.11 | 1.39 | 33.33 | 
| • Non-Native | 25.00 | 2.78 | 34.72 | 
| • By disciplines | |||
| • Technology & Engineering | 16.67 | 0.00 | 41.67 | 
| • Social Sciences | 12.50 | 6.25 | 29.17 | 
| • Interdisciplinary | 25.00 | 0.00 | 31.25 | 
| • By LLMs | |||
| • ChatGPT o1 | 8.33 | 0.00 | 22.22 | 
| • Gemini 2.0 Pro | 27.78 | 4.17 | 45.83 | 
Meanwhile, although ZeroGPT seemed to be the most prudent, with the lowest ODR value of all three tools, it also exhibited the highest UDR, with values consistently exceeding 50% across all author statuses and disciplines except for interdisciplinary. Furthermore, while other tools typically use a 50% cutoff to classify text as AI-generated, ZeroGPT relies on a lower, less transparent threshold. For instance, it labeled some original abstracts in Scenario 1 as AI-generated even though their scores were below 30%. This combination of moderately high FPR in Scenario 1 and high UDR in Scenario 2 highlights the tool’s difficulty in balancing false positives and false negatives across varying contexts.
On the other hand, DetectGPT showed moderately high values for both UDR and ODR, indicating a strong tendency to make bold yet incorrect classifications. Coupled with the lowest accuracy and the highest FPR and FNR observed in the previous scenario, these findings underscore DetectGPT’s limitations. Ultimately, DetectGPT emerged as the worst-performing AI text detection tool in this study.
Discussion
The findings in this study highlight significant challenges in the use of AI text detection tools, driven in part by the rapidly evolving landscape of LLMs and generative AI. Updates to LLMs can quickly render existing detection tools ineffective, as demonstrated by the poor performance of ZeroGPT and DetectGPT in this study. These tools struggled to handle texts generated by ChatGPT o1 and Gemini 2.0 Pro Experimental, the two most advanced models available at the time of the study, revealing limitations in their adaptability. These findings are consistent with previous research evaluating AI text detection tools on earlier versions of these LLMs (Perkins et al., 2024; Elkhatat, Elsaid & Almeer, 2023; Weber-Wulff et al., 2023; Liang et al., 2023), suggesting that the detection-based approach may be fundamentally flawed. This dynamic creates a perpetual cycle of advancement and adaptation, where detection tools must continuously try to catch up. However, this urgency often leads to trade-offs, such as introducing biases against specific groups of authors or disciplines, as seen with GPTZero, or increasing the risk of false accusations that disproportionately affect non-native authors.
These findings have significant implications for scholarly publication. Non-native authors, who already face systemic challenges, are disproportionately affected by the biases of these tools, particularly GPTZero. The risk of being falsely accused of submitting AI-generated content, even when AI was used solely to improve the readability of their own writing, as shown in Scenario 2, could further exacerbate existing inequities in access to publication opportunities. This is especially concerning given that non-native authors are far more likely than native authors to rely on AI tools to enhance the clarity and readability of their work (Liao et al., 2024). The ethical implications of using AI text detection tools go beyond technical performance. The lack of transparency in how these tools evaluate text, combined with their potential biases, poses serious risks to authors’ reputations and careers. Institutions, journal editors, and publishers must take these risks into account when adopting such tools, ensuring they are used as supplementary aids rather than definitive decision-makers.
Given the widely acknowledged language barriers in scholarly publication, LLMs have the potential to level the playing field for non-native authors. Rather than stigmatizing their use, it may be more beneficial to embrace these tools as valuable aids for improving the clarity and accessibility of academic writing. Notably, some researchers, journal editors, and even publishers have voiced support for the responsible adoption of LLMs, recognizing their potential to enhance inclusivity and reduce linguistic barriers (Seghier, 2023; Kaebnick et al., 2023; Koller et al., 2024). The focus, therefore, should not be on penalizing authors who use LLMs, but on promoting ethical use, ensuring transparency, and holding authors accountable for the accuracy and integrity of their work.
At the end of the day, LLMs are merely tools, and like any tool, their misuse or overreliance without oversight can lead to significant issues. One of the most critical concerns in content generated by LLMs is AI hallucinations (Salvagno, Taccone & Gerli, 2023; Athaluri et al., 2023; Hatem et al., 2023), where LLMs generate false or misleading information, essentially fabrications and falsifications (Emsley, 2023). Another potential issue is plagiarism, where human authors use a simple prompt to have the LLM generate text from scratch and then present it as their own original work. These behaviors constitute scientific misconduct and are unacceptable. Authors who fail to verify and validate any AI involvement in the production of their content and manuscript preparation should be held accountable for the consequences. Ultimately, ethical use and transparency are essential for the responsible integration of LLMs into scholarly publication.
Furthermore, a shift in the default assumption may be necessary to achieve greater transparency. Currently, many journals and publishers require authors to disclose the use of LLMs for improving readability (Cheng & Wu, 2024; Ganjavi et al., 2024; Flanagin et al., 2023). Given the widespread availability of LLMs, their increasing integration into writing workflows, and evidence that some authors choose not to disclose their AI use (Pesante, Mauffrey & Parry, 2024; Chemaya & Martin, 2024), it may be more practical to assume that every submitted manuscript has had some level of AI involvement for copyediting or improving readability, unless authors explicitly state otherwise. For instance, requiring a disclaimer such as “no AI involvement” could serve as a clear affirmation that the manuscript was entirely written and edited by humans without any AI support. That said, any other forms of LLM involvement, such as generating text, images, or other content from scratch using a prompt, should be fully disclosed. Ultimately, the emphasis should remain on the substance of the research and the scientific rigor it reflects, rather than the tools used in its production.
Limitations and future work
Several limitations should be acknowledged to appropriately contextualize the results of this study. First, the selection of articles published up to 2021, while intentionally chosen to predate the public accessibility of generative AI tools like ChatGPT, still overlaps with earlier GPT-based technologies (e.g., GPT-2 released in 2019 and GPT-3 released in 2020). Although fully generative tools capable of producing coherent, lengthy texts from scratch were not widely available to the general public until late 2022, earlier GPT-based tools (e.g., Grammarly) were primarily used for language refinement rather than text generation. Therefore, it remains reasonable to assume that the original texts in journals published in 2021 were human-authored, though the possibility that GPT-based tools contributed to editing or polishing the writing cannot be entirely ruled out.
Second, the disciplinary scope of the selected journals was intentionally limited due to practical constraints. This study included journals from technology & engineering and social sciences to represent contrasting fields, as well as interdisciplinary journals positioned in between. However, this limited representation may not fully generalize to broader STEM and non-STEM categories. Future research should expand disciplinary coverage to include a broader spectrum of fields, such as natural sciences, humanities, health sciences, and professional disciplines like medicine, law, or economics to strengthen the generalizability of the results. Expanding the range of disciplines would not only improve representativeness but also support the development of a larger and more diverse dataset, enabling more robust and nuanced analyses across fields.
Finally, the analysis focused on a limited number of popular and publicly available AI detection tools. Considering the rapid evolution of AI detection technologies, future studies should incorporate additional detection tools, particularly those commonly used by higher education institutions (e.g., Turnitin) to enhance the real-world applicability and relevance of the findings.
Addressing these limitations in future work will further enrich our understanding of AI-generated and AI-assisted writing, highlight the inherent flaws and biases of AI text detection tools, and reinforce the need for alternative approaches that prioritize transparency, fairness, and ethical use of AI in scholarly publication.
Conclusion
This study evaluates the performance of popular AI text detection tools and highlights the challenges associated with their practical use in the scholarly publication realm. Despite claims of high accuracy, most tools in this study struggled with both accuracy and consistency when applied to texts involving the latest versions of LLMs. The study also reveals notable accuracy-bias trade-offs, where tools with higher overall accuracy (such as GPTZero) exhibit stronger biases against non-native authors and certain disciplines, in both original and AI-assisted human writings.
The findings underscore the need for promoting ethical and transparent use of LLMs over detection-focused approaches. When used responsibly, LLMs have the potential to help level the playing field for non-native authors. Clear guidelines that encourage disclosure of AI involvement while maintaining author accountability are essential to fostering trust in scholarly publishing. As LLMs continue to evolve, collaborative efforts are needed to ensure that scholarly communication remains equitable by keeping the focus on the quality and integrity of the research, rather than on the tools used in its production.
Supplemental Information
Acknowledgments
In addition to using ChatGPT and Gemini to generate part of the dataset in this study, the author also used ChatGPT to improve the readability of this manuscript and Gemini in Google Colab to troubleshoot and fix errors in the Python code used for data analysis and visualization.
Funding Statement
The author received no funding for this work.
Additional Information and Declarations
Competing Interests
The authors declare that they have no competing interests.
Author Contributions
Ahmad R. Pratama conceived and designed the experiments, performed the experiments, analyzed the data, performed the computation work, prepared figures and/or tables, authored or reviewed drafts of the article, and approved the final draft.
Data Availability
The following information was supplied regarding data availability:
The complete dataset, along with the experimental results and Python code used to analyze and visualize the findings, is available at GitHub and Zenodo:
https://github.com/ahmadrpratama/ai-text-detection-bias
ahmadrpratama. (2025). ahmadrpratama/ai-text-detection-bias: v1.0.0: Data and code for “The Accuracy-Bias Trade-Offs in AI Text Detection Tools and Their Impact on Fairness in Scholarly Publication” (PeerJ Computer Science) (v1.0.0). Zenodo. https://doi.org/10.5281/zenodo.15490003.
References
- Alhassan (2021).Alhassan A. Challenges and professional development needs of EMI lecturers in omani higher education. Sage Open. 2021;11(4):21582440211061527. doi: 10.1177/21582440211061527. [DOI] [Google Scholar]
- Alluqmani & Shamir (2018).Alluqmani A, Shamir L. Writing styles in different scientific disciplines: a data science approach. Scientometrics. 2018;115(2):1071–1085. doi: 10.1007/s11192-018-2688-8. [DOI] [Google Scholar]
- Alshaer, Moawad & Ismail (2021).Alshaer N, Moawad A, Ismail T. Reliability and security analysis of an entanglement-based QKD protocol in a dynamic ground-to-UAV FSO communications system. IEEE Access. 2021;9:168052–168067. doi: 10.1109/ACCESS.2021.3137357. [DOI] [Google Scholar]
- Altalhi (2021).Altalhi M. Toward a model for acceptance of MOOCs in higher education: the modified UTAUT model for Saudi Arabia. Education and Information Technologies. 2021;26(2):1589–1605. doi: 10.1007/s10639-020-10317-x. [DOI] [Google Scholar]
- Amano et al. (2023).Amano T, Ramírez-Castañeda V, Berdejo-Espinola V, Borokini I, Chowdhury S, Golivets M, González-Trujillo JD, Montaño-Centellas F, Paudel K, White RL, Veríssimo D. The manifold costs of being a non-native English speaker in science. PLOS Biology. 2023;21(7):e3002184. doi: 10.1371/journal.pbio.3002184. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Athaluri et al. (2023).Athaluri SA, Manthena SV, Kesapragada VSRKM, Yarlagadda V, Dave T, Duddumpudi RTS, Athaluri SA, Manthena SV, Kesapragada VSRKM, Yarlagadda V, Dave T, Duddumpudi RTS. Exploring the boundaries of reality: investigating the phenomenon of artificial intelligence hallucination in scientific writing through ChatGPT references. Cureus. 2023;15:e37432. doi: 10.7759/cureus.37432. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Bae (2021).Bae W. Today’s computing challenges: opportunities for computer hardware design. PeerJ Computer Science. 2021;7(1):e420. doi: 10.7717/peerj-cs.420. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Bhat et al. (2021).Bhat V, Yadav A, Yadav S, Chandrasekaran D, Mago V. AdCOFE: advanced contextual feature extraction in conversations for emotion classification. PeerJ Computer Science. 2021;7(4):e786. doi: 10.7717/peerj-cs.786. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Brablec (2021).Brablec D. Who counts as an authentic indigenous? Collective identity negotiations in the chilean urban context. Sociology. 2021;55(1):129–145. doi: 10.1177/0038038520915435. [DOI] [Google Scholar]
- Bui et al. (2021).Bui M-QV, Ngo DT, Pham H-A, Nguyen DD. GAC3D: improving monocular 3D object detection with ground-guide model and adaptive convolution. PeerJ Computer Science. 2021;7(2):e686. doi: 10.7717/peerj-cs.686. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Çelik (2021).Çelik Ç. Rethinking institutional habitus in education: a relational approach for studying its sources and impacts. Sociology. 2021;55(3):522–538. doi: 10.1177/0038038520966565. [DOI] [Google Scholar]
- Challob (2021).Challob AI. The effect of flipped learning on EFL students’ writing performance, autonomy, and motivation. Education and Information Technologies. 2021;26(4):3743–3769. doi: 10.1007/s10639-021-10434-1. [DOI] [Google Scholar]
- Chemaya & Martin (2024).Chemaya N, Martin D. Perceptions and detection of AI use in manuscript preparation for academic journals. PLOS ONE. 2024;19(7):e0304807. doi: 10.1371/journal.pone.0304807. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Chen (2021).Chen Y-H. Self-competence learning for bandwidth slicing and confidential computing requirement. IEEE Access. 2021;9:169156–169160. doi: 10.1109/ACCESS.2021.3137827. [DOI] [Google Scholar]
- Cheng & Wu (2024).Cheng K, Wu H. Policy framework for the utilization of generative AI. Critical Care. 2024;28(1):128. doi: 10.1186/s13054-024-04917-z. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Cho (2004).Cho S. Challenges of entering discourse communities through publishing in English: perspectives of nonnative-speaking doctoral students in the United States of America. Journal of Language, Identity & Education. 2004;3(1):47–72. doi: 10.1207/s15327701jlie0301_3. [DOI] [Google Scholar]
- Christoe et al. (2021).Christoe MJ, Yuan J, Michael A, Kalantar-Zadeh K. Bluetooth signal attenuation analysis in human body tissue analogues. IEEE Acces. 2021;9:85144–85150. doi: 10.1109/ACCESS.2021.3087780. [DOI] [Google Scholar]
- Coburn (2021).Coburn E. Sociology versus economics: economic life as social fact and social struggle. International Sociology. 2021;36(5):720–731. doi: 10.1177/02685809211057474. [DOI] [Google Scholar]
- Cornejo-Lupa et al. (2020).Cornejo-Lupa MA, Ticona-Herrera RP, Cardinale Y, Barrios-Aranibar D. A survey of ontologies for simultaneous localization and mapping in mobile robots. ACM Computing Surveys. 2020;53(5):103:1–103:26. doi: 10.1145/3408316. [DOI] [Google Scholar]
- de Brito Lima, Lautert & Gomes (2021).de Brito Lima F, Lautert SL, Gomes AS. Contrasting levels of student engagement in blended and non-blended learning scenarios. Computers & Education. 2021;172(3):104241. doi: 10.1016/j.compedu.2021.104241. [DOI] [Google Scholar]
- Dong et al. (2024).Dong S, Mao J, Ke Q, Pei L. Decoding the writing styles of disciplines: a large-scale quantitative analysis. Information Processing & Management. 2024;61(4):103718. doi: 10.1016/j.ipm.2024.103718. [DOI] [Google Scholar]
- Dong, Mao & Pei (2023).Dong S, Mao J, Pei L. Comparing the writing styles of multiple disciplines: a large-scale quantitative analysis. Proceedings of the Association for Information Science and Technology. 2023;60(1):941–943. doi: 10.1002/pra2.905. [DOI] [Google Scholar]
- Elkhatat, Elsaid & Almeer (2023).Elkhatat AM, Elsaid K, Almeer S. Evaluating the efficacy of AI content detection tools in differentiating between human and AI-generated text. International Journal for Educational Integrity. 2023;19(1) doi: 10.1007/s40979-023-00140-5. 1. [DOI] [Google Scholar]
- Elshoush, Al-Tayeb & Obeid (2021).Elshoush HT, Al-Tayeb BM, Obeid KT. Enhanced Serpent algorithm using Lorenz 96 Chaos-based block key generation and parallel computing for RGB image encryption. PeerJ Computer Science. 2021;7(3):e812. doi: 10.7717/peerj-cs.812. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Emsley (2023).Emsley R. ChatGPT: these are not hallucinations–they’re fabrications and falsifications. Schizophrenia. 2023;9(1):1–2. doi: 10.1038/s41537-023-00379-4. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Erkmen (2021).Erkmen TD. Flexible selves in flexible times? Yoga and neoliberal subjectivities in Istanbul. Sociology. 2021;55(5):1035–1052. doi: 10.1177/0038038521998930. [DOI] [Google Scholar]
- Falcón-Cardona & Coello (2020).Falcón-Cardona JG, Coello CAC. Indicator-based multi-objective evolutionary algorithms: a comprehensive survey. ACM Computing Surveys. 2020;53(2):29:1–29:35. doi: 10.1145/3376916. [DOI] [Google Scholar]
- Ferguson, Pérez-Llantada & Plo (2011).Ferguson G, Pérez-Llantada C, Plo R. English as an international language of scientific publication: a study of attitudes. World Englishes. 2011;30(1):41–59. doi: 10.1111/j.1467-971X.2010.01656.x. [DOI] [Google Scholar]
- Flanagin et al. (2023).Flanagin A, Bibbins-Domingo K, Berkwits M, Christiansen SL. Nonhuman “Authors” and implications for the integrity of scientific publication and medical knowledge. The Journal of The American Medical Association. 2023;329(8):637–639. doi: 10.1001/jama.2023.1344. [DOI] [PubMed] [Google Scholar]
- Fletcher & Stanzione (2021).Fletcher KA, Stanzione CM. A mixed-methods approach to understanding laptop-free zones in college classrooms. Computers & Education. 2021;172(5):104253. doi: 10.1016/j.compedu.2021.104253. [DOI] [Google Scholar]
- Flowerdew (2001).Flowerdew J. Attitudes of journal editors to nonnative speaker contributions. TESOL Quarterly. 2001;35(1):121–150. doi: 10.2307/3587862. [DOI] [Google Scholar]
- Fujita (2021).Fujita K. Approximate spectral clustering using both reference vectors and topology of the network generated by growing neural gas. PeerJ Computer Science. 2021;7(10):e679. doi: 10.7717/peerj-cs.679. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Gallo-Cruz (2021).Gallo-Cruz S. Films at the fault line of ‘deep’ and ‘bright’ greens in the environmental movement. International Sociology. 2021;36(5):745–756. doi: 10.1177/02685809211057482. [DOI] [Google Scholar]
- Ganjavi et al. (2024).Ganjavi C, Eppler MB, Pekcan A, Biedermann B, Abreu A, Collins GS, Gill IS, Cacciamani GE. Publishers’ and journals’ instructions to authors on use of generative artificial intelligence in academic and scientific publishing: bibliometric analysis. BMJ. 2024;384:e077192. doi: 10.1136/bmj-2023-077192. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Gibson et al. (2021).Gibson C, Ross K, Williams M, de Vries N. The experiences of mothers in a neonatal unit and their use of the babble app. Sage Open. 2021;11(2):21582440211023170. doi: 10.1177/21582440211023170. [DOI] [Google Scholar]
- Giray (2024b).Giray L. The problem with false positives: AI detection unfairly accuses scholars of AI plagiarism. The Serials Librarian. 2024b;85(5–6):181–189. doi: 10.1080/0361526X.2024.2433256. [DOI] [Google Scholar]
- Giray (2024a).Giray L. AI shaming: the silent stigma among academic writers and researchers. Annals of Biomedical Engineering. 2024a;52(9):2319–2324. doi: 10.1007/s10439-024-03582-1. [DOI] [PubMed] [Google Scholar]
- Hatem et al. (2023).Hatem R, Simmons B, Thornton JE, Hatem R, Simmons B, Thornton JE. A call to address AI “Hallucinations” and how healthcare professionals can mitigate their risks. Cureus. 2023;15:e44720. doi: 10.7759/cureus.44720. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Hayashi, Shibanoki & Tsuji (2021).Hayashi H, Shibanoki T, Tsuji T. A neural network based on the Johnson SU translation system and related application to electromyogram classification. IEEE Access. 2021;9:154304–154317. doi: 10.1109/ACCESS.2021.3126348. [DOI] [Google Scholar]
- Hehir et al. (2021).Hehir E, Zeller M, Luckhurst J, Chandler T. Developing student connectedness under remote learning using digital resources: a systematic review. Education and Information Technologies. 2021;26(5):6531–6548. doi: 10.1007/s10639-021-10577-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Herodotou et al. (2021).Herodotou C, Maguire C, McDowell N, Hlosta M, Boroowa A. The engagement of university teachers with predictive learning analytics. Computers & Education. 2021;173(2):104285. doi: 10.1016/j.compedu.2021.104285. [DOI] [Google Scholar]
- Herrera & Rivera (2021).Herrera JM, Rivera M. Migration, emotions and policies of sensibilities in Central America. International Sociology. 2021;36(4):569–584. doi: 10.1177/0268580921993332. [DOI] [Google Scholar]
- Holmes (2021).Holmes M. Emotions in Aotearoa New Zealand: reflexive emotionalisation in a colonised context. International Sociology. 2021;36(4):498–514. doi: 10.1177/0268580921995301. [DOI] [Google Scholar]
- Horn (2017).Horn SA. Non-English nativeness as stigma in academic settings. Academy of Management Learning & Education. 2017;16(4):579–602. doi: 10.5465/amle.2015.0194. [DOI] [Google Scholar]
- Hu & Yin (2021).Hu A, Yin C. The distinction between the absolute and relative advantages of cultural capital: different conceptualizations, different consequences. Sociology. 2021;55(4):803–822. doi: 10.1177/0038038520973588. [DOI] [Google Scholar]
- Huang et al. (2021).Huang L, Li S, Poitras EG, Lajoie SP. Latent profiles of self-regulated learning and their impacts on teachers’ technology integration. British Journal of Educational Technology. 2021;52(2):695–713. doi: 10.1111/bjet.13050. [DOI] [Google Scholar]
- Hudson & Moubayed (2021).Hudson GT, Moubayed NA. Ask me in your own words: paraphrasing for multitask question answering. PeerJ Computer Science. 2021;7(4):e759. doi: 10.7717/peerj-cs.759. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Hwang, Chien & Li (2021).Hwang G-J, Chien S-Y, Li W-S. A multidimensional repertory grid as a graphic organizer for implementing digital games to promote students’ learning performances and behaviors. British Journal of Educational Technology. 2021;52(2):915–933. doi: 10.1111/bjet.13062. [DOI] [Google Scholar]
- Jacobs et al. (2021).Jacobs M, Benovoy M, Chang L-C, Corcoran D, Berry C, Arai AE, Hsu L-Y. Automated segmental analysis of fully quantitative myocardial blood flow maps by first-pass perfusion cardiovascular magnetic resonance. IEEE Access. 2021;9:52796–52811. doi: 10.1109/ACCESS.2021.3070320. [DOI] [PMC free article] [PubMed] [Google Scholar]
- James, Delfabbro & King (2021).James O, Delfabbro P, King DL. A comparison of psychological and work outcomes in open-plan and cellular office designs: a systematic review. Sage Open. 2021;11(1):2158244020988869. doi: 10.1177/2158244020988869. [DOI] [Google Scholar]
- Janakiraman et al. (2021).Janakiraman S, Watson SL, Watson WR, Newby T. Effectiveness of digital games in producing environmentally friendly attitudes and behaviors: a mixed methods study. Computers & Education. 2021;160(1):104043. doi: 10.1016/j.compedu.2020.104043. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Jingting & Chao (2021).Jingting Z, Chao J. Between tradition and modernity: emotional changes and expressions of Chinese society. International Sociology. 2021;36(4):515–531. doi: 10.1177/0268580921994502. [DOI] [Google Scholar]
- Jung (2021).Jung G. Transnational logic of gratitude: marriage migrants’ narratives on marital satisfaction in South Korea. International Sociology. 2021;36(6):844–867. doi: 10.1177/0268580921994692. [DOI] [Google Scholar]
- Kaebnick et al. (2023).Kaebnick GE, Magnus DC, Kao A, Hosseini M, Resnik D, Dubljević V, Rentmeester C, Gordijn B, Cherry MJ. Editors’ statement on the responsible use of generative AI technologies in scholarly journal publishing. Hastings Center Report. 2023;53:3–6. doi: 10.1002/hast.1507. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Kendall & Teixeira da Silva (2024).Kendall G, Teixeira da Silva JA. Risks of abuse of large language models, like ChatGPT, in scientific publishing: authorship, predatory publishing, and paper mills. Learned Publishing. 2024;37(1):55–62. doi: 10.1002/leap.1578. [DOI] [Google Scholar]
- Khader & Al-Naymat (2020).Khader M, Al-Naymat G. Density-based algorithms for big data clustering using MapReduce framework: a comprehensive study. ACM Computing Surveys. 2020;53(5):93:1–93:38. doi: 10.1145/3403951. [DOI] [Google Scholar]
- Kingsbury (2021).Kingsbury I. Online learning: how do brick and mortar schools stack up to virtual schools? Education and Information Technologies. 2021;26(6):6567–6588. doi: 10.1007/s10639-021-10450-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Kojima & Popiel (2022).Kojima T, Popiel HA. Proper scholarly writing for non-native English-speaking authors: choosing active and passive voice, rewording, and refining texts. Journal of Korean Medical Science. 2022;37(44):e312. doi: 10.3346/jkms.2022.37.e312. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Koller et al. (2024).Koller D, Beam A, Manrai A, Ashley E, Liu X, Gichoya J, Holmes C, Zou J, Dagan N, Wong TY, Blumenthal D, Kohane I, the editors and editorial board of NEJM AI Why we support and encourage the use of large language models in NEJM AI submissions. NEJM AI. 2024;1(1):AIe2300128. doi: 10.1056/AIe2300128. [DOI] [Google Scholar]
- Kuteeva & Andersson (2024).Kuteeva M, Andersson M. Diversity and standards in writing for publication in the age of AI—between a rock and a hard place. Applied Linguistics. 2024;45(3):561–567. doi: 10.1093/applin/amae025. [DOI] [Google Scholar]
- Kwon et al. (2021).Kwon Y, Dwivedi A, McDaid AJ, Liarokapis M. Electromyography-based decoding of dexterous, in-hand manipulation of objects: comparing task execution in real world and virtual reality. IEEE Access. 2021;9:37297–37310. doi: 10.1109/ACCESS.2021.3062364. [DOI] [Google Scholar]
- Latifi, Noroozi & Talaee (2021).Latifi S, Noroozi O, Talaee E. Peer feedback or peer feedforward? Enhancing students’ argumentative peer learning processes and outcomes. British Journal of Educational Technology. 2021;52(2):768–784. doi: 10.1111/bjet.13054. [DOI] [Google Scholar]
- Le et al. (2021).Le Y, Chen Z, Liu S, Pang W, Deng C. Investigating the effectiveness of emotional design principle to attenuate ego depletion effect. Computers & Education. 2021;174(1):104311. doi: 10.1016/j.compedu.2021.104311. [DOI] [Google Scholar]
- Lei et al. (2024).Lei F, Du L, Dong M, Liu X. Global retractions due to randomly generated content: characterization and trends. Scientometrics. 2024;129(12):7943–7958. doi: 10.1007/s11192-024-05172-3. [DOI] [Google Scholar]
- Li et al. (2021b).Li F-Y, Hwang G-J, Chen P-Y, Lin Y-J. Effects of a concept mapping-based two-tier test strategy on students’ digital game-based learning performances and behavioral patterns. Computers & Education. 2021b;173(5):104293. doi: 10.1016/j.compedu.2021.104293. [DOI] [Google Scholar]
- Li et al. (2021a).Li Y, Li X, Zhang Y, Li X. The effects of a group awareness tool on knowledge construction in computer-supported collaborative learning. British Journal of Educational Technology. 2021a;52(3):1178–1196. doi: 10.1111/bjet.13066. [DOI] [Google Scholar]
- Liang et al. (2023).Liang W, Yuksekgonul M, Mao Y, Wu E, Zou J. GPT detectors are biased against non-native English writers. Patterns. 2023;4(7):100779. doi: 10.1016/j.patter.2023.100779. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Liao et al. (2024).Liao Z, Antoniak M, Cheong I, Cheng EY-Y, Lee A-H, Lo K, Chang JC, Zhang AX. LLMs as research tools: a large scale survey of researchers’ usage and perceptions. 2024. ArXiv preprint. [DOI]
- Lund & Naheem (2024).Lund BD, Naheem Kt. Can ChatGPT be an author? A study of artificial intelligence authorship policies in top academic journals. Learned Publishing. 2024;37(1):13–21. doi: 10.1002/leap.1582. [DOI] [Google Scholar]
- Magrath (2021).Magrath R. Gay male football fans’ experiences: authenticity, belonging and conditional acceptance. Sociology. 2021;55(5):978–994. doi: 10.1177/0038038521994012. [DOI] [Google Scholar]
- Mahmud, Ramamohanarao & Buyya (2020).Mahmud R, Ramamohanarao K, Buyya R. Application management in fog computing environments: a taxonomy, review and future directions. ACM Computing Surveys. 2020;53(4):88:1–88:43. doi: 10.1145/3403955. [DOI] [Google Scholar]
- Makarov et al. (2021).Makarov I, Kiselev D, Nikitinsky N, Subelj L. Survey on graph embeddings and their applications to machine learning problems on graphs. PeerJ Computer Science. 2021;7(3):e357. doi: 10.7717/peerj-cs.357. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Marino et al. (2021).Marino DL, Wickramasinghe CS, Singh VK, Gentle J, Rieger C, Manic M. The virtualized cyber-physical testbed for machine learning anomaly detection: a wind powered grid case study. IEEE Access. 2021;9:159475–159494. doi: 10.1109/ACCESS.2021.3127169. [DOI] [Google Scholar]
- Matsuo (2021).Matsuo M. Reflecting on success in difficult times: a key to enhance proactivity and employability. Sage Open. 2021;11(4):21582440211059167. doi: 10.1177/21582440211059167. [DOI] [Google Scholar]
- Mitchell et al. (2023).Mitchell E, Lee Y, Khazatsky A, Manning CD, Finn C. DetectGPT: zero-shot machine-generated text detection using probability curvature. Proceedings of the 40th International Conference on Machine Learning; 2023. pp. 24950–24962. [Google Scholar]
- Moro et al. (2021).Moro C, Phelps C, Redmond P, Stromberga Z. HoloLens and mobile augmented reality in medical and health science education: a randomised controlled trial. British Journal of Educational Technology. 2021;52(2):680–694. doi: 10.1111/bjet.13049. [DOI] [Google Scholar]
- Mumin (2022).Mumin MA. The academic dilemma on the use of proof readers in academic assignments. World Journal of English Language. 2022;12(4) doi: 10.5430/wjel.v12n4p16. 4. [DOI] [Google Scholar]
- Murthy et al. (2021).Murthy D, O’Brien JD, Gross A, Meyers N. Variations in the temporal structure of sociability across American cities. Sociology. 2021;55(1):30–55. doi: 10.1177/0038038520956940. [DOI] [Google Scholar]
- Navarro & Rojas-Ledesma (2020).Navarro G, Rojas-Ledesma J. Predecessor search. ACM Computing Surveys. 2020;53(5):105:1–105:35. doi: 10.1145/3409371. [DOI] [Google Scholar]
- Nazarovets & Teixeira da Silva (2024).Nazarovets S, Teixeira da Silva JA. ChatGPT as an “author”: bibliometric analysis to assess the validity of authorship. Accountability in Research. 2024 doi: 10.1080/08989621.2024.2345713. Epub ahead of print 1 May 2024. [DOI] [PubMed] [Google Scholar]
- Neves & Mead (2021).Neves BB, Mead G. Digital technology and older people: towards a sociological approach to technology adoption in later life. Sociology. 2021;55(5):888–905. doi: 10.1177/0038038520975587. [DOI] [Google Scholar]
- Park et al. (2021).Park JH, Tran NM, Hwang SI, Kim DI, Choi KW. Design and implementation of 5.8 GHz RF wireless power transfer system. IEEE Access. 2021;9:168520–168534. doi: 10.1109/ACCESS.2021.3138221. [DOI] [Google Scholar]
- Perkins et al. (2024).Perkins M, Roe J, Postma D, McGaughran J, Hickerson D. Detection of GPT-4 generated text in higher education: combining academic judgement and software to identify generative AI tool misuse. Journal of Academic Ethics. 2024;22:89–113. doi: 10.1007/s10805-023-09492-6. [DOI] [Google Scholar]
- Pesante, Mauffrey & Parry (2024).Pesante BD, Mauffrey C, Parry JA. Rise of the machines: the prevalence and disclosure of artificial intelligence–generated text in high-impact orthopaedic journals. JAAOS-Journal of the American Academy of Orthopaedic Surgeons. 2024;32(19):910–914. doi: 10.5435/JAAOS-D-24-00318. [DOI] [PubMed] [Google Scholar]
- Pileh Roud & Hidri (2021).Pileh Roud LF, Hidri S. Toward a sociocultural approach to computerized dynamic assessment of the TOEFL iBT listening comprehension test. Education and Information Technologies. 2021;26(4):4943–4968. doi: 10.1007/s10639-021-10498-z. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Pratama (2021).Pratama AR. Fun first, useful later: mobile learning acceptance among secondary school students in Indonesia. Education and Information Technologies. 2021;26(2):1737–1753. doi: 10.1007/s10639-020-10334-w. [DOI] [Google Scholar]
- Raitskaya & Tikhonova (2020).Raitskaya L, Tikhonova E. Overcoming cultural barriers to scholarly communication in international peer-reviewed journals | journal of language and education. Journal of Language and Education. 2020;6(2):4–8. doi: 10.17323/jle.2020.11043. [DOI] [Google Scholar]
- Ramírez-Castañeda (2020).Ramírez-Castañeda V. Disadvantages in preparing and publishing scientific papers caused by the dominance of the English language in science: the case of Colombian researchers in biological sciences. PLOS ONE. 2020;15(9):e0238372. doi: 10.1371/journal.pone.0238372. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Roussel, Achim & Auty (2021).Roussel J-R, Achim A, Auty D. Classification of high-voltage power line structures in low density ALS data acquired over broad non-urban areas. PeerJ Computer Science. 2021;7:e672. doi: 10.7717/peerj-cs.672. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Salvagno, Taccone & Gerli (2023).Salvagno M, Taccone FS, Gerli AG. Artificial intelligence hallucinations. Critical Care. 2023;27:180. doi: 10.1186/s13054-023-04473-y. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Scribano (2021).Scribano A. Emotions and politics of sensibilities in Latin America. International Sociology. 2021;36(4):585–604. doi: 10.1177/0268580921994515. [DOI] [Google Scholar]
- Seghier (2023).Seghier ML. Using ChatGPT and other AI-assisted tools to improve manuscripts readability and language. International Journal of Imaging Systems and Technology. 2023;33(3):773–775. doi: 10.1002/ima.22902. [DOI] [Google Scholar]
- Smith (2021).Smith S. Finding the voice of students engaging in online alternative provision via digital data collection methods. British Journal of Educational Technology. 2021;52(2):899–914. doi: 10.1111/bjet.13061. [DOI] [Google Scholar]
- Staudt Willet & Carpenter (2021).Staudt Willet KB, Carpenter JP. A tale of two subreddits: change and continuity in teaching-related online spaces. British Journal of Educational Technology. 2021;52(2):714–733. doi: 10.1111/bjet.13051. [DOI] [Google Scholar]
- Stokel-Walker (2023).Stokel-Walker C. ChatGPT listed as author on research papers: many scientists disapprove. Nature. 2023;613(7945):620–621. doi: 10.1038/d41586-023-00107-z. [DOI] [PubMed] [Google Scholar]
- Tharavanij (2021).Tharavanij P. Optimal book-value debt ratio. Sage Open. 2021;11(1):2158244020985788. doi: 10.1177/2158244020985788. [DOI] [Google Scholar]
- Thomas & Cassady (2021).Thomas CL, Cassady JC. Validation of the state version of the state-trait anxiety inventory in a university sample. Sage Open. 2021;11(3):21582440211031900. doi: 10.1177/21582440211031900. [DOI] [Google Scholar]
- Turnbull, Chugh & Luck (2021).Turnbull D, Chugh R, Luck J. Transitioning to E-Learning during the COVID-19 pandemic: how have higher education institutions responded to the challenge? Education and Information Technologies. 2021;26(5):6401–6419. doi: 10.1007/s10639-021-10633-w. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Unal & Uzun (2021).Unal E, Uzun AM. Understanding university students’ behavioral intention to use Edmodo through the lens of an extended technology acceptance model. British Journal of Educational Technology. 2021;52(2):619–637. doi: 10.1111/bjet.13046. [DOI] [Google Scholar]
- Van Noorden (2013).Van Noorden R. Open access: the true cost of science publishing. Nature. 2013;495(7442):426–429. doi: 10.1038/495426a. [DOI] [PubMed] [Google Scholar]
- Velitchkova (2021).Velitchkova A. Rationalization of belonging: transnational community endurance. International Sociology. 2021;36(3):419–438. doi: 10.1177/0268580920962005. [DOI] [Google Scholar]
- Weber-Wulff et al. (2023).Weber-Wulff D, Anohina-Naumeca A, Bjelobaba S, Folttfytfnek T, Guerrero-Dib J, Popoola O, Šigut P, Waddington L. Testing of detection tools for AI-generated text. International Journal for Educational Integrity. 2023;19(26) doi: 10.1007/s40979-023-00146-z. [DOI] [Google Scholar]
- Welsh & Benkhelifa (2020).Welsh T, Benkhelifa E. On resilience in cloud computing: a survey of techniques across the cloud domain. ACM Computing Surveys. 2020;53(3):59:1–59:36. doi: 10.1145/3388922. [DOI] [Google Scholar]
- Wicklum et al. (2021).Wicklum S, Willis E, Amson A, McGuire KA, Crowshoe LL, McBrien K, Henderson R. A systematic literature review of physical activity-based health programs for indigenous women: impacts on physical activity levels, obesity, and community building. Sage Open. 2021;11(2):21582440211016845. doi: 10.1177/21582440211016845. [DOI] [Google Scholar]
- Williams & Corwith (2021).Williams KM, Corwith A. Beyond bricks and mortar: the efficacy of online learning and community-building at College Park Academy during the COVID-19 pandemic. Education and Information Technologies. 2021;26(5):5055–5076. doi: 10.1007/s10639-021-10516-0. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Wilson et al. (2021).Wilson J, Ahrendt C, Fudge EA, Raiche A, Beard G, MacArthur C. Elementary teachers’ perceptions of automated feedback and automated scoring: transforming the teaching and learning of writing using automated writing evaluation. Computers & Education. 2021;168(5):104208. doi: 10.1016/j.compedu.2021.104208. [DOI] [Google Scholar]
- Wood, Najarian & Kahrobaei (2020).Wood A, Najarian K, Kahrobaei D. Homomorphic encryption for machine learning in medicine and bioinformatics. ACM Computing Surveys. 2020;53(4):70:1–70:35. doi: 10.1145/3394658. [DOI] [Google Scholar]
- Yang et al. (2021).Yang S, Lee JW, Kim H-J, Kang M, Chong E, Kim E. Can an online educational game contribute to developing information literate citizens? Computers & Education. 2021;161(4):104057. doi: 10.1016/j.compedu.2020.104057. [DOI] [Google Scholar]
- Yen, Chen & Ho (2021).Yen H-P, Chen P-C, Ho K-C. Analyzing destination accessibility from the perspective of efficiency among tourism origin countries. Sage Open. 2021;11(2):21582440211005752. doi: 10.1177/21582440211005752. [DOI] [Google Scholar]
- Yeo-Teh & Tang (2024).Yeo-Teh NSL, Tang BL. Letter to editor: NLP systems such as ChatGPT cannot be listed as an author because these cannot fulfill widely adopted authorship criteria. Accountability in Research. 2024;31(7):968–970. doi: 10.1080/08989621.2023.2177160. [DOI] [PubMed] [Google Scholar]
- Zakariás & Feischmidt (2021).Zakariás I, Feischmidt M. ‘We Are That In-Between Nation’: discourses of deservingness of Hungarian migrants working in institutions of refugee accommodation in Germany. Sociology. 2021;55(2):400–420. doi: 10.1177/0038038520941690. [DOI] [Google Scholar]
- Zave & Rexford (2020).Zave P, Rexford J. Patterns and interactions in network security. ACM Computing Surveys. 2020;53(6):118:1–118:37. doi: 10.1145/3417988. [DOI] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Supplementary Materials
Data Availability Statement
The following information was supplied regarding data availability:
The complete dataset, along with the experimental results and Python code used to analyze and visualize the findings, is available at GitHub and Zenodo:
https://github.com/ahmadrpratama/ai-text-detection-bias
ahmadrpratama. (2025). ahmadrpratama/ai-text-detection-bias: v1.0.0: Data and code for “The Accuracy-Bias Trade-Offs in AI Text Detection Tools and Their Impact on Fairness in Scholarly Publication” (PeerJ Computer Science) (v1.0.0). Zenodo. https://doi.org/10.5281/zenodo.15490003.






