Skip to main content
Indian Journal of Surgical Oncology logoLink to Indian Journal of Surgical Oncology
. 2023 Jul 3;14(4):854–858. doi: 10.1007/s13193-023-01791-z

NLP AI Models for Optimizing Medical Research: Demystifying the Concerns

Karthik Nagaraja Rao 1,, Ripu Daman Arora 2, Prajwal Dange 1, Nitin M Nagarkar 3
PMCID: PMC10767031  PMID: 38187847

Abstract

Natural language processing (NLP) AI models have gained popularity in research; however, ethical considerations are necessary to avoid potential negative consequences. This paper identifies and explores the key areas of ethical concern for researchers using NLP AI models, such as bias in training data and algorithms, plagiarism, data privacy, accuracy of generated content, prompt and content generation, and training data quality. To mitigate bias, researchers should use diverse training data and regularly evaluate models for potential biases. Proper attribution and privacy protection are essential when using AI-generated content, while accuracy should be regularly tested and evaluated. Specific and appropriate prompts, algorithms, and techniques should be used for content generation, and training data quality should be high, diverse, and updated regularly. Finally, appropriate authorship credit and avoidance of conflicts of interest must be ensured. Adherence to ethical standards, such as those outlined by ICMJE, is crucial. These ethical considerations are vital for ensuring the quality and integrity of NLP AI model research and avoiding negative consequences.

Keywords: Artificial intelligence, Medical research, Ethics, Authorship, ChatGPT

Introduction

In recent years, natural language processing (NLP) AI models have become increasingly popular in a wide range of applications. These models use machine learning algorithms to analyze and understand human language, allowing them to perform tasks such as text classification, sentiment analysis, and language translation. Among the most prominent NLP AI models are OpenAI’s ChatGPT [1], Microsoft Bing [2], and Google Bard [3]. These models have demonstrated impressive capabilities in a variety of domains, but their effectiveness and reliability are still subject to ongoing research and development. In this manuscript, we dive deep into the potential of these models for various applications, limitations, and ethical implications.

Issues Faced While Using NLP AI Model for Research (Table 1)

Table 1.

Highlights of concerns

Concern Highlights
Bias

• NLP AI models can reflect and amplify societal biases.

• Biases can be introduced through the selection of training data, or the algorithms used to train the model.

• Researchers must be aware of potential biases and take steps to mitigate them.

• Use diverse training data and evaluate the model for bias.

• Regularly test and evaluate the model for fairness and accuracy.

• Ensure that the results of the research are interpreted in an unbiased and objective manner.

• Consider the potential impact of the research on different groups of people.

AI and plagiarism

• Ensure proper attribution and citation of sources.

• Use plagiarism detection tools to check the content for similarity to existing works.

• Be mindful of ethical considerations around using AI-generated content.

• Strive to create original content that does not rely heavily on existing works.

Data privacy

• Use appropriate and ethical methods for collecting data.

• Obtain appropriate consent from individuals whose data is being used.

• Protect the privacy of individuals whose data is being used by anonymizing or de-identifying the data.

• Comply with relevant data protection regulations and guidelines.

Accuracy of generated content

• Regularly test and evaluate the accuracy of the NLP AI model.

• Use appropriate metrics to measure the accuracy of the model.

• Continuously improve the accuracy of the model through adjustments to the training data or algorithms.

• Be transparent about the limitations of the model and the potential for errors.

Prompt and content generation

• Choose appropriate prompts that are relevant to the research question.

• Ensure that the prompts are specific and unambiguous to avoid generating irrelevant content.

• Use appropriate algorithms and techniques to generate content that is relevant to the prompt.

• Regularly test and evaluate the quality of the generated content to ensure that it meets the research objectives.

Data used for training the NLP model

• Use high-quality and diverse training data that reflects the research objectives.

• Be transparent about the sources and quality of the training data.

• Regularly review and update the training data to ensure that it remains relevant and accurate.

• Consider the potential biases and limitations of the training data.

Authorship issues

• Clearly define the roles and responsibilities of each individual involved in the research.

• Ensure that all contributors are appropriately credited for their work.

• Be mindful of potential conflicts of interest or competing interests.

• Adhere to ethical standards for authorship, such as those outlined by the International Committee of Medical Journal Editors (ICMJE).

Bias

The issue of bias in NLP AI models used for research is a significant concern because it can lead to inaccurate or discriminatory results [4]. Bias can occur in several ways, including the following:

  1. Biased training data: NLP AI models are trained on large datasets, and if the dataset itself contains bias, then the resulting model will also be biased [5]. For example, if the dataset used to train a model only contains data from a particular demographic group, then the model may struggle to accurately identify and understand language from other demographic groups.

  2. Biased language use: Language itself can be biased, and certain words or phrases can have different meanings or connotations depending on the context in which they are used [5]. NLP AI models may struggle to recognize or understand these nuances, leading to inaccurate or biased results.

  3. Biased algorithmic decisions: AI models rely on algorithms to make decisions, and if the algorithms themselves contain bias, then the resulting model will also be biased [5]. For example, an algorithm that prioritizes certain words or phrases over others may lead to biased results.

To address these biases, researchers must carefully select and preprocess their training data to ensure that it is representative and unbiased. They must also regularly test their models for bias and adjust the training data or algorithms as necessary to improve the accuracy and fairness of the model. Additionally, researchers must be aware of their own biases and how they may influence the design and interpretation of their research findings. They should also strive to include diverse perspectives and voices in their research to avoid reinforcing existing biases.

AI and Plagiarism

AI models can create high-quality content quickly and efficiently, making them a valuable tool for researchers and content creators. However, the use of AI-generated content also comes with challenges, particularly regarding plagiarism.

Plagiarism is the act of using someone else’s work without giving proper credit or permission. With AI-generated content, it can be challenging to determine whether the material is original or copied from existing works [6]. AI models can generate text that is similar in structure, language, and tone to existing works, making it challenging to identify whether the content is genuinely original. To address this challenge, researchers using AI-generated content must take several steps.

  1. One should always give credit to their sources and cite any existing works that are used as a basis for the AI-generated content. Proper attribution is crucial, as it helps to acknowledge the original authors and give them the recognition they deserve. Additionally, citing sources can also help to demonstrate the legitimacy of the content.

  2. Researchers must ensure that the AI-generated content is original and not copied from existing works. This can be achieved by using plagiarism detection tools to check for similarities between the AI-generated content and other works. Plagiarism detection software can compare the content to a vast database of existing works and flag any potential matches.

  3. Lastly, researchers must be mindful of the ethical considerations of using AI-generated content. They must ensure that the content is not misleading or deceptive and that it does not harm others or perpetuate stereotypes. They must also be transparent about the use of AI-generated content and clearly communicate to readers what parts of the content were generated by AI.

Researchers must take steps to mitigate the risk of plagiarism accusations by properly citing their sources, using plagiarism detection tools, and being mindful of ethical considerations [7]. By doing so, they can ensure that they are using AI-generated content responsibly and in a way that advances their research goals without fear of plagiarism.

Data Privacy

The issue of data privacy is a significant concern for researchers using NLP AI models, particularly as these models rely heavily on large datasets of text data. When collecting and using data for NLP research, researchers must be mindful of the potential privacy risks to individuals whose data is being used. One of the main challenges with NLP AI models and data privacy is that these models are often trained on large datasets of text data that may contain personal information about individuals, such as their names, addresses, or other identifiable information [8]. This information can be used to identify individuals and can be used for nefarious purposes if it falls into the wrong hands. To address the issue of data privacy in NLP research, researchers must take several steps.

  1. One must ensure that they are using appropriate and ethical methods for collecting data, and that they have obtained appropriate consent from individuals whose data is being used. This means that researchers must clearly explain how the data will be used and must give individuals the option to opt-out of having their data collected or used for research purposes.

  2. Researchers must take steps to protect the privacy of individuals whose data is being used. This can be achieved by anonymizing or de-identifying the data before it is used for research purposes, removing any identifying information such as names, addresses, or other personal details.

  3. Finally, researchers must ensure that they are complying with relevant data protection regulations and guidelines. This means that they must be aware of the legal and ethical requirements around collecting and using data for research purposes and must take steps to ensure that they are complying with these requirements.

By taking steps to protect the privacy of individuals whose data is being used and complying with relevant regulations and guidelines, researchers can ensure that they are conducting their research in an ethical and responsible manner.

Accuracy and Authenticity of AI-Generated Content

The accuracy of the generated content is a critical consideration in using an NLP AI model for research. The accuracy of the generated content refers to how closely the content produced by the model aligns with the intended research objectives. A high level of accuracy is essential to ensure that the research results are reliable and valid [9]. There are several factors that can impact the accuracy of the generated content. One key factor is the quality of the training data used to develop the NLP AI model. If the training data is biased or incomplete, it can result in inaccurate and unreliable content being generated. Regular testing and evaluation of the model can help identify areas where the accuracy could be improved, such as by adjusting the training data or refining the algorithms used to generate content.

Measuring the accuracy of the generated content can be challenging, as there may be no clear-cut definition of what constitutes “accurate” content. However, researchers can use appropriate metrics, such as precision and recall, to measure the effectiveness of the model in generating content that aligns with the research objectives [10]. The metrics used should be appropriate for the specific research question and should be carefully chosen to ensure that they capture the most relevant aspects of the generated content. In addition to improving the accuracy of the model, researchers should also be transparent about the limitations of the model and the potential for errors. It is essential to be honest and clear about the potential for inaccuracies or limitations in the generated content to avoid misleading interpretations of the research results. Ultimately, the accuracy of the generated content is a crucial consideration for any research project that uses NLP AI models, and researchers must take appropriate steps to ensure that the content generated is as accurate as possible.

Prompt Use and Content Creation

Prompt and content generation are key aspects of using NLP AI models for research. The prompt refers to the input provided to the model, which typically takes the form of a sentence or phrase that guides the content generation process. The quality of the prompt is critical to ensure that the generated content is relevant to the research question and aligned with the intended research objectives [11]. To ensure that the prompt is appropriate, researchers should carefully consider the research question and develop a prompt that is specific, unambiguous, and relevant. This can involve breaking down the research question into sub-questions and developing prompts that address each of these questions in turn. By doing so, researchers can ensure that the generated content is focused and relevant to the research question.

The algorithms and techniques used for content generation are also critical to the quality of the generated content. There are several approaches to content generation, including rule-based models, template-based models, and neural network–based models [12]. Each approach has its strengths and weaknesses, and the choice of approach will depend on the specific research question and the desired outcomes. Once the content is generated, it is important to evaluate its quality to ensure that it aligns with the intended research objectives. This can involve using appropriate metrics to measure the relevance and accuracy of the generated content, as well as manual evaluation by subject matter experts. Regular testing and evaluation of the generated content can help identify areas for improvement and ensure that the content remains relevant and aligned with the research objectives.

Overall, prompt and content generation are critical components of using NLP AI models for research. By developing appropriate prompts and using appropriate algorithms and techniques for content generation, researchers can ensure that the generated content is relevant, accurate, and aligned with the research objectives.

Data Used for Training the NLP Model

The data used for training an NLP AI model is a critical consideration in using such models for research. The quality and quantity of the training data can impact the accuracy and effectiveness of the model in generating content that aligns with the research objectives. There are several factors to consider when selecting the training data, including the source of the data, the size of the data set, and the quality of the data. The source of the data can impact its relevance to the research question, and researchers should consider using data that is specific to the research domain or topic. The size of the data set is also critical, as larger data sets can help improve the accuracy of the model [13]. The quality of the data is another important consideration. The training data should be representative of the intended research population and should be free from biases or errors.

Researchers should carefully evaluate the quality of the data and take appropriate steps to address any issues, such as by cleaning the data or removing any irrelevant or misleading data points. In addition to selecting appropriate training data, researchers should also consider the ethical implications of using such data. This can involve ensuring that the data is obtained and used in an ethical and legal manner, such as by obtaining appropriate consent from participants and protecting their privacy. Regular testing and evaluation of the model can help identify areas where the training data may need to be adjusted or refined to improve the accuracy and effectiveness of the model. It is important to remember that the quality of the training data is an ongoing concern and should be regularly evaluated and updated as needed to ensure that the model remains effective and aligned with the research objectives.

Authorship Issues

Authorship issues can arise when using NLP AI models for research, particularly in cases where the generated content is used for publication or dissemination. This can involve questions around who should be credited as the author of the generated content, as well as concerns around intellectual property and ownership. One potential approach to addressing authorship issues is to consider the role of the human researcher in guiding the content generation process. While the AI model may be responsible for generating the content, the human researcher can play a critical role in developing the prompt and evaluating the quality of the generated content. As such, it may be appropriate to credit both the AI model and the human researcher as authors of the generated content [14]. Another approach to addressing authorship issues is to consider the intended use of the generated content. In cases where the content is intended for academic or research purposes, it may be appropriate to credit the AI model as the author of the content, with appropriate acknowledgements to the source of the training data and the human researcher’s role in guiding the content generation process.

However, in cases where the generated content is intended for commercial or other purposes, such as marketing or advertising, it may be necessary to consider the ownership and intellectual property rights associated with the content. This can involve ensuring that appropriate licenses or permissions are obtained for any copyrighted material used in the training data, as well as ensuring that any trademarks or other intellectual property associated with the content are appropriately protected [15]. Ultimately, addressing authorship issues when using NLP AI models for research requires careful consideration of the intended use of the generated content and the role of the human researcher in guiding the content generation process. By carefully evaluating these factors and taking appropriate steps to ensure intellectual property and ownership rights are protected, researchers can effectively navigate authorship issues and use NLP AI models to generate high-quality content for a variety of research and commercial purposes.

Conclusion

We emphasize the need for caution and consideration when using these models, especially when it comes to accuracy of generated content and ethicality of its use. Researchers can use the models to generate outlines and headings, check grammar and phrasing, and improve efficiency. It is important to acknowledge and disclose the use of NLP AI models in research, including appropriate authorship and prompt disclosure in the manuscript. Ultimately, researchers must ensure that the generated content is fact-checked and appropriately referenced before use.

Declarations

Conflict of Interest

The authors declare no competing interests.

Footnotes

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

References


Articles from Indian Journal of Surgical Oncology are provided here courtesy of Springer

RESOURCES