Skip to main content
PLOS Digital Health logoLink to PLOS Digital Health
. 2024 Sep 19;3(9):e0000607. doi: 10.1371/journal.pdig.0000607

Ethical, legal, and social issues (ELSI) and reporting guidelines of AI research in healthcare

Junko Kameyama 1, Satoshi Kodera 2, Yusuke Inoue 1,3,*
Editor: Dhiya Al-Jumeily O B E4
PMCID: PMC11412540  PMID: 39298409

This paper aims to (1) summarize the descriptions of ethical, legal, and social issues (ELSI) in major reporting guidelines, identify trends and tendencies, and (2) highlight issues for future researchers and guideline creators.

Medical care is a fundamental aspect of human life, closely linked to their health and well-being. Advances in the research and development (R&D) of medical artificial intelligence (AI) are anticipated to yield numerous benefits. However, concerns regarding incorrect AI design and usage, causing various ethical and human-rights issues for both individuals and society, do arise [14]. Policies for research efforts in developing AI have been proposed; however, no consensus on approach selection has been attained [4]. Therefore, this study focuses on “reporting guidelines,” which provide specific result-publication instructions for researchers. The guidelines provide a basis for AI deployment and implementation through the increased transparency and evaluability of processes and data management during R&D [5]. In recent years, AI design and usage have been closely related to ELSI [4]. ELSI is the examination of ethical, legal, and social issues raised by the deployment of new knowledge. This perspective focuses on the impact of new scientific methods and knowledge on the current and future generations. While questioning the role of researchers in ELSI, the position of reporting guidelines requires changes.

Therefore, we review recent reporting guidelines for medical AI research and adopt guidelines with a “checklist” of specific initiatives as the selection criteria. First, six reports registered in the Enhancing the Quality and Transparency of Health Research (EQUATOR) network were selected. The EQUATOR network is an international initiative seeking improvements to the credibility and value of published health-research literature by promoting transparent and accurate reports and disseminating robust reporting guidelines [6]. Furthermore, two other reports from major literature databases (Web of Science and PubMed) were identified and added for a total of eight reporting guidelines [714]. While organizing these guidelines from an ELSI perspective, the 11 items of the “Considerations for AI developers,” which append the WHO guidance “Ethics and governance of artificial intelligence for health,” were used [Table 1].

Table 1. Ethical, legal, social design, development, and deployment considerations for artificial intelligence (AI) research and health development: Notes for AI developer.

AI design AI development AI deployment
I Clarify the objectives VII Identify regulatory requirements X Engage and educate multiple stakeholders for deployment and maintenance
II Engage multiple stakeholders and understand contexts VIII Establish data management plans XI Evaluate and improve performance
III Define relevant ethical issues through consultation IX Adopt standards and best practices
IV Assess risks
V Address biases
VI Privacy by design and privacy by default

Eleven items are presented as points to keep AI developers in mind. In accordance with the WHO guidelines, the AI research and development process for health is classified into three stages: “AI design,” “AI development,” and “AI deployment,” with relevant ethical considerations at the bottom of each stage.

Ethics and governance of artificial intelligence for health: WHO guidance [4] Annex, Considerations for the ethical design, development and use of artificial intelligence for health “Considerations for AI developers” is prepared by revising

A minimum and maximum of 14 and 58 (average of 25) items were included in the checklists of the 8 reporting guidelines, respectively. "✓" was marked against each reference in the checklist that corresponded to the 11 items proposed by WHO. Additionally, "(✓)" was marked for reports that were not mentioned in the checklist but were mentioned in the text that provided background information and review-process details. For no corresponding descriptions, the checklist was left blank. JK and YI

performed this work, and after both verified the consistency, the results are listed in Table 2. The obtained results, including the clinician’s perspective on SK, were examined, and two major issues were identified.

Table 2. ELSI perspectives on reporting guidelines for the R&D of AI technology in healthcare.

Reporting guideline acronym
/Date of publication
Targeted
readers
Study design Medical
specialization
ELSI*
AI design AI development AI deployment
I II III IV V VI VII VIII IX X XI
CLAIM
2020/05
Authors and reviewers Diagnostic accuracy studies Multi-fields
CONSORT-AI
2020/09
Investigators, editors, and peer reviewers Randomised controlled trials
(Interventional Study Design)
Multi-fields
SPIRIT-AI
2020/09
External reviewers Clinical Trial Protocol Multi-fields
MI-CLAIM
2020/09
Algorithm designers, repository
managers, manuscript writers and readers,
editors, and model users
AI modelling Multi-fields
CAIR
2021/05
Clinicians Orthopedics (✓) (✓) (✓) (✓)
CLEAR Derm
2021/12
Developers and reviewers Diagnostic Accuracy Dermatology
DECIDE-AI
2022/05†
Multiple stakeholders‡ Evaluation of human factors in early algorithm deployment Multi-fields
CLEAR
2023/05
Authors and reviewers Diagnostic Accuracy Multi-fields

* From ethics and governance of artificial intelligence for health: WHO guidance [4] Annex, Considerations for the ethical design, development and use of artificial intelligence for health “Considerations for AI developers,” the following 11 items were used in the analysis as ELSI perspectives.

I.Clarify the objectives II.Engage multiple stakeholders and understand contexts III.Define relevant ethical issues through consultation IV.Assess risks V.Address biases VI.Privacy by design and privacy by default VII.Identify regulatory requirements VIII.Establish data management plans IX.Adopt standards and best practices X.Engage and educate multiple stakeholders for deployment and maintenance XI.Evaluate and improve performance

† Three reports are documented over 2021–2023; the 2022 report is the central document that proposes the checklist.

‡ Stakeholders involved in the development of the guidelines. There are no specifications for the targeted reader

1. The necessity of efforts for further practical guidance in cooperation with society and patients

Prominently, “V. Address biases” (seven guidelines) were mentioned in the reporting guidelines. However, the purpose and awareness of the issue were unclear in several studies. Recently, the issues of bias surrounding AI have garnered increasing attention in addition to scientific imperatives, owing to its impact on ELSI. For example, biases in data, design and interpretation, and beneficiary exist [4,15]. Considering the diverse concerns surrounding bias, further enhancements are required to support the efforts of researchers.

In addition, as shown in Table 2, an increasingly mentioned item in the guidelines is “III. Defining relevant ethical issues through consultation,” which includes patient and public involvement (PPI) awareness. Precisely, a strong sense of crisis has arisen regarding ELSI, such that the design bias of medical AI and data may directly develop into discrimination and disparities in the recipients of health care [15]. The same purpose is applicable for “II. Engage multiple stakeholders and understand the context.” and “XI. Evaluate and improve performance.” Researchers are required to recognize the growing concerns and incorporate the perspective of potential medical AI targets in their R&D processes. To help researchers smoothly develop such initiatives, highly practical guidance content from guideline creators is needed.

2. Limitations of guidelines and the need for researchers to take the initiative

In the AI reporting guidelines for R&D, items with few checks were identified; overall, ELSI items lagged behind while being addressed. Particularly, “VI. Privacy by Design and Privacy by Default” and “VII. Identify regulatory requirements” were significant, and both were met only in one or two guidelines.

Despite the large amount of detailed personal data handled in AI research, reporting guidelines did not include a section addressing data-protection measures. Globally, personal data protection in research activities is highly concerning, and reporting guidelines require the consideration of issues and alerting researchers. However, the legal requirements for personal-information protection may differ from country to country, and reporting guidelines may offer limitations in providing detailed guidance on the matter for individual research reports. While addressing privacy and legal requirements, researchers should not solely rely on the reporting guidelines but should consider proactive consultation with experts and ethical review committees, such that the legal and ethical measures remain relevant to their plans. The point “IX. Adopt standards and best practices” is as mentioned above, and the lack of reference to best practices in R&D was assumed as a limitation of the guidelines for reporting papers.

ELSI surrounding medical AI are evolving and becoming more inclusive, even in the early stages of R&D. The "Reporting Guidelines" not only present the methods to write papers, but also consider the meaning of ELSI guidelines that link researchers, medical care, and patients. A continuous review of reporting guidelines is essential to ensure that medical AI continues to develop with the support of people.

Moreover, as a next step, it is also important to promote harmonization among guidelines, and for reporting guidelines to be linked to discussions on AI practice in the clinical aspect. In doing so, we hope that the ELSI aspect of the discussion, as described in this paper, will be developed in a way that includes the perspectives of many stakeholders.

Supporting information

S1 File. Relevant sections of the guidelines from which the decisions in Table 2 are based.

"✓" was marked against each reference in the checklist that corresponded to the 11 items proposed by WHO. Additionally, "(✓)" was marked for reports that were not mentioned in the checklist but were mentioned in the text that provided background information and review-process details. For no corresponding descriptions, the checklist was left blank. JK and YI performed this work.

(PDF)

pdig.0000607.s001.pdf (252.6KB, pdf)

Funding Statement

The authors received no specific funding for this work.

References

  • 1.van Kolfschooten H, Shachar C. The Council of Europe’s AI Convention (2023–2024): Promises and pitfalls for health protection. Health Policy. 2023;138(1):104935. doi: 10.1016/j.healthpol.2023.104935 [DOI] [PubMed] [Google Scholar]
  • 2.Saw SN, Ng KH. Current challenges of implementing artificial intelligence in medical imaging. Phys Med. 2022;100:12–17. doi: 10.1016/j.ejmp.2022.06.003 [DOI] [PubMed] [Google Scholar]
  • 3.Rajpurkar P, Chen E, Banerjee O, Topol EJ. AI in health and medicine. Nat Med. 2022;28(1):31–38. doi: 10.1038/s41591-021-01614-0 [DOI] [PubMed] [Google Scholar]
  • 4.World Health Organization. Ethics and governance of artificial intelligence for health: WHO guidance. Geneva: World Health Organization; 2021. [Google Scholar]
  • 5.Shelmerdine SC, Arthurs OJ, Denniston A, Sebire NJ. Review of study reporting guidelines for clinical studies using artificial intelligence in healthcare. BMJ Health Care Inform. 2021;28(1):e100385. doi: 10.1136/bmjhci-2021-100385 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6.Network EQUATOR. Enhancing the QUAlity and Transparency Of health Research [Internet]. [cited 2023. Nov 15]. Available from: https://www.equator-network.org/ [Google Scholar]
  • 7.Mongan J, Moy L, Kahn CE Jr. Checklist for artificial intelligence in medical imaging (CLAIM): A guide for authors and reviewers. Radiol Artif Intell. 2020;2(2):e200029. doi: 10.1148/ryai.2020200029 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.Liu X, Rivera SC, Moher D, Calvert MJ, Denniston AK, et al. Reporting guidelines for clinical trial reports for interventions involving artificial intelligence: the CONSORT-AI extension. Lancet Digital Health. 2020;2(10):e537–e548. doi: 10.1016/S2589-7500(20)30218-1 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9.Rivera SC, Liu X, Chan AW, Denniston AK, Calvert MJ, et al. Guidelines for clinical trial protocols for interventions involving artificial intelligence: the SPIRIT-AI Extension. Lancet Digital Health. 2020;2(10):e549–e560. doi: 10.1016/S2589-7500(20)30219-3 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10.Norgeot B, Quer G, Beaulieu-Jones BK, Torkamani A, Dias R, Gianfrancesco M, et al. Minimum information about clinical artificial intelligence modeling: the MI-CLAIM checklist. Nat Med. 2020;26(9):1320–1324. doi: 10.1038/s41591-020-1041-y [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11.Olczak J, Pavlopoulos J, Prijs J, Ijpma FFA, Doornberg JN, Lundström C, et al. Presenting artificial intelligence, deep learning, and machine learning studies to clinicians and healthcare stakeholders: An introductory reference with a guideline and a Clinical AI Research (CAIR) checklist proposal. Acta Orthop. 2021;92(5):513–525. doi: 10.1080/17453674.2021.1918389 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12.Daneshjou R, Barata C, Betz-Stablein B, Celebi ME, Codella N, Combalia M, et al. Checklist for evaluation of image-based artificial intelligence reports in dermatology CLEAR Derm Consensus Guidelines from the International Skin Imaging Collaboration Artificial Intelligence Working Group. JAMA Dermatol. 2022;158(1):90–96. doi: 10.1001/jamadermatol.2021.4915 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.Vasey B, Nagendran M, Campbell B, Clifton DA, Collins GS, Denaxas S, et al. Reporting guideline for the early-stage clinical evaluation of decision support systems driven by artificial intelligence: DECIDE-AI. BMJ. 2022;377:e070904. doi: 10.1136/bmj-2022-070904 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14.Kocak B, Baessler B, Bakas S, Cuocolo R, Fedorov A, Maier-Hein L, et al. CheckList for EvaluAtion of Radiomics research (CLEAR): a step-by-step reporting guideline for authors and reviewers endorsed by ESR and EuSoMII. Insights Imaging. 2023;14(1):75. doi: 10.1186/s13244-023-01415-8 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15.Schwartz R, Vassilev A, Greene K, Perine L, Burt A, Hall P. Towards a Standard for Identifying and Managing Bias in Artificial Intelligence. NIST Special Publication 1270. Gaithersburg: National Institute of Standards and; Technology; 2022. [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

S1 File. Relevant sections of the guidelines from which the decisions in Table 2 are based.

"✓" was marked against each reference in the checklist that corresponded to the 11 items proposed by WHO. Additionally, "(✓)" was marked for reports that were not mentioned in the checklist but were mentioned in the text that provided background information and review-process details. For no corresponding descriptions, the checklist was left blank. JK and YI performed this work.

(PDF)

pdig.0000607.s001.pdf (252.6KB, pdf)

Articles from PLOS Digital Health are provided here courtesy of PLOS

RESOURCES