Skip to main content
Bulletin of the World Health Organization logoLink to Bulletin of the World Health Organization
. 2020 Jan 27;98(4):257–262. doi: 10.2471/BLT.19.237289

How to achieve trustworthy artificial intelligence for health

Garantir la fiabilité de l'intelligence artificielle dans le secteur de la santé

Cómo lograr una inteligencia artificial confiable para la salud

كيفية تحقيق ذكاء اصطناعي جدير بالثقة للصحة

如何实现可信赖的医疗人工智能

Как обеспечить надежное и безопасное использование искусственного интеллекта в сфере здравоохранения

Kristine Bærøe a,, Ainar Miyata-Sturm b, Edmund Henden b
PMCID: PMC7133476  PMID: 32284649

Abstract

Artificial intelligence holds great promise in terms of beneficial, accurate and effective preventive and curative interventions. At the same time, there is also awareness of potential risks and harm that may be caused by unregulated developments of artificial intelligence. Guiding principles are being developed around the world to foster trustworthy development and application of artificial intelligence systems. These guidelines can support developers and governing authorities when making decisions about the use of artificial intelligence. The High-Level Expert Group on Artificial Intelligence set up by the European Commission launched the report Ethical guidelines for trustworthy artificial intelligence in2019. The report aims to contribute to reflections and the discussion on the ethics of artificial intelligence technologies also beyond the countries of the European Union (EU). In this paper, we use the global health sector as a case and argue that the EU’s guidance leaves too much room for local, contextualized discretion for it to foster trustworthy artificial intelligence globally. We point to the urgency of shared globalized efforts to safeguard against the potential harms of artificial intelligence technologies in health care.

Introduction

Artificial intelligence has been defined as “the part of digital technology that denotes the use of coded computer software routines with specific instructions to perform tasks for which a human brain is normally considered necessary.”1 The most complex connotation of the term artificial intelligence, that of machines with human-like general intelligence, is still a distant vision. However, artificial intelligence in the more restricted sense defined above is already broadly embedded in society in a variety of forms. The pace of development of new and improved artificial intelligence-based technologies is rapid; the question is no longer whether artificial intelligence will have an impact, but “by whom, how, where, and when this positive or negative impact will be felt.”2

Many areas of health care could benefit from the use of artificial intelligence technology. According to a recent literature review, artificial intelligence is already being used: “(1) in the assessment of risk of disease onset and in estimating treatment success [before] initiation; (2) in an attempt to manage or alleviate complications; (3) to assist with patient care during the active treatment or procedure phase; and (4) in research aimed at elucidating the pathology or mechanism of and/or the ideal treatment for a disease.”3 On the risk side, others have summarized several health-related concerns: the potential for bias in the data used to train artificial intelligence algorithms; the need for protection for patients’ privacy; potential mistrust of digital tools by clinicians and the general public; and ensuring health-care personnel handle artificial intelligence in a trustworthy manner.4 Other concerns relate to physical applications of artificial intelligence. For example, while robots could be useful in the care of the elderly, there are risks of reduced contact between humans, the deception of encouraging companionship with a machine and loss of control over a person’s own life.5 Questions have also been raised about the extent to which artificial intelligence technologies could replace clinicians6 and, if so, whether the opacity of machine learning-based decisions weaken the authority of clinicians, threaten patients’ autonomy7 or jeopardize shared decision-making between doctor and patient.8

Discussions of the risks posed by artificial intelligence systems range from current concerns, such as violations of privacy or harmful effects on society, to debates about whether machines could ever escape from human control. However, fully predicting the consequences of these technological developments is not possible. The need for a precautionary approach to artificial intelligence highlights the importance of thoughtful governance. By applying our human intelligence, we have the opportunity, through control of decision-making, to steer the development of artificial intelligence in ways that accord with human values and needs.

Guidance has been developed through initiatives that aim to foster responsible and trustworthy artificial intelligence and to mitigate unwanted consequences. Examples include AI4People,2 Asilomar AI principles9 and the Montreal Declaration for a Responsible Development of Artificial Intelligence.10 In this paper we focus on the report of the independent High-Level Expert Group on Artificial Intelligence set up by the European Commission.11 Ethical guidelines for trustworthy AI identified trustworthy artificial intelligence as consisting of three components: (i) compliance with all applicable laws and regulations; (ii) adherence to ethical principles and values; and (iii) promotion of technical and societal robustness. These components are important throughout the cycle of development, deployment and use of artificial intelligence.

The expert group’s report focuses on the ethics and robustness of artificial intelligence rather than the legal issues, basing ethical guidance for trustworthy artificial intelligence on a fundamental rights approach.11 Four principles rooted in these fundamental rights shape the framework and are translated into more specific requirements: (i) respect for human autonomy; (ii) prevention of harm; (iii) fairness; and (iv) explicability (the report stresses that this list is not necessarily exhaustive). These requirements can translate into a tailored list to allow for assessments of specific artificial intelligence interventions (Fig. 1). The first three principles are well-established in the bioethical literature.12 The principle of explicability is intended to gain an understanding of how artificial intelligence generates output, which is important for contesting decisions based on artificial intelligence and tracing appropriate chains of accountability.2

Fig. 1.

Framework for trustworthy artificial intelligence

Source: Adapted from the report of the European Commission High-Level Expert Group on Artificial Intelligence.9

Fig. 1

These general ethical principles may, however, conflict with each other. Any conflicts should be managed by deliberations that are accountable to and conditioned by democratic systems of public engagement and processes of open political participation.11 The expert group’s report strongly emphasized the need for responsible governance and the role of stakeholders “including the general public.” The safety mechanisms proposed include a “basic artificial intelligence literacy [that] should be fostered across society.” Building trustworthiness into artificial intelligence therefore relies on assumptions about a high-functioning, deliberative democracy and its governing potential to drive the development of artificial intelligence-based technologies. The expert group also highlighted the context-specific nature of artificial intelligence (“different situations raise different challenges”) and the need for an additional sectoral approach to the general framework they propose. As part of a coordinated artificial intelligence strategy for the European Commission and European Union (EU) Member States, the report’s recommendations are expected to be central in shaping the development and use of artificial intelligence in Europe.

Since the use and impact of artificial intelligence spans national borders, the expert group calls for work towards a global ethical guidance.11 We believe that the ethical concerns and challenges addressed by the EU’s framework have global relevance. Basing the principles on a fundamental rights approach means that their relevance and significance can be considered universal. Indeed, the principles are rooted in the same rights and obligations that structure most of the United Nations’ (UN) sustainable development goals (SDGs) and that influence development strategies in low-and middle-income countries beyond the EU. Thus, the content of the ethical framework (principles and guidelines) fronted by the expert group can be expected to carry legitimacy across cultural contexts and economic divides. The risks and potential negative impact artificial intelligence systems can have, for example “on democracy, the rule of law and distributive justice, or on the human mind”11 are also applicable wherever such institutions are in place.

The EU expert group’s framework is a process of first identifying fundamental ethical principles that are acknowledged through public debate as relevant for different contexts and across domains of analysis and then translating these principles into “viable guidelines to shape artificial intelligence-based innovation.”13 However, while an ethical framework can recommend a process for how to resolve conflicting ethical principles in real-world situations, it cannot provide concrete, practical solutions for specific contexts. Moreover, ethical frameworks may not provide guidance on how to deal with conflicts that can occur between realizing the aim of the framework itself (such as ethically justified artificial intelligence) and realization of other goods (such as economic growth or achieving the SDGs). In the following section, we discuss how such conflicts occur in the health-care arena in ways that undermine the trustworthiness assumed by the expert group’s framework. Finally, we outline how a global initiative emerging from within various sectors can constructively promote globally trustworthy applications of artificial intelligence.

Ethical challenges

Despite its global relevance, the EU expert group’s framework may fail to provide trustworthy safeguards for the use of artificial intelligence in all health settings. We reflect here on how general, structural features of global health systems, general human motivation and known drivers of interest for actions might together impact on the development and implementation of artificial intelligence systems. We have identified five areas of ethical concern: (i) conflicting goals; (ii) unequal contexts; (iii) risk and uncertainty; (iv) opportunity costs; and (v) democratic deficits (Box 1). These distinct concerns, when combined, demonstrate the need to foster trustworthy development, deployment and use of artificial intelligence as an explicitly global and transnational endeavour.

Box 1. Ethical challenges for the global development and implementation of artificial intelligence systems in health care.

Conflicting goals

Forces, such as the economic interests of the artificial intelligence industry and the political objective of the United Nations’ sustainable development goals, can work against the promotion of ethically safe artificial intelligence technologies.

Unequal contexts

Unequal contextual factors across countries create different bases for the ethical assessment of acceptable employment of artificial intelligence and may thus sustain a non-universal standard and inequitable quality of health care across borders.

Risk and uncertainty

In lower-income countries with challenging living conditions, promises of effective artificial intelligence solutions that can improve the situation could override precautionary concerns about the potential risks.

Opportunity costs

The opportunity costs of replacing human intelligence with artificial intelligence has implications for the experiences that citizens bring into the political debate and the organization of powers and political institutions.

Democratic deficits

Many countries do not have sufficiently high-functioning, deliberative democracies in place and lack the ability to adequately manage and control the precautionary risks and societal impact assessments of artificial intelligence.

Conflicting goals

The health sector is affected by strong forces of global political governance, as exemplified by SDG 3 to: “ensure healthy lives and promote well-being for all ages,” and target 3.8 to: “achieve universal health coverage, including financial risk protection, access to quality essential health-care services and access to safe, effective, quality and affordable essential medicines and vaccines for all.”14 These global forces that shape local priority-setting in health care may, however, represent conflicting goals. For example, the goals of equality of access to care and equality of care quality are not inherently connected and can conflict with each other when implemented.15 Another example is the efficient use of resources when deciding what to include in universal health coverage. Governance of health care such that it meets all political and cost–effectiveness aims inevitably leads to trade-offs and priority-setting. Such trade-offs become more difficult when resources are scarcer.

Application of artificial intelligence can be more cost–effective than human labour. The call for cost–effective, un-biased, equality-promoting solutions in the health sector can therefore be seen as an open invitation for constructive cooperation with the artificial intelligence industry. Tools already being discussed include personal care robots,5 ambient assisted living technologies16 and humanoid nursing robots.17 A potential challenge, however, is that such machines might not be able to provide the same overall quality of care for everyone, as do interactions with human beings. Increasing accessibility by implementing these more cost–effective methods will not solve the issue of equality if the quality of artificial intelligence-based interactions is worse than human interactions.

Another consequence of striving to reach political goals needs to be addressed. The global challenge of achieving universal access to health care provides the artificial intelligence industry with an opportunity not only to promote a societal good, but also to make favourable investments. Yet if the profits are not fairly distributed, these investments will only benefit the artificial intelligence industry economically and thereby contribute to accumulated wealth for some. The inequalities in living conditions that are associated with health inequalities may therefore persist or even worsen, thus undermining the political goal of ensuring healthy lives. Redistributing the economic benefits of large-scale investments in artificial intelligence in public health could be one way of compensating for the adverse determinants of health and therefore a strategy for promoting overall health equality. Transnational regulations are, however, required for such a redistribution to occur in a systematic manner within and across higher- and lower-income countries.

Finally, the demand for policy decisions in the health sector to be based on empirical evidence is also a force that may influence decisions to employ artificial intelligence technologies in the health-care sector and may distract from the risks associated with artificial intelligence technologies. The empirical evidence on which to base future risk calculations and assessments of new technologies may not be available before the opportunity to implement safeguards (such as regulations) has passed. Even when there are attempts to consider the uncertain, long-term impacts of artificial intelligence, the uncertain conclusions may be traded-off for clearly effective, short-term solutions to important problems.

Unequal contexts

The conditions in which people live, and therefore the determinants of their health, vary across countries. Many countries will struggle to find the resources to address the SDGs and they will have to achieve the greatest possible health benefit out of the funds they have. In lower-income countries life expectancy is increasing due to decreases in infectious diseases.18 The prevalence of noncommunicable diseases associated with older age will likely increase. Cost–effective artificial intelligence technologies that can help reach, screen, diagnose, prescribe treatment for and even care for such patients will be invaluable where there are insufficient resources to increase the health workforce to meet the demographic challenge. On the other hand, in a high-income country with a well-developed, publicly funded health-care system introducing artificial intelligence-based methods to, for example, follow up the day-to-day social and nutritional care of elderly people could create other concerns. For example, the artificial intelligence system might replace an established, well-functioning workforce, which raises the concern that something valuable is lost in that transition. Unequal contextual factors create a different basis for the ethical assessment of the appropriateness of artificial intelligence. If applications of artificial intelligence actually provide poorer quality care for the elderly than does human intelligence, then there is a risk of accepting different ethical standards for higher- and lower-income countries within the same ethical framework and thereby sustaining inequitable quality of health care across borders.

Risk and uncertainty

SDG 3 establishes an urgency to the goals of promoting health and reducing health inequalities, while measuring countries on how they perform contributes another layer of motivation. Yet being willing to risk more to achieve aims to which strong values are attached creates a structural dilemma in the area of health and artificial intelligence ethics. When resources are scarce, there might be a willingness to discount potential future harms. For example, implementing resource-efficient digital tools to monitor the movements of people with dementia could be seen as a step towards greater surveillance of society in the future. When people are in need of health care, concerns about the uncertain, potentially problematic, long-term impact of receiving help from an artificial intelligence-based system might not be their main priority. On a political level, precautionary thinking about highly uncertain future impacts may be ignored in favour of an effective solution, which helps to solve a national health challenge. Furthermore, tolerance towards the potential unwanted consequences of implementing artificial intelligence will likely depend on how intolerable the current state of affairs is perceived to be.

Opportunity costs

In the health-care sector, we need to consider the opportunity cost of not implementing potentially beneficial artificial intelligence technology.2 There are, however, specific opportunity costs of choosing artificial intelligence at the expense of human intelligence. The implementation of artificial intelligence in health care might gradually replace or complement functions previously performed by human intelligence, such as exercising clinical judgement and providing assistance for those who need the help of others. The consequences for the workforce due to such replacement must be considered on the path to the SDGs. The input of human judgement will move further up the chain of decision-making, even when the EU’s ethical guidance is adhered to and artificial intelligence is overseen and controlled by human beings. The developers and managers of artificial intelligence will therefore become the authorities on the value and trade-offs involved in artificial intelligence decisions. This transition into a more centralized control of health care might create less autonomy for the remaining health-care professionals and thus negatively impact their own health. Moreover, if human-to-human interactions are replaced by human-to-artificial intelligence interactions, opportunities for the valuable, health-promoting benefits of human interactions, such as emotional intimacy, reassurance and affirmation of self-worth through others, will decrease.

Another concern is that a decrease in the health-care workforce means having fewer people (as stakeholders) who can feed their experiences with fellow humans’ social and health-care needs into processes of public, political deliberation. Open democratic discussions are important safeguards against the undesirable outcomes of artificial intelligence technology. A feedback loop is therefore created wherein increasing implementation of artificial intelligence methods leads to even weaker safeguards. Loss of the educated workforce in health (and other sectors) means that the key elements of the deliberative process to establish safety mechanisms for artificial intelligence technologies can therefore be lost.

All the above are potential opportunity costs of implementing artificial intelligence at the expense of a workforce that is driven by individual human judgement. Such lost opportunities must be identified and discussed as part of a global, self-reflective, trustworthy artificial intelligence strategy. This is in addition to the risks discussed in the EU’s expert group report.

Democratic deficits

The EU expert group’s guidance on trustworthy artificial intelligence is based on the assumptions that high-functioning democracies and societies are present to administer and counteract or control any undesirable outcomes of artificial intelligence systems. A potential challenge, however, is that these safeguarding mechanisms and governing powers may not be in place in all countries. Also, how best to organize stakeholder involvement in the health sector is still being debated19,20 and current deliberations over artificial intelligence could be a new path for developing such governing institutions. Developing safety mechanisms based on particular cultural and social traditions for organizing and managing political issues would, however, be a time-consuming process. The development of artificial intelligence and the forces that drive it cannot be slowed to the pace of these deliberations.

A related, structural danger (which is not unique to artificial intelligence) is that the impact of artificial intelligence technology developed without safety restrictions in one country might affect other countries. An artificial intelligence health tool that would not even be considered in the design laboratory of one society could, however, be placed on the agenda of policy-making debates simply because it exists as an available option elsewhere.

A global approach

The artificial intelligence industry is driven by strong economic and political interests. Gaining trustworthy control over the potential risks and harms related to artificial intelligence is therefore crucial. The EU expert group’s framework is designed to translate general principles into more concrete guidance and recommendations for how to address artificial intelligence. However, the framework does not address threats to the attainment of trustworthy artificial intelligence embedded in real-world interests and complex circumstances. Securing the governance of trustworthy artificial intelligence technologies, locally and globally, in health and other sectors, will have to be based on an expanded understanding of what translation of ethical norms into practice requires by addressing and managing structural concerns as those we have identified.

More concretely, there is a need for transnational development of shared, explicitly articulated rules that are context-independent, rather than for a framework that is too context-specific, at least before there has been a chance to develop local, protected political institutions. Low-income nations might be deterred from implementing cost–effective, but potentially unsafe artificial intelligence technologies to solve their short-term problems. As part of a global endeavour, high-income nations bear a responsibility to compensate for the potential losses to these countries, for example by financially supporting education of a scaled-up workforce of health-care personnel. The sector-specific challenges, as pointed out by the EU’s report and highlighted by our analysis, mean that targeted translation of shared general principles into specific, global regulations could guard against the potential dangers of artificial intelligence-based technology. The World Health Organization, together with the other UN bodies, is well-placed in the field of health, to lead such shared efforts towards globally trustworthy artificial intelligence.

Competing interests:

None declared.

References

  • 1.Zandi D, Reis A, Vayena E, Goodman K. New ethical challenges of digital technologies, machine learning and artificial intelligence in public health: a call for papers. Bull World Health Organ. 2019;97(1):2 10.2471/BLT.18.227686 [DOI] [Google Scholar]
  • 2.Floridi L, Cowls J, Beltrametti M, Chatila R, Chazerand P, Dignum V, et al. AI4People – an ethical framework for a good artificial intelligence society: opportunities, risks, principles, and recommendations. Minds Mach (Dordr). 2018;28(4):689–707. 10.1007/s11023-018-9482-5 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3.Becker A. Artificial intelligence in medicine: what is it doing for us today? Health Policy Technol. 2019;8(2):198–205. 10.1016/j.hlpt.2019.03.004 [DOI] [Google Scholar]
  • 4.Reddy S, Allan S, Coghlan S, Cooper P. A governance model for the application of AI in health care. J Am Med Inform Assoc. 2019. November 4. pii: ocz192. 10.1093/jamia/ocz192 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5.Sharkey A, Sharkey N. Granny and the robots: ethical issues in robot care for the elderly. Ethics Inf Technol. 2012;14(1):27–40. 10.1007/s10676-010-9234-6 [DOI] [Google Scholar]
  • 6.Bluemke DA. Radiology in 2018: are you working with artificial intelligence or being replaced by artificial intelligence? Radiology. 2018. May;287(2):365–6. 10.1148/radiol.2018184007 [DOI] [PubMed] [Google Scholar]
  • 7.Grote T, Berens P. On the ethics of algorithmic decision-making in healthcare. J Med Ethics. 2019. November 20. pii: medethics-2019-105586. 10.1136/medethics-2019-105586 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.McDougall RJ. Computer knows best? The need for value-flexibility in medical artificial intelligence. J Med Ethics. 2019. March;45(3):156–60. 10.1136/medethics-2018-105118 [DOI] [PubMed] [Google Scholar]
  • 9.Asilomar AI principles [internet]. Cambridge, Massachusetts: Future of Life Institute; 2017. Available from: https://futureoflife.org/ai-principles [cited 2020 Jan 21].
  • 10.Montreal Declaration for a Responsible Development of Artificial Intelligence [internet]. Montreal: University of Montreal; 2018. Available from: https://www.montrealdeclaration-responsibleai.com/the-declaration [cited 2020 Jan 21].
  • 11.Independent High-Level Expert Group on Artificial Intelligence. Ethics guidelines for trustworthy AI. Brussels: European Commission; 2019. Available from: https://ec.europa.eu/futurium/en/ai-alliance-consultation [cited 2020 Jan 17].
  • 12.Beauchamp TL, Childres JF. Principles of biomedical ethics. New York: Oxford University Press; 2001. [Google Scholar]
  • 13.Taddeo M, Floridi L. How artificial intelligence can be a force for good. Science. 2018. August 24;361(6404):751–2. 10.1126/science.aat5991 [DOI] [PubMed] [Google Scholar]
  • 14.Health in 2015: from MDGs, millennium development goals to SDGs, sustainable development goals. Geneva: World Health Organization; 2015. Available from: https://www.who.int/gho/publications/mdgs-sdgs/en/ [cited 2020 Jan 21].
  • 15.Kruk ME, Gage AD, Arsenault C, Jordan K, Leslie HH, Roder-DeWan S, et al. High-quality health systems in the Sustainable Development Goals era: time for a revolution. Lancet Glob Health. 2018. November;6(11):e1196–252. 10.1016/S2214-109X(18)30386-3 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16.Ienca M, Fabrice J, Elger B, Caon M, Scoccia Pappagallo A, Kressig RW, et al. Intelligent assistive technology for alzheimer’s disease and other dementias: a systematic review. J Alzheimers Dis. 2017;56(4):1301–40. 10.3233/JAD-161037 [DOI] [PubMed] [Google Scholar]
  • 17.Tanioka T, Osaka K, Locsin R, Yasuhara Y, Ito H. Recommended design and direction of development for humanoid nursing robots: perspective from nursing researchers. Intell Control Autom. 2017;8(02):96–110. 10.4236/ica.2017.82008 [DOI] [Google Scholar]
  • 18.GBD 2013 Mortality and Causes of Death Collaborators. Global, regional, and national age-sex specific all-cause and cause-specific mortality for 240 causes of death, 1990–2013: a systematic analysis for the Global Burden of Disease Study 2013. Lancet. 2015. January 10;385(9963):117–71. 10.1016/S0140-6736(14)61682-2 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19.Jansen MPM, Baltussen R, Bærøe K. Stakeholder participation for legitimate priority setting: a checklist. Int J Health Policy Manag. 2018. November 1;7(11):973–6. 10.15171/ijhpm.2018.57 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 20.McCoy MS, Jongsma KR, Friesen P, Dunn M, Neuhaus CP, Rand L, et al. National standards for public involvement in research: missing the forest for the trees. J Med Ethics. 2018. December;44(12):801–4. 10.1136/medethics-2018-105088 [DOI] [PubMed] [Google Scholar]

Articles from Bulletin of the World Health Organization are provided here courtesy of World Health Organization

RESOURCES