Skip to main content
Journal of Medical Internet Research logoLink to Journal of Medical Internet Research
. 2020 Jul 7;22(7):e17707. doi: 10.2196/17707

Artificial Intelligence and Health Technology Assessment: Anticipating a New Level of Complexity

Hassane Alami 1,2,3,, Pascale Lehoux 1,2, Yannick Auclair 3, Michèle de Guise 3, Marie-Pierre Gagnon 4,5, James Shaw 6,7, Denis Roy 3, Richard Fleet 4,8,9, Mohamed Ali Ag Ahmed 10, Jean-Paul Fortin 4,11
Editor: Gunther Eysenbach
Reviewed by: Hong-Seng Gan, Thomas Poder, Malaika Reynolds, Beenish Chaudhry, Asiya Zaidi, Chariklia Tziraki, Syed Ghulam Sarwar Shah
PMCID: PMC7380986  PMID: 32406850

Abstract

Artificial intelligence (AI) is seen as a strategic lever to improve access, quality, and efficiency of care and services and to build learning and value-based health systems. Many studies have examined the technical performance of AI within an experimental context. These studies provide limited insights into the issues that its use in a real-world context of care and services raises. To help decision makers address these issues in a systemic and holistic manner, this viewpoint paper relies on the health technology assessment core model to contrast the expectations of the health sector toward the use of AI with the risks that should be mitigated for its responsible deployment. The analysis adopts the perspective of payers (ie, health system organizations and agencies) because of their central role in regulating, financing, and reimbursing novel technologies. This paper suggests that AI-based systems should be seen as a health system transformation lever, rather than a discrete set of technological devices. Their use could bring significant changes and impacts at several levels: technological, clinical, human and cognitive (patient and clinician), professional and organizational, economic, legal, and ethical. The assessment of AI’s value proposition should thus go beyond technical performance and cost logic by performing a holistic analysis of its value in a real-world context of care and services. To guide AI development, generate knowledge, and draw lessons that can be translated into action, the right political, regulatory, organizational, clinical, and technological conditions for innovation should be created as a first step.

Keywords: artificial intelligence, health technology assessment, eHealth, health care, medical device, patient, health services

Introduction

Artificial intelligence (AI) raises many expectations in all sectors of society. There is no universally agreed upon definition of what AI encompasses. Generically, it refers to a branch of informatics that develops systems that—through their ability to learn—imitate the characteristics associated with human intelligence: reasoning, learning, adaptation, self-correction, sensory comprehension, and interaction [1,2].

AI is seen as a strategic lever to improve access, quality, and efficiency of health care and services [3]. For example, by exploiting exhaustive data sets from complex systems, it could contribute to improving clinical decision making (eg, diagnosis, screening, and treatment), service organization (eg, flow optimization, triage, and resource allocation), and patient management and follow-up (eg, drug administration and compliance) [4].

However, research on the application of AI in health focuses primarily on technological performance in experimental contexts or on ethical issues. Although relevant, these studies do not fully address the broader systemic policy questions surrounding their use in a real-world context of care and services. In a recent meta-analysis, Lieu et al [5] concluded that despite a diagnostic performance equivalent to that of health care professionals, the diagnostic applications of AI have not been externally validated in a real-world context of care and services. Poor reporting is also prevalent in studies on AI, which limits the reliable interpretation of results. Thus, before being integrated into clinical routine, AI applications should overcome what is called the AI chasm, that is, the gap between reported performance in laboratory conditions and its performance and impacts in a real-world context of care and services [6]. AI raises issues of different types, but they are, in practice, closely interconnected: economic, professional, organizational, clinical, human, cognitive, legal, ethical, and technological. To date, few scholars have examined these issues in a systemic and holistic manner [7].

In this viewpoint paper, relying on the health technology assessment (HTA) core model [8], which is a methodological framework used to facilitate production and sharing of HTA information [9], we examine, based on our own experience as HTA academics and practitioners and in light of the emerging literature on the subject, issues raised by the use of AI. More specifically, we contrast the expectations specific to the health sector and the risks that should be mitigated for AI to be deployed responsibly. We limit our analysis to AI-based applications for clinical use (eg, diagnostic), some of which would be classified by the US Food and Drug Administration (FDA) as software as a medical device: “software intended to be used for one or more medical purposes that perform these purposes without being part of a hardware medical device” [10]. They are subject to formal regulatory approval [6,11].

In this paper, we offer critical observations and reflections that are informed by our various roles in HTA as health technology governance experts, researchers-evaluators, and/or decision makers. The analysis primarily adopts the perspective of payers (ie, health system organizations and agencies) because of their central role in regulating, funding, and reimbursing technologies [12].

On the basis of the HTA core model, we summarize key challenges posed by AI in a real-world context of care and services, which include (1) technological, (2) clinical, (3) human and cognitive (patient and clinician), (4) professional and organizational, (5) economic, and (6) legal and ethical dimensions (Textbox 1). We provide examples for each of these dimensions and underline how decision makers could approach them in a more systemic and holistic manner.

Synthesis of some key challenges posed by artificial intelligence.

Technological

  • Laboratory performance versus a real-world context of care and services

  • Data quality and representativeness of the general population or other contexts

  • Black box: how and why the decision is made?

  • Is artificial intelligence (AI) reliable and free of biases or technical failures?

  • How AI would react in situations where input data deviate from initial data?

  • Cybersecurity: data and model (algorithm)

  • Interoperability: fragmented systems and unstructured data

Clinical

  • Reproduction of tropism of practice models

  • Actual clinical added value in a real-world context of care and services: difficult to distinguish the effect of the AI's decision from the rest of the preventive and/or therapeutic strategy

  • The level of accuracy of AI in diagnosis and recommendations (reference standard) in a real-world context of care and services

Human and cognitive (patients)

  • Evolution of the nature and quality of the clinician-patient relationship

  • Loss of human contact: isolation of some people

  • Unrealistic expectations in some patients regarding clinical outcomes

  • Black box: could be perceived as a restriction on the patient’s right to make a free and informed decision

  • AI could be beneficial for one part of the population and not be for others: what is the good target population?

Human and cognitive (clinicians)

  • How to integrate AI into the electronic health record (EHR) and clinical routine with minimal effort and disruption for clinicians?

  • Nonintuitive technologies: weigh-down workflows and burden for clinicians and cognitive overload

  • Disruption of interpersonal communication styles (eg, clinician-clinician and clinician-patient)

  • AI as clinical mind: challenge of clinician’s decision-making autonomy

  • Absolute confidence in AI: technical dependence

Professional and organizational

  • How will it fit into the patient care and services trajectory?

  • How will it be integrated into the clinical-administrative processes and workflows of organizations and health system?

  • What changes will result in terms of service organization (eg, waiting time, primary care and specialized services relationships)?

  • How will it impact on professional jurisdictions (eg, reserved activities, responsibility, training, new skills, and expertise)?

Economic

  • Investments required: continuous performance tests, software and data quality tests, infrastructure and equipment upgrades, human expertise, and training

  • Clinical tropism and reimbursement/billing biases: costs for patients, clinicians, organizations, and health system

  • Need of new financing mechanisms, appropriate remuneration and/or reimbursement models, and insurance models

Legal and ethical

  • When is AI considered as a decision-making support tool? When is it considered as a decision-making tool?

  • What are the limits of technology and their potential legal implications?

  • If the AI makes a mistake (eg, black box), who will be held responsible? If the patient is harmed, who will pay for the repairs?

  • What would be the consequence if the clinician does not comply with the recommendations of an AI and this leads to an error?

  • AI needs access to data from different sources: consent is becoming more complex, as patients will be asked to authorize the use of diversified amounts of data

  • Protection and confidentiality: origin of the data, how consent was obtained, and authorization to use and/or reuse the data

  • Who owns the data? Who is responsible for it? Who can use (or reuse) it and under what conditions?

Technological Dimension

Generalizability and Reproducibility

Studies that focus on technological issues indicate that AI should provide the same level of performance in a real-world context of care and services as that obtained in laboratory conditions. However, this requirement is difficult to achieve [13-16]. The majority of AI applications reported in the literature are not exploitable in clinical practice [17]. AI is often trained with so-called clean (exclusion of poor-quality images) and complete data sets (elimination of imperfect data) [18]. It may not be operational in other contexts where data are incomplete or of poor quality (electronic health record [EHR] with missing data and/or erroneously entered data) [19-21]. This applies to some categories of the patient population (eg, low economic status and psychosocial problems) who receive care and services in a fragmented way in several organizations (institutional wandering) [21-24]. In addition, AI is usually trained on data specific to certain sites (hospital) and patients who are not necessarily representative of the general population. This includes decontextualized data (lack of psychosocial and organizational indicators) and data about disproportionately sick individuals (data enriched by metastases cases), men, and those from a European origin (ethnodiversity) [23,25-28].

Health organizations and systems produce and manage data in different ways. Variations may exist in clinical protocols (eg, diagnosis, procedures, and vital parameters) and devices (eg, different types of scanners, EHRs, and laboratory devices) on which AI applications are trained and those on which they are expected to operate [29,30]. These variations could affect the AI performance in a real-world context of care and services [31]. For example, an AI application trained on data from 2 hospitals in the United States performed poorly in a third hospital [13,32]. In its decision, the AI application had as predictors the image characteristics (magnetic resonance imaging machines specifications), imperceptible to humans, specific to the technological systems of the hospitals where it was trained. The AI solution had adapted to noise rather than to the signal of clinical interest [33]. When used in the third hospital, it was deprived of these expected predictors (noise), which affected its anticipated performance [32]. In the same vein, the use of data from the Framingham Heart Study to predict the risk of cardiovascular events produced biased results, which both overestimated and underestimated risk when AI was used in non-white populations [34,35]. The ability of AI to operate without bias or confounding factors on different devices and protocols remains a major challenge [36,37]. Thus, the fact that an algorithm was trained on large data sets does not mean that its results are generalizable.

Interpretability and Transparency

The interpretability and transparency of AI are important issues. The black box logic makes some AI applications vulnerable and at risk to false discoveries via spurious associations: how is the decision made and on what basis (justification and process description) [24,38,39]. This issue is central because these technologies will be diffused on a large scale. The error of a defective AI could have a greater impact (several patients) than a clinician's error on a single patient [20,31,40].

Interpretability and transparency are also necessary to identify the origin of errors, biases, or failures that should be prevented and/or avoided in the future [3,21,41]. For example, an AI application could lead to many undesirable impacts related to: (1) poor-quality training data, which could lead to erroneous or biased knowledge (garbage in, garbage out), whereas technology may further amplify how poor data produce poor results (noisy data and missing values); (2) the presence of a technical flaw in the algorithm (code), which could lead to erroneous inferences, even if good-quality data are used; (3) decision-making criteria that may not be universally acceptable; and (4) the emergence of new situations for which AI could not adapt, even with good-quality data and code [21,30,42-45]. For example, the emergence of new treatments or practices may require changes in clinical protocols; however, at present, AI applications are not developed to manage temporal data naturally in a real-world context of care and services. However, diseases and treatments evolve in a nonlinear manner [18,45]. The question thus remains regarding how AI would react, with observable indicators, in situations where input data deviate from initial data (EHRs and real-time monitoring devices), in the medium and long term [45,46].

The risk of cyberattacks is also a major concern. The data could be modified and/or fed by other false or biased data in a way that is difficult to detect [1]. For example, a slight intentional modification of laboratory results in a patient's EHR resulted in significant changes in the estimates of a well-trained AI of the same patient's risk of mortality [24]. For AI, the issue is two-fold because it is necessary to ensure the security of the data and that of the model (the algorithm). Interoperability is also a significant issue. The integration of AI in fragmented and noninteroperable information technology systems and organizations could create more problems than it will solve; to deliver its full potential, AI needs integrated and interoperable systems with fluent and optimal data circulation and exchange [17].

Finally, addressing interpretably and transparency in AI could be compromised by intellectual property issues, competitive strategy, and financial advantages that make companies reluctant to disclose their source codes [3].

Clinical Dimension

AI can entrench and disseminate practice models specific to particular contexts (organizations or health systems) and not necessarily accepted or used in others (tropism) [38]. For example, clinicians in some countries stopped using IBM Watson for Oncology because it reflected US specificity in cancer treatment [1,47].

To use AI in their decision making, clinicians should understand how it makes decisions in the first place [38,45,48]. They need the evidence to support a given conclusion to be able to carry out the necessary verifications or even corrections [14]: Why this decision (what information or image—or part of the image—tipped the final decision of the AI)? Why not another option (or choice)? When may I consider that the decision is correct? When should I accept this decision? How can I correct the error when it occurs?

AI should provide clinically added value for the patient. In a real-world context of care and services, much information, decisions, and diagnoses could intersect (eg, symptom assessment, laboratory tests, and radiology). At present, it is difficult to distinguish the effect of an AI-based decision from the overall preventive and/or therapeutic strategy of patient care [49,50].

Another clinical issue is determining the level of accuracy of AI for diagnosis and recommendations. In practice, decisions physicians make could diverge or even contradict each other in many situations. The gold standard is not always easy to define in a process that involves complex judgments [38,51,52]. In this case, should the standard reflect that of the lead clinician (or clinicians) in the organization? Or the one accepted by the majority of clinicians? Or the one reported in similar contexts? Some authors believe that for technologies that aim to provide pragmatic solutions under suboptimal conditions, AI performance should correspond to clinically acceptable practice in a given context and not necessarily to recommended practices [32]. This last point is likely to be problematic, particularly in a context where health systems are trying to overcome the challenge of practice variations to be able to provide equitable and quality services for all citizens.

Human and Cognitive Dimensions

AI could affect the nature and quality of the clinician-patient relationship and their expectations for care and follow-up [53,54]. The loss of human contact could lead to increased isolation of some people (replacement of health care providers) [1]. Some patients may feel able to control and manage their disease, with passive surveillance and/or less contact with the clinician, whereas others may feel overwhelmed by additional responsibilities [55]. AI may also create unrealistic expectations in some patients regarding clinical outcomes, which could have a negative impact on their care and service experience [56]. In addition, some AI-based decisions could be perceived as a restriction on the patient's right to make a free and informed decision [1,53]. Cultural and social aspects could play an important role in how patients will respond to AI and therefore how effective it can prove in practice [57]. Hence, it is important to know on which basis one may define the target population that can benefit from it [58]. In this regard, the question of social acceptability (acceptable risk and public confidence) also needs to be considered, which goes beyond the simple question of the effectiveness and usability of AI [59].

For clinicians, the challenge is to integrate AI into the EHR and clinical routine with minimal effort while respecting their decision-making autonomy [24]. Nonintuitive technologies could encumber workflows and become a burden for clinicians without improving service delivery [30,60]. Otherwise, the ability of AI to combine data from the scientific literature with learning from practice data could generate a repository of clinical practices (clinical mind), which could give AI an unwanted power or authority [35]. In some situations, AI may reduce the clinician's ability to take into account patient values and preferences. In contrast, some clinicians may develop absolute confidence and become dependent on AI, thus relinquishing their responsibility to verify or double-check its decisions [1].

In short, if clinicians feel overloaded and workflows become more complex, AI may be rejected because of self-perceived inefficacy and performance, alert fatigue, cognitive overload, and disruption of interpersonal communication routines [54,61-63].

Professional and Organizational Dimensions

Global appreciation of the added value of AI should take into account the nature and magnitude of the professional and organizational changes required for its use [6]. For example, the FDA has approved an AI application used for diabetic retinopathy screening, which may be used in primary care clinics [11]. As in some countries, the screening procedure is performed by an ophthalmologist (specialist), some questions arise: How will this technology fit into patient care and services trajectory? How will it be integrated into the clinical-administrative processes of organizations and the health system? If used at the primary care level, will general practitioners, nurses, or optometrists be allowed to supervise the AI? If so, under what conditions? What will be the impact on professional jurisdictions (regulated activities, remuneration, and training)? What changes will result in terms of service organization and clinical-administrative workflows (waiting time at primary care level, primary care, and specialized services relationships)?

Thus, AI could lead to a redistribution of work between different professional scopes of practice and highlight the need for other clinical, administrative, and technical skills and expertise. This will require clarifying new rules and processes (clinical and administrative), negotiating and reframing professional jurisdictions, responsibilities, and privileges associated with them and reassessing the number of positions needed and the new skills required to work (with) and/or perform other tasks that accompany its use. This will have to take into consideration how new roles in terms of skills in informatics and data science and the ability to liaise may be introduced within clinical teams [64].

Finally, today, most AI applications are developed to perform a single task or a set of very specific tasks (eg, diagnosing only diabetic retinopathy and macular edema) [65]. They are unusable for other diagnoses for which they are not trained (eg, nondiabetic retinopathy lesions and eye melanoma) and are unable, at least for the moment, to replace a complete clinical examination [66]. Payers will thus be tasked to determine whether AI provides sufficient added value in relation to the nature and magnitude of the clinical, cognitive, professional, and organizational changes it could generate.

Economic Dimension

To adapt an AI to a local environment, considerable investments and expenditures may be necessary. The evolution of AI in a real-world context of care and services, by integrating large amounts of data of various types and sources, requires additional resources to ensure its proper functioning and stability: continuous performance tests, software and data quality tests, infrastructure and equipment upgrades, human expertise, and training [3,67]. However, many health organizations do not have a secure and scalable technological and data infrastructure as well as adequate human resources to ensure proper collection of the data necessary for the training and adaptation of AI to their local population and clinical environment [17]. The literature on AI’s promises as well as innovation policies that support its development downplays the capital-intensive requirements that are required to properly deploy AI, compared with the day-to-day work of managers in organizations.

In health systems where activity-based financing is the basis for funding health organizations, some clinicians tend to enter the highest paying codes for each clinical activity (ie, the most complex case of an intervention) to increase performance and maximize revenue. An AI application trained on data from these organizations (EHR with invoicing or reimbursement data) could amplify biases inherent in such practices that do not necessarily reflect the actual clinical condition [23,44,68]. The replication and entrenchment at a large scale of these biases could result in significant costs for patients, clinicians, organizations, and the health systems [35].

Similarly, some AI applications may be too cautious, resulting in an increase in requests for unnecessary testing and treatment, leading to overdiagnosis or overprescription [69]. Their recommendations, which are not necessarily associated with improved patient outcomes, could lead to increased costs and expenses for patients and the health system.

Legal and Ethical Dimensions

Many AI technologies are still considered today as decision-making support tools for clinicians. It could then be argued that the legal responsibility for the decision still rests with the clinician. However, with the growing performance of AI, clinicians may be increasingly influenced and may more easily accept AI decisions, even when there is clinical ambiguity. Determining the clinician's degree of responsibility becomes more complex [30]. The challenge here is to distinguish between several situations: When is it considered a decision-making support tool? When is it considered a decision-making tool? This distinction is key in defining who is legally responsible in the event of an error or a malfunction (professional misconduct) [30,51,70].

For example, if the clinical decision is based on an erroneous clinical recommendation from the AI (delayed or erroneous treatment), who will be held responsible? Is it the technology developer, technology provider, clinician, organization, or do they all share responsibility (and how)? In some jurisdictions, to confirm professional misconduct, it is necessary to prove that the standard of care was not followed. This standard is blurred when AI comes into play [2]. In addition, the likely consequence if the clinician does not comply with the recommendations of an AI and if this leads to an error must be anticipated [2]. It could be argued that the responsibility should rest with the human controller of AI, but such a responsibility becomes difficult to clarify when autonomous technologies are used [57]. In this regard, standards may shift over time: “What happens if medical practice reaches a point where AI becomes part of the standard of care?” Medical insurers and regulators will have to be able to distinguish errors inherent in the tool from those resulting from misuse by the clinician, the organization, or even the patient, an issue exacerbated by the black box of AI [51,71].

To generate a complete picture of the patient, AI will need access to data from different organizations (hospitals and insurers) [45]. The risk of disclosing sensitive information about patients or certain populations is real [45]. For example, some AI applications can reidentify an individual from only three different data sources [25,38,72]. In the same vein, the issue of consent is becoming more complex, as patients will be asked to authorize the use of increasingly large and diversified amounts of data about them: medical records, audio, videos, and socioeconomic data [58]. Problems could arise if the patient only consents to sharing parts of his or her data. Usually, confidentiality means that the clinician can withhold certain information—at the patient's request (or not)—and avoid entering it into the EHR. Incomplete data make AI less efficient and does not allow patients to benefit from the best possible services. AI may not be fully operational in a real-world context of care and services if specific restrictions on data access and use are applied [38].

Protection and confidentiality requirements imply the obligation to know several things: the origin of the data, how consent was obtained, and authorization to use and/or reuse the data for training and in a real-world context of care and services. As the data may come from different sources and contexts, different conditions and precautions will need to be considered [73]. Regulators will need to determine who owns the data and, in the context of public-private partnerships, who is responsible for its collection, use, transmission to third parties, and under what conditions [17]. As the answers will vary according to the nature of the data, the jurisdictions, and the purpose of use, the task at hand is sizable [73]. Finally, payers will have to recognize that the ethical implications of AI affect, directly or indirectly, all the other dimensions discussed earlier.

Conclusions

The purpose of this viewpoint paper is to provide a structured roadmap of the issues surrounding the integration of AI into health care organizations and systems. To the best of our knowledge, this is one of the few papers that offers a multidimensional and holistic analysis on the subject [7]. It contributes to current knowledge by providing a necessary basis for reflections, exchanges, and knowledge sharing among the various stakeholders concerned with AI in health care.

In light of the issues we identified, it becomes clear that regulatory and decision-making organizations as well as HTA agencies are facing unprecedented complexity: evaluating and approving so-called disruptive technologies, especially AI, requires taking several issues into consideration altogether. Many studies have reported significant technical performance of AI technologies, but very few have adopted a holistic standpoint that can situate their impacts and associated changes and transformations in health systems. Technical studies are rarely adapted to the complexity surrounding AI applications, as they overlook the context-dependent changes or adjustments the implementation and use of technology requires (variations, clinical and organizational interactions, and interdependencies) [74]. According to the frame problem [62,75], which highlights the difficulty for AI, beyond the specific tasks it masters, to update its set of axioms to capture the context in which it is implemented and used (eg, patient preferences, environment and social support, clinical history, personality/cultural characteristics and values that influence clinical outcomes, and empathy in medicine), the complexity inherent in the use of AI applications in the real-world context of care and services may seem difficult to overcome [62].

For informed decision making, there is a real need for evaluations that address AI as a lever of health system transformation. Given the magnitude of the implications it could have at all levels, the evaluation of AI’s value proposition should go beyond its technical performance and cost logic to incorporate its global value based on a holistic analysis in a real-world context of care and services. In this vein, technology brings value when its use in a real-world context of care and services contributes to the aims of the health system and aligns with the values of society. Global value appreciation could be based on the quintuple aim: (1) better quality and experience of care and services for patients; (2) a better state of health and well-being for the entire population; (3) reducing costs for responsible and sustainable resource management; (4) a better quality of work and satisfaction of health care providers; and (5) equity and inclusion to avoid exacerbating health disparities in the population [76]. From this perspective, further research on the evaluation of AI should no longer be limited to a technological approach, which only demonstrates quality from an engineering point of view and costs—motivated mainly by a logic of short-term savings—but should broaden its horizons to include the dimensions this paper underscored [77,78].

Real-world evaluations could be a major asset in informing AI decision making. In the context of uncertainty, iterative and reflective evaluation approaches should be developed to encourage dialog and collaboration among all relevant stakeholders (eg, payers, health care providers, technology providers, regulators, citizens/patients, academic researchers, and evaluation agencies) [63,78,79]. In addition, an early dialog between these stakeholders is needed to identify the evidence required to inform decision making [63,78]. This approach would also help AI providers to better understand the expectations of the health system [78]. This change implies that HTA should play an active role as a mediator and facilitator of transparent dialog between different stakeholders who are implicated throughout the technology’s life cycle [78,80].

Decision making for innovative technologies is inherently complex, in particular because of visions, perceptions, and objectives that may differ between the stakeholders involved: risk sharing is essential to strive to find a balance between uncertainty and added value [81]. In this regard, “major radical innovations never bring new technologies into the world in a fully developed form” but “appear in a crude and embryonic state with only a few specific uses” [81]. It is their use in a real-world context of care and services, through a process of learning by doing (improving users’ skills) and learning by using (improving users’ knowledge), which makes it possible to appreciate their global value [81]. With the complexity associated with AI, value appreciation becomes even more complex, challenging the traditional methodological foundations that are the basis for decision making about innovative technologies [82]. This also presents a unique opportunity for HTA to evolve and adapt (evaluative framework and contextualized data), particularly in view of the importance of contexts in the appreciation of the value of innovative technologies [83,84]. It is necessary for HTA scholars and practitioners to explore and exploit other avenues, complementary to traditional methods, to collect data and information that can better inform AI-related decisions [85].

Finally, this new context implies mechanisms for continuous collective learning and sharing of lessons. To do so, there is a need for learning and flexible health organizations and systems that are able to adjust and operate under uncertainty. In this regard, creating the political, regulatory, organizational, clinical, and technological conditions necessary for proper innovation is the first step. This requires building trust to ensure stakeholder engagement to guide AI developments, rapidly generate knowledge in a real-world context of care and services, and draw lessons to translate them into action.

Acknowledgments

HA was supported by the Canadian Institutes of Health Research (CIHR)’s Health System Impact Fellowship. This program is led by CIHR’s Institute of Health Services and Policy Research, in partnership with the Fonds de recherche du Québec–Santé and the Institut national d’excellence en santé et services sociaux. The authors would like to thank the reviewers and the editorial team for their insightful comments and suggestions, as these comments and suggestions led to an improvement of the manuscript.

Abbreviations

AI

artificial intelligence

CIHR

Canadian Institutes of Health Research

EHR

electronic health record

FDA

Food and Drug Administration

HTA

health technology assessment

Footnotes

Authors' Contributions: HA and PL produced the first draft of this manuscript and received input from YA, MG, MG, JS, DR, RF, MA, and JF. All authors have read and approved the final manuscript.

Conflicts of Interest: None declared.

References

  • 1.Scott RE, Mars M. Principles and framework for ehealth strategy development. J Med Internet Res. 2013 Jul 30;15(7):e155. doi: 10.2196/jmir.2250. https://www.jmir.org/2013/7/e155/ [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 2.Massalha S, Clarkin O, Thornhill R, Wells G, Chow BJ. Decision support tools, systems, and artificial intelligence in cardiac imaging. Can J Cardiol. 2018 Jul;34(7):827–38. doi: 10.1016/j.cjca.2018.04.032. [DOI] [PubMed] [Google Scholar]
  • 3.He J, Baxter SL, Xu J, Xu J, Zhou X, Zhang K. The practical implementation of artificial intelligence technologies in medicine. Nat Med. 2019 Jan;25(1):30–6. doi: 10.1038/s41591-018-0307-0. http://europepmc.org/abstract/MED/30617336. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4.Bini SA. Artificial intelligence, machine learning, deep learning, and cognitive computing: what do these terms mean and how will they impact health care? J Arthroplasty. 2018 Aug;33(8):2358–61. doi: 10.1016/j.arth.2018.02.067. [DOI] [PubMed] [Google Scholar]
  • 5.Liu X, Faes L, Kale AU, Wagner SK, Fu DJ, Bruynseels A, Mahendiran T, Moraes G, Shamdas M, Kern C, Ledsam Jr, Schmid MK, Balaskas K, Topol EJ, Bachmann LM, Keane PA, Denniston AK. A comparison of deep learning performance against health-care professionals in detecting diseases from medical imaging: a systematic review and meta-analysis. Lancet Digit Health. 2019 Oct;1(6):e271–97. doi: 10.1016/S2589-7500(19)30123-2. doi: 10.1016/S2589-7500(19)30123-2. [DOI] [PubMed] [Google Scholar]
  • 6.Keane PA, Topol EJ. With an eye to AI and autonomous diagnosis. Digit Med. 2018;1:40. doi: 10.1038/s41746-018-0048-y. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7.Shaw J, Rudzicz F, Jamieson T, Goldfarb A. Artificial intelligence and the implementation challenge. J Med Internet Res. 2019 Jul 10;21(7):e13659. doi: 10.2196/13659. https://www.jmir.org/2019/7/e13659/ [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.Lampe K, Mäkelä M, Garrido MV, Anttila H, Autti-Rämö I, Hicks NJ, Hofmann B, Koivisto J, Kunz R, Kärki P, Malmivaara A, Meiesaar K, Reiman-Möttönen P, Norderhaug I, Pasternack I, Ruano-Ravina A, Räsänen P, Saalasti-Koskinen U, Saarni SI, Walin L, Kristensen FB, European Network for Health Technology Assessment (EUnetHTA) The HTA core model: a novel method for producing and reporting health technology assessments. Int J Technol Assess Health Care. 2009 Dec;25(Suppl 2):9–20. doi: 10.1017/S0266462309990638. [DOI] [PubMed] [Google Scholar]
  • 9.Kristensen FB, Lampe K, Wild C, Cerbo M, Goettsch W, Becla L. The HTA core model-10 years of developing an international framework to share multidimensional value assessment. Value Health. 2017 Feb;20(2):244–50. doi: 10.1016/j.jval.2016.12.010. https://linkinghub.elsevier.com/retrieve/pii/S1098-3015(16)34220-6. [DOI] [PubMed] [Google Scholar]
  • 10.Software as a Medical Device (SaMD): Key Definitions. International Medical Device Regulators Forum (IMDRF) 2013. [2020-04-23]. http://www.imdrf.org/docs/imdrf/final/technical/imdrf-tech-131209-samd-key-definitions-140901.pdf.
  • 11.FDA Permits Marketing of Artificial Intelligence-Based Device to Detect Certain Diabetes-Related Eye Problems. US Food and Drug Administration. 2018. [2020-04-23]. https://www.fda.gov/news-events/press-announcements/fda-permits-marketing-artificial-intelligence-based-device-detect-certain-diabetes-related-eye.
  • 12.Godman B, Bucsics A, Bonanno PV, Oortwijn W, Rothe CC, Ferrario A, Bosselli S, Hill A, Martin AP, Simoens S, Kurdi A, Gad M, Gulbinovič J, Timoney A, Bochenek T, Salem A, Hoxha I, Sauermann R, Massele A, Guerra AA, Petrova G, Mitkova Z, Achniotou G, Laius O, Sermet C, Selke G, Kourafalos V, Yfantopoulos J, Magnusson E, Joppi R, Oluka M, Kwon H, Jakupi A, Kalemeera F, Fadare JO, Melien O, Pomorski M, Wladysiuk M, Marković-Peković V, Mardare I, Meshkov D, Novakovic T, Fürst J, Tomek D, Zara C, Diogene E, Meyer JC, Malmström R, Wettermark B, Matsebula Z, Campbell S, Haycox A. Barriers for access to new medicines: searching for the balance between rising costs and limited budgets. Front Public Health. 2018;6:328. doi: 10.3389/fpubh.2018.00328. doi: 10.3389/fpubh.2018.00328. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.Zech JR, Badgeley MA, Liu M, Costa AB, Titano JJ, Oermann EK. Variable generalization performance of a deep learning model to detect pneumonia in chest radiographs: a cross-sectional study. PLoS Med. 2018 Nov;15(11):e1002683. doi: 10.1371/journal.pmed.1002683. http://dx.plos.org/10.1371/journal.pmed.1002683. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14.Yasaka K, Abe O. Deep learning and artificial intelligence in radiology: current applications and future directions. PLoS Med. 2018 Nov;15(11):e1002707. doi: 10.1371/journal.pmed.1002707. http://dx.plos.org/10.1371/journal.pmed.1002707. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15.Wong J, Horwitz MM, Zhou L, Toh S. Using machine learning to identify health outcomes from electronic health record data. Curr Epidemiol Rep. 2018 Dec;5(4):331–42. doi: 10.1007/s40471-018-0165-9. http://europepmc.org/abstract/MED/30555773. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16.Pirracchio R, Cohen MJ, Malenica I, Cohen J, Chambaz A, Cannesson M, Lee C, Resche-Rigon M, Hubbard A, ACTERREA Research Group Big data and targeted machine learning in action to assist medical decision in the ICU. Anaesth Crit Care Pain Med. 2019 Aug;38(4):377–84. doi: 10.1016/j.accpm.2018.09.008. [DOI] [PubMed] [Google Scholar]
  • 17.Panch T, Mattie H, Celi LA. The 'inconvenient truth' about AI in healthcare. Digit Med. 2019;2:77. doi: 10.1038/s41746-019-0155-4. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18.Sherbet GV, Woo WL, Dlay S. Application of artificial intelligence-based technology in cancer management: a commentary on the deployment of artificial neural networks. Anticancer Res. 2018 Dec;38(12):6607–13. doi: 10.21873/anticanres.13027. [DOI] [PubMed] [Google Scholar]
  • 19.Redd TK, Campbell JP, Brown JM, Kim SJ, Ostmo S, Chan RV, Dy J, Erdogmus D, Ioannidis S, Kalpathy-Cramer J, Chiang MF, Imaging and Informatics in Retinopathy of Prematurity (i-ROP) Research Consortium Evaluation of a deep learning image assessment system for detecting severe retinopathy of prematurity. Br J Ophthalmol. 2018 Nov 23;:-. doi: 10.1136/bjophthalmol-2018-313156. epub ahead of print. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 20.Kanagasingam Y, Xiao D, Vignarajan J, Preetham A, Tay-Kearney M, Mehrotra A. Evaluation of artificial intelligence-based grading of diabetic retinopathy in primary care. JAMA Netw Open. 2018 Sep 7;1(5):e182665. doi: 10.1001/jamanetworkopen.2018.2665. https://jamanetwork.com/journals/jamanetworkopen/fullarticle/10.1001/jamanetworkopen.2018.2665. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21.Walsh CG, Chaudhry B, Dua P, Goodman KW, Kaplan B, Kavuluru R, Solomonides A, Subbian V. Stigma, biomarkers, and algorithmic bias: recommendations for precision behavioral health with artificial intelligence. JAMIA Open. 2020;3(1):9–15. doi: 10.1093/jamiaopen/ooz054. https://academic.oup.com/jamiaopen/article/3/1/9/5714181. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22.Gianfrancesco MA, Tamang S, Yazdany J, Schmajuk G. Potential biases in machine learning algorithms using electronic health record data. JAMA Intern Med. 2018 Nov 1;178(11):1544–7. doi: 10.1001/jamainternmed.2018.3763. http://europepmc.org/abstract/MED/30128552. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 23.Obermeyer Z, Powers B, Vogeli C, Mullainathan S. Dissecting racial bias in an algorithm used to manage the health of populations. Science. 2019 Oct 25;366(6464):447–53. doi: 10.1126/science.aax2342. [DOI] [PubMed] [Google Scholar]
  • 24.Wang F, Casalino LP, Khullar D. Deep learning in medicine-promise, progress, and challenges. JAMA Intern Med. 2019 Mar 1;179(3):293–4. doi: 10.1001/jamainternmed.2018.7117. [DOI] [PubMed] [Google Scholar]
  • 25.Crigger E, Khoury C. Making policy on augmented intelligence in health care. AMA J Ethics. 2019 Feb 1;21(2):E188–91. doi: 10.1001/amajethics.2019.188. https://journalofethics.ama-assn.org/article/making-policy-augmented-intelligence-health-care/2019-02. [DOI] [PubMed] [Google Scholar]
  • 26.Cabitza F, Rasoini R, Gensini GF. Unintended consequences of machine learning in medicine. J Am Med Assoc. 2017 Aug 8;318(6):517–8. doi: 10.1001/jama.2017.7797. [DOI] [PubMed] [Google Scholar]
  • 27.Zou J, Schiebinger L. AI can be sexist and racist - it's time to make it fair. Nature. 2018 Jul;559(7714):324–6. doi: 10.1038/d41586-018-05707-8. [DOI] [PubMed] [Google Scholar]
  • 28.Bejnordi BE, Veta M, van Diest PJ, van Ginneken B, Karssemeijer N, Litjens G, van der Laak JA, The CAMELYON16 Consortium. Hermsen M, Manson QF, Balkenhol M, Geessink O, Stathonikos N, van Dijk MC, Bult P, Beca F, Beck AH, Wang D, Khosla A, Gargeya R, Irshad H, Zhong A, Dou Q, Li Q, Chen H, Lin H, Heng P, Haß C, Bruni E, Wong Q, Halici U, Öner MU, Cetin-Atalay R, Berseth M, Khvatkov V, Vylegzhanin A, Kraus O, Shaban M, Rajpoot N, Awan R, Sirinukunwattana K, Qaiser T, Tsang Y, Tellez D, Annuscheit J, Hufnagl P, Valkonen M, Kartasalo K, Latonen L, Ruusuvuori P, Liimatainen K, Albarqouni S, Mungal B, George A, Demirci S, Navab N, Watanabe S, Seno S, Takenaka Y, Matsuda H, Phoulady HA, Kovalev V, Kalinovsky A, Liauchuk V, Bueno G, Fernandez-Carrobles MM, Serrano I, Deniz O, Racoceanu D, Venâncio R. Diagnostic assessment of deep learning algorithms for detection of lymph node metastases in women with breast cancer. J Am Med Assoc. 2017 Dec 12;318(22):2199–210. doi: 10.1001/jama.2017.14585. http://europepmc.org/abstract/MED/29234806. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 29.Keel S, van Wijngaarden P. The eye in AI: artificial intelligence in ophthalmology. Clin Exp Ophthalmol. 2019 Jan;47(1):5–6. doi: 10.1111/ceo.13435. [DOI] [PubMed] [Google Scholar]
  • 30.Harwich E, Laycock K. Thinking on its Own: AI in the NHS. London, UK: Reform Research Trust; 2018. [Google Scholar]
  • 31.Topol EJ. High-performance medicine: the convergence of human and artificial intelligence. Nat Med. 2019 Jan;25(1):44–56. doi: 10.1038/s41591-018-0300-7. [DOI] [PubMed] [Google Scholar]
  • 32.Nevin L, PLOS Medicine Editors Advancing the beneficial use of machine learning in health care and medicine: toward a community understanding. PLoS Med. 2018 Nov;15(11):e1002708. doi: 10.1371/journal.pmed.1002708. http://dx.plos.org/10.1371/journal.pmed.1002708. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 33.Tiffin PA, Paton LW. Rise of the machines? Machine learning approaches and mental health: opportunities and challenges. Br J Psychiatry. 2018 Sep;213(3):509–10. doi: 10.1192/bjp.2018.105. [DOI] [PubMed] [Google Scholar]
  • 34.Gijsberts CM, Groenewegen KA, Hoefer IE, Eijkemans MJ, Asselbergs FW, Anderson TJ, Britton AR, Dekker JM, Engström G, Evans GW, de Graaf J, Grobbee DE, Hedblad B, Holewijn S, Ikeda A, Kitagawa K, Kitamura A, de Kleijn D, Lonn EM, Lorenz MW, Mathiesen EB, Nijpels G, Okazaki S, O'Leary DH, Pasterkamp G, Peters SA, Polak JF, Price JF, Robertson C, Rembold CM, Rosvall M, Rundek T, Salonen JT, Sitzer M, Stehouwer CD, Bots ML, den Ruijter HM. Race/ethnic differences in the associations of the Framingham risk factors with carotid IMT and cardiovascular events. PLoS One. 2015;10(7):e0132321. doi: 10.1371/journal.pone.0132321. http://dx.plos.org/10.1371/journal.pone.0132321. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 35.Char DS, Shah NH, Magnus D. Implementing machine learning in health care - addressing ethical challenges. N Engl J Med. 2018 Mar 15;378(11):981–3. doi: 10.1056/NEJMp1714229. http://europepmc.org/abstract/MED/29539284. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 36.Madabhushi A, Lee G. Image analysis and machine learning in digital pathology: challenges and opportunities. Med Image Anal. 2016 Oct;33:170–5. doi: 10.1016/j.media.2016.06.037. http://europepmc.org/abstract/MED/27423409. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 37.Mateos-Pérez JM, Dadar M, Lacalle-Aurioles M, Iturria-Medina Y, Zeighami Y, Evans AC. Structural neuroimaging as clinical predictor: a review of machine learning applications. Neuroimage Clin. 2018;20:506–22. doi: 10.1016/j.nicl.2018.08.019. https://linkinghub.elsevier.com/retrieve/pii/S2213-1582(18)30260-2. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 38.Ching T, Himmelstein DS, Beaulieu-Jones BK, Kalinin AA, Do BT, Way GP, Ferrero E, Agapow P, Zietz M, Hoffman MM, Xie W, Rosen GL, Lengerich BJ, Israeli J, Lanchantin J, Woloszynek S, Carpenter AE, Shrikumar A, Xu J, Cofer EM, Lavender CA, Turaga SC, Alexandari AM, Lu Z, Harris DJ, DeCaprio D, Qi Y, Kundaje A, Peng Y, Wiley LK, Segler MH, Boca SM, Swamidass SJ, Huang A, Gitter A, Greene CS. Opportunities and obstacles for deep learning in biology and medicine. J R Soc Interface. 2018 Apr;15(141):-. doi: 10.1098/rsif.2017.0387. http://europepmc.org/abstract/MED/29618526. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 39.Foster KR, Koprowski R, Skufca JD. Machine learning, medical diagnosis, and biomedical engineering research - commentary. Biomed Eng Online. 2014 Jul 5;13:94. doi: 10.1186/1475-925X-13-94. https://biomedical-engineering-online.biomedcentral.com/articles/10.1186/1475-925X-13-94. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 40.Xiao C, Choi E, Sun J. Opportunities and challenges in developing deep learning models using electronic health records data: a systematic review. J Am Med Inform Assoc. 2018 Oct 1;25(10):1419–28. doi: 10.1093/jamia/ocy068. http://europepmc.org/abstract/MED/29893864. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 41.Kuan R. Adopting AI in Health Care Will Be Slow and Difficult. Harvard Business Review. 2019. [2020-04-23]. https://hbr.org/2019/10/adopting-ai-in-health-care-will-be-slow-and-difficult.
  • 42.Fox J, Das SK. Safe and Sound: Artificial Intelligence in Hazardous Applications. Menlo Park, CA: AAAI Press; 2000. [Google Scholar]
  • 43.Bibault J, Giraud P, Burgun A. Big data and machine learning in radiation oncology: state of the art and future prospects. Cancer Lett. 2016 Nov 1;382(1):110–7. doi: 10.1016/j.canlet.2016.05.033. [DOI] [PubMed] [Google Scholar]
  • 44.Yu K, Kohane IS. Framing the challenges of artificial intelligence in medicine. BMJ Qual Saf. 2019 Mar;28(3):238–41. doi: 10.1136/bmjqs-2018-008551. [DOI] [PubMed] [Google Scholar]
  • 45.Jordan MI, Mitchell TM. Machine learning: trends, perspectives, and prospects. Science. 2015 Jul 17;349(6245):255–60. doi: 10.1126/science.aaa8415. [DOI] [PubMed] [Google Scholar]
  • 46.Derrington D. Artificial Intelligence for Health and Health Care. Office of the National Coordinator for Health Information. 2017. [2019-04-23]. https://www.healthit.gov/sites/default/files/jsr-17-task-002_aiforhealthandhealthcare12122017.pdf.
  • 47.Ross C, Swetlitz I. IBM Pitched Its Watson Supercomputer as a Revolution in Cancer Care. It’s Nowhere Close. Stat. 2017. [2020-04-23]. https://www.statnews.com/2017/09/05/watson-ibm-cancer/
  • 48.Wong TY, Bressler NM. Artificial intelligence with deep learning technology looks into diabetic retinopathy screening. J Am Med Assoc. 2016 Dec 13;316(22):2366–7. doi: 10.1001/jama.2016.17563. [DOI] [PubMed] [Google Scholar]
  • 49.Belfort MA, Clark SL. Computerised cardiotocography-study design hampers findings. Lancet. 2017 Apr 29;389(10080):1674–6. doi: 10.1016/S0140-6736(17)30762-6. https://linkinghub.elsevier.com/retrieve/pii/S0140-6736(17)30762-6. [DOI] [PubMed] [Google Scholar]
  • 50.Park SH, Han K. Methodologic guide for evaluating clinical performance and effect of artificial intelligence technology for medical diagnosis and prediction. Radiology. 2018 Mar;286(3):800–9. doi: 10.1148/radiol.2017171920. [DOI] [PubMed] [Google Scholar]
  • 51.Price WN, Gerke S, Cohen IG. Potential liability for physicians using artificial intelligence. J Am Med Assoc. 2019 Oct 4;:-. doi: 10.1001/jama.2019.15064. epub ahead of print. [DOI] [PubMed] [Google Scholar]
  • 52.Kim J, Hong J, Park H. Prospects of deep learning for medical imaging. Precis Future Med. 2018 Jun;2(2):37–52. doi: 10.23838/pfm.2018.00030. doi: 10.23838/pfm.2018.00030. [DOI] [Google Scholar]
  • 53.Bi WL, Hosny A, Schabath MB, Giger ML, Birkbak NJ, Mehrtash A, Allison T, Arnaout O, Abbosh C, Dunn IF, Mak RH, Tamimi RM, Tempany CM, Swanton C, Hoffmann U, Schwartz LH, Gillies RJ, Huang RY, Aerts HJ. Artificial intelligence in cancer imaging: clinical challenges and applications. CA Cancer J Clin. 2019 Mar;69(2):127–57. doi: 10.3322/caac.21552. doi: 10.3322/caac.21552. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 54.Blease C, Kaptchuk TJ, Bernstein MH, Mandl KD, Halamka JD, DesRoches CM. Artificial intelligence and the future of primary care: exploratory qualitative study of UK general practitioners' views. J Med Internet Res. 2019 Mar 20;21(3):e12802. doi: 10.2196/12802. https://www.jmir.org/2019/3/e12802/ [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 55.Glenn T, Monteith S. New measures of mental state and behavior based on data collected from sensors, smartphones, and the internet. Curr Psychiatry Rep. 2014 Dec;16(12):523. doi: 10.1007/s11920-014-0523-3. [DOI] [PubMed] [Google Scholar]
  • 56.Luxton DD. Should Watson be consulted for a second opinion? AMA J Ethics. 2019 Feb 1;21(2):E131–7. doi: 10.1001/amajethics.2019.131. https://journalofethics.ama-assn.org/article/should-watson-be-consulted-second-opinion/2019-02. [DOI] [PubMed] [Google Scholar]
  • 57.Luxton DD. Artificial intelligence in psychological practice: current and future applications and implications. Prof Psychol Res Pr. 2014;45(5):332–9. doi: 10.1037/a0034559. [DOI] [Google Scholar]
  • 58.Lovejoy CA, Buch V, Maruthappu M. Technology and mental health: the role of artificial intelligence. Eur Psychiatry. 2019 Jan;55:1–3. doi: 10.1016/j.eurpsy.2018.08.004. [DOI] [PubMed] [Google Scholar]
  • 59.Poder TG, Bellemare CA, Bédard SK, Lemieux R. Social acceptance and population confidence in telehealth in Quebec. BMC Health Serv Res. 2015 Feb 21;15:72. doi: 10.1186/s12913-015-0727-1. https://bmchealthservres.biomedcentral.com/articles/10.1186/s12913-015-0727-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 60.Shortliffe EH, Sepúlveda MJ. Clinical decision support in the era of artificial intelligence. J Am Med Assoc. 2018 Dec 4;320(21):2199–200. doi: 10.1001/jama.2018.17163. [DOI] [PubMed] [Google Scholar]
  • 61.McCullagh LJ, Sofianou A, Kannry J, Mann DM, McGinn TG. User centered clinical decision support tools: adoption across clinician training level. Appl Clin Inform. 2014;5(4):1015–25. doi: 10.4338/ACI-2014-05-RA-0048. http://europepmc.org/abstract/MED/25589914. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 62.Yu K, Beam AL, Kohane IS. Artificial intelligence in healthcare. Nat Biomed Eng. 2018 Oct;2(10):719–31. doi: 10.1038/s41551-018-0305-z. [DOI] [PubMed] [Google Scholar]
  • 63.Laï MC, Brian M, Mamzer M. Perceptions of artificial intelligence in healthcare: findings from a qualitative survey study among actors in France. J Transl Med. 2020 Jan 9;18(1):14. doi: 10.1186/s12967-019-02204-y. https://translational-medicine.biomedcentral.com/articles/10.1186/s12967-019-02204-y. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 64.Gillan C, Milne E, Harnett N, Purdie T, Jaffray D, Hodges B. Professional implications of introducing artificial intelligence in healthcare: an evaluation using radiation medicine as a testing ground. J Radiother Pract. 2018 Oct 3;18(1):5–9. doi: 10.1017/S1460396918000468. doi: 10.1017/S1460396918000468. [DOI] [Google Scholar]
  • 65.Syed AB, Zoga AC. Artificial intelligence in radiology: current technology and future directions. Semin Musculoskelet Radiol. 2018 Nov;22(5):540–5. doi: 10.1055/s-0038-1673383. [DOI] [PubMed] [Google Scholar]
  • 66.Gulshan V, Peng L, Coram M, Stumpe MC, Wu D, Narayanaswamy A, Venugopalan S, Widner K, Madams T, Cuadros J, Kim R, Raman R, Nelson PC, Mega JL, Webster DR. Development and validation of a deep learning algorithm for detection of diabetic retinopathy in retinal fundus photographs. J Am Med Assoc. 2016 Dec 13;316(22):2402–10. doi: 10.1001/jama.2016.17216. [DOI] [PubMed] [Google Scholar]
  • 67.Sendak M, Gao M, Nichols M, Lin A, Balu S. Machine learning in health care: a critical appraisal of challenges and opportunities. EGEMS (Wash DC) 2019 Jan 24;7(1):1. doi: 10.5334/egems.287. http://europepmc.org/abstract/MED/30705919. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 68.Miller DD, Brown EW. Artificial intelligence in medical practice: the question to the answer? Am J Med. 2018 Feb;131(2):129–33. doi: 10.1016/j.amjmed.2017.10.035. [DOI] [PubMed] [Google Scholar]
  • 69.Loder J, Nicholas L. Confronting Dr Robot: Creating a People-Powered Future For AI in Health. Nesta: The Innovation Foundation. 2018. [2020-04-23]. https://media.nesta.org.uk/documents/confronting_dr_robot.pdf.
  • 70.Toratani M, Konno M, Asai A, Koseki J, Kawamoto K, Tamari K, Li Z, Sakai D, Kudo T, Satoh T, Sato K, Motooka D, Okuzaki D, Doki Y, Mori M, Ogawa K, Ishii H. A convolutional neural network uses microscopic images to differentiate between mouse and human cell lines and their radioresistant clones. Cancer Res. 2018 Dec 1;78(23):6703–7. doi: 10.1158/0008-5472.CAN-18-0653. http://cancerres.aacrjournals.org/cgi/pmidlookup?view=long&pmid=30254144. [DOI] [PubMed] [Google Scholar]
  • 71.Sullivan HR, Schweikart SJ. Are current tort liability doctrines adequate for addressing injury caused by AI? AMA J Ethics. 2019 Feb 1;21(2):E160–6. doi: 10.1001/amajethics.2019.160. https://journalofethics.ama-assn.org/article/are-current-tort-liability-doctrines-adequate-addressing-injury-caused-ai/2019-02. [DOI] [PubMed] [Google Scholar]
  • 72.Shokri R, Stronati M, Song C, Shmatikov V. Membership Inference Attacks Against Machine Learning Models. Proceedings of the IEEE Symposium on Security and Privacy; SP'17; May 22-26, 2017; San Jose, CA, USA. 2017. [DOI] [Google Scholar]
  • 73.Vayena E, Blasimme A, Cohen IG. Machine learning in medicine: addressing ethical challenges. PLoS Med. 2018 Nov;15(11):e1002689. doi: 10.1371/journal.pmed.1002689. http://dx.plos.org/10.1371/journal.pmed.1002689. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 74.Drummond M, Griffin A, Tarricone R. Economic evaluation for devices and drugs--same or different? Value Health. 2009 Jun;12(4):402–4. doi: 10.1111/j.1524-4733.2008.00476_1.x. https://linkinghub.elsevier.com/retrieve/pii/S1098-3015(10)60778-4. [DOI] [PubMed] [Google Scholar]
  • 75.McCarthy J, Hayes PJ. Some philosophical problems from the standpoint of artificial intelligence. Read Artif Intel. 1981:431–50. doi: 10.1016/B978-0-934613-03-3.50033-7. doi: 10.1016/B978-0-934613-03-3.50033-7. [DOI] [Google Scholar]
  • 76.Matheny M, Israni S, Ahmed M, Whicher D. Artificial Intelligence in Health Care: The Hope, the Hype, the Promise, the Peril. E-Health Research. 2019. [2020-04-23]. https://ehealthresearch.no/files/documents/Rapporter/Andre/2019-12-AI-in-Health-Care.pdf.
  • 77.Fuchs S, Olberg B, Panteli D, Perleth M, Busse R. HTA of medical devices: challenges and ideas for the future from a European perspective. Health Policy. 2017 Mar;121(3):215–29. doi: 10.1016/j.healthpol.2016.08.010. [DOI] [PubMed] [Google Scholar]
  • 78.Facey K, Henshall C, Sampietro-Colom L, Thomas S. Improving the effectiveness and efficiency of evidence production for health technology assessment. Int J Technol Assess Health Care. 2015 Jan;31(4):201–6. doi: 10.1017/S0266462315000355. http://journals.cambridge.org/abstract_S0266462315000355. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 79.Pomey M, Brouillard P, Ganache I, Lambert L, Boothroyd L, Collette C, Bédard S, Grégoire A, Pelaez S, Demers-Payette O, Goetghebeur M, de Guise M, Roy D. Co-construction of health technology assessment recommendations with patients: an example with cardiac defibrillator replacement. Health Expect. 2020 Feb;23(1):182–92. doi: 10.1111/hex.12989. http://europepmc.org/abstract/MED/31691439. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 80.Goetghebeur MM, Wagner M, Khoury H, Levitt RJ, Erickson LJ, Rindress D. Bridging health technology assessment (HTA) and efficient health care decision making with multicriteria decision analysis (MCDA): applying the EVIDEM framework to medicines appraisal. Med Decis Making. 2012;32(2):376–88. doi: 10.1177/0272989X11416870. [DOI] [PubMed] [Google Scholar]
  • 81.Angelis A, Kanavos P. Multiple criteria decision analysis (MCDA) for evaluating new medicines in health technology assessment and beyond: the advance value framework. Soc Sci Med. 2017 Sep;188:137–56. doi: 10.1016/j.socscimed.2017.06.024. https://linkinghub.elsevier.com/retrieve/pii/S0277-9536(17)30393-3. [DOI] [PubMed] [Google Scholar]
  • 82.Wahlster P, Goetghebeur M, Kriza C, Niederländer C, Kolominsky-Rabas P, National Leading-Edge Cluster Medical Technologies ‘Medical Valley EMN’ Balancing costs and benefits at different stages of medical innovation: a systematic review of multi-criteria decision analysis (MCDA) BMC Health Serv Res. 2015 Jul 9;15:262. doi: 10.1186/s12913-015-0930-0. https://bmchealthservres.biomedcentral.com/articles/10.1186/s12913-015-0930-0. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 83.van der Wilt GJ, Gerhardus A, Oortwijn W. Toward integration in the context of health technology assessment: the need for evaluative frameworks. Int J Technol Assess Health Care. 2017 Jan;33(5):586–90. doi: 10.1017/S0266462317004354. [DOI] [PubMed] [Google Scholar]
  • 84.Poder TG, Bellemare CA. Importance of contextual data in producing health technology assessment recommendations: a case study. Int J Technol Assess Health Care. 2018 Jan;34(1):63–7. doi: 10.1017/S0266462317004469. [DOI] [PubMed] [Google Scholar]
  • 85.Poder TG. Using the health technology assessment toolbox to facilitate procurement: the case of smart pumps in a Canadian hospital. Int J Technol Assess Health Care. 2017 Jan;33(1):54–62. doi: 10.1017/S0266462317000125. [DOI] [PubMed] [Google Scholar]

Articles from Journal of Medical Internet Research are provided here courtesy of JMIR Publications Inc.

RESOURCES