Skip to main content
Frontiers in Psychology logoLink to Frontiers in Psychology
. 2021 Sep 28;12:710982. doi: 10.3389/fpsyg.2021.710982

Human, All Too Human? An All-Around Appraisal of the “Artificial Intelligence Revolution” in Medical Imaging

Francesca Coppola 1,2,, Lorenzo Faggioni 3,*,, Michela Gabelloni 3, Fabrizio De Vietro 3, Vincenzo Mendola 3, Arrigo Cattabriga 1, Maria Adriana Cocozza 1, Giulio Vara 1, Alberto Piccinino 1, Silvia Lo Monaco 1, Luigi Vincenzo Pastore 1, Margherita Mottola 4, Silvia Malavasi 4, Alessandro Bevilacqua 4, Emanuele Neri 2,3,, Rita Golfieri 1,
PMCID: PMC8505993  PMID: 34650476

Abstract

Artificial intelligence (AI) has seen dramatic growth over the past decade, evolving from a niche super specialty computer application into a powerful tool which has revolutionized many areas of our professional and daily lives, and the potential of which seems to be still largely untapped. The field of medicine and medical imaging, as one of its various specialties, has gained considerable benefit from AI, including improved diagnostic accuracy and the possibility of predicting individual patient outcomes and options of more personalized treatment. It should be noted that this process can actively support the ongoing development of advanced, highly specific treatment strategies (e.g., target therapies for cancer patients) while enabling faster workflow and more efficient use of healthcare resources. The potential advantages of AI over conventional methods have made it attractive for physicians and other healthcare stakeholders, raising much interest in both the research and the industry communities. However, the fast development of AI has unveiled its potential for disrupting the work of healthcare professionals, spawning concerns among radiologists that, in the future, AI may outperform them, thus damaging their reputations or putting their jobs at risk. Furthermore, this development has raised relevant psychological, ethical, and medico-legal issues which need to be addressed for AI to be considered fully capable of patient management. The aim of this review is to provide a brief, hopefully exhaustive, overview of the state of the art of AI systems regarding medical imaging, with a special focus on how AI and the entire healthcare environment should be prepared to accomplish the goal of a more advanced human-centered world.

Keywords: artificial intelligence, medical imaging, ethics, medico-legal issues, patient data, communication, psychology

Introduction

The term artificial intelligence (AI) was coined in 1956 to differentiate the intelligence of machines (generated by software development programs) from natural, human intelligence. In the past decade, AI algorithms have begun to influence many activities based on computer platforms, having a significant impact on daily life. These new technologies have aroused great interest among biomedical scientists, since AI has proven to be able to simplify the work of researchers and healthcare professionals, and to provide crucial information for the management of patients (e.g., early diagnosis, prediction of individual prognosis, and therapy personalization), which would realistically be very difficult or impossible to obtain without the support of such systems (Yu et al., 2018).

Radiology is one of the medical specialties with a greater interest in AI, since the latter can offer radiologists new tools for quantitative analysis and image interpretation in addition to offering automation and standardization of processes and procedures, which allow saving time and effort during fatiguing and/or repetitive tasks, improving diagnostic performance, and optimizing the overall workflow (Curtis et al., 2018; Hosny et al., 2018; Pesapane et al., 2018a; Yu et al., 2018; European Society of Radiology (ESR), 2019b; Grassi et al., 2019). However, this enthusiasm is paralleled by concerns of psychological, ethical, and medico-legal nature (including those related to the involvement of AI systems in patient management and the responsibilities that this may entail), as well as by the fear that AI could revolutionize radiologists’ jobs, possibly threatening their existence as specific professional figures (Gong et al., 2019; Savadjiev et al., 2019; van Hoek et al., 2019). More generally, it has been emphasized that as long as AI becomes more and more autonomous (e.g., able to talk, “think,” and actively participate in decision making), its role within complex relationships such as those between patients and physicians may be unclear to the human interlocutors, and new obstacles to decision making could arise due to AI acting as a “third wheel” between them (Triberti et al., 2020).

This review illustrates the main characteristics of AI systems and the current issues related to their use in the radiology profession.

Artificial Intelligence: Basic Concepts

Broadly speaking, AI encompasses the ability of hardware and software devices to autonomously mimic activities which have traditionally been deemed as specific to humans, such as learning and thinking. More than 60 years after its inception, AI has recently come back under the spotlight owing to the increasing availability of relatively low-cost computers capable of processing large amounts of data in real time, enabling the practical implementation of AI systems. In the medical field, AI mainly refers to the ability of systems to detect and analyze data related to patient clinical information and management with the aim of accomplishing a predetermined goal (Savadjiev et al., 2019).

Artificial intelligence systems can be broadly classified into strong (or general) and weak (or restricted). Strong AI systems can apply AI to resolve any problem, and as such, aim to mimic human intelligence. Conversely, weak AI denotes systems from which humans can take advantage to efficiently perform specific tasks (Zackova, 2015; Brink et al., 2017; Park and Park, 2018; European Society of Radiology (ESR), 2019b). More specifically, weak AI systems can improve their intrinsic ability to solve problems autonomously by means of progressive learning, starting from acquired information. This category includes most systems used in practice which are based on different machine learning (ML) techniques, including those of bio-inspired artificial neural networks.

Machine Learning

In 1959, Arthur Samuel gave a boost to the development of weak AI by introducing the concept of “ML,” defined as a subclass of AI systems which help the machine to learn and make decisions based on the data. To this end, the machine builds its own model from a subset of data used for training (Bishop, 2006; Litjens et al., 2017). Consequently, ML can make predictions on new data based on previous training without the need of being specifically programmed or recall previously defined models.

Another notable feature of ML is that the system performance increases with increasing experience of the system itself. In classic ML (which is used for classifying and interpreting data related to image analysis), data are labeled by human experts and organized according to their properties using statistical methods (Chartrand et al., 2017). In order for an ML algorithm to successfully reproduce the process of analyzing an image (e.g., a chest X-ray) by a radiologist, it must first be trained with a supervised approach starting from different labeled learning datasets (which contain many heterogeneous types of radiographic abnormalities), reinforced with different datasets each containing a class of abnormal findings (e.g., cardiac, mediastinal, pulmonary, and bone) and, if necessary, additionally reinforced with specific datasets for various subclasses of anomalies (e.g., congenital heart disease).

In general, ML is the highest expression of the power of a computer system. However, as in human learning, ML can also encounter some problems. For example, if the training dataset is poorly representative of the characteristics to be analyzed, an ML algorithm could learn from the training dataset in too much detail, leading to the problem of overfitting. In this case, non-significant statistical fluctuations of the same sample are cataloged by the learning model as separate data, which subsequently causes a worsening of the performance in analyzing new data (Duda et al., 2001; European Society of Radiology (ESR), 2019b). In diagnostic imaging, overfitting can be amplified by the possibility of non-pathological anatomical variants (such as accessory bones, or congenitally absent or hypoplastic structures).

The need for numerical accuracy and precision in processing radiological images represents one of the main challenges for the applications of ML systems in diagnostic imaging. Accuracy is essential for addressing the complexity of semantic aspects (related to the enormous variety of normal and pathological findings that an ML system could encounter in the analysis of images acquired on real patients) and technical issues due to differences among various imaging modalities. Another challenge is related to the large number of images which need to be processed from cross-sectional imaging modalities (even with the aid of semi-automatic algorithms), with magnetic resonance imaging (MRI) or multislice computed tomography (MSCT) being able to furnish hundreds or thousands of images per single dataset.

Artificial Neural Networks and Deep Learning

The term deep learning (DL) was first introduced in 1986 by Rina Dechter, and represents a form of ML which can yield better performance than classic ML (Figure 1). Compared to a traditional artificial neural network, in which the number of levels is limited and the nodes of one level are connected to those of the next level (“completely connected”), DL systems are generally made up of several specialized levels. The last levels are generally the only fully connected ones and combine functionalities learned to make decisions. Instead of requiring labeling or engineering of the properties, DL algorithms independently learn the most suitable characteristics for classifying the data provided, depending on the specific task (Chartrand et al., 2017; Philbrick et al., 2018).

FIGURE 1.

FIGURE 1

Schematic representation of the hierarchical relationship between AI (which, in general, describes systems capable of imitating human behavior), ML (a subset of AI, which describes systems capable of learning by experience), and DL (a subset of ML, which describes systems based on neural networks).

The commonest approach in image processing is represented by convolutional neural networks (CNNs), a particular type of neural network developed for the recognition of patterns within images, which can accept two- or three-dimensional images as input (Domingos, 2012). While the first CNN was implemented in 1980 by Fukushima (1980), CNNs were formalized as we know them now by LeCun et al. (1998). The introduction of advanced graphic processing units with the ability to process enormous amounts of data in parallel has made CNNs an essential tool for the development of modern DL algorithms (Hinton and Salakhutdinov, 2006; Trebeschi et al., 2017).

The two factors which mainly affect the functionality of CNNs are the power of the hardware and, most importantly, the availability of adequate data for the learning process. If computer power increases progressively over years or months, and can therefore only be relatively limiting, time and cost constraints make it difficult to find a solution to the problem of the low availability of well-structured datasets for training, which represents an actual hurdle to the development and diffusion of these systems (Napel et al., 2018).

Deep learning has proven to be a promising tool for the extraction of features from biomedical images (LeCun et al., 2015; Wang et al., 2017; Kermany et al., 2018; Lustberg et al., 2018). For that application, computational units are defined as levels integrated with each other to extract the intrinsic characteristics of the images. Using a CNN structured in a hierarchical manner, a DL system can, for example, extract the intrinsic characteristics of a neoplasm to build a model capable of providing prognostic or predictive information, having a clear potential impact on patient clinical management (Wang et al., 2019).

Artificial Intelligence in Medical Imaging: “Images Are More Than Pictures, They Are Data” (Gillies et al., 2016)

During its development, medical imaging has enjoyed great benefit from technological progress (Nance et al., 2013; Nguyen and Shetty, 2018), and the scientific relevance of the development of AI systems in radiology has been underscored by an ever-increasing number of publications on AI. For diagnostic imaging alone, the number of publications on AI has increased from about 100–150 per year in 2007–2008 to 1000–1100 per year in 2017–2018 (Tang, 2020).

Artificial intelligence systems can support medical decision-making processes related to requests for imaging tests, not only by means of the evaluation of the patient’s medical record and the accuracy of the radiological examinations, but also by guiding the choice of the most suitable diagnostic modality. Of note, AI algorithms can be programmed to work in keeping with the appropriateness criteria developed and approved by scientific societies (such as those developed by the American College of Radiology) in order to maximize the adherence to validated criteria (Blackmore and Medina, 2006; American College of Radiology (ACR), 2021 Reporting and Data Systems).

Furthermore, AI has opened new perspectives on how to make the most of the information which can be obtained from biomedical imaging for a more in-depth understanding of the various pathological processes, aimed at more effective diagnostic and therapeutic management. Once trained with appropriate learning datasets, AI systems can analyze biomedical images with the aim of recognizing specific characteristics (either visible or invisible to the human eye) and build probabilistic models capable of detecting abnormal findings (Dodd, 2007; Sardanelli et al., 2010).

Automated image interpretation is one of the potential radiological applications of AI which has been received with the greatest enthusiasm. Rajpurkar et al. (2018) illustrated an AI algorithm with a comparable accuracy to that of human radiologists for diagnosing pneumonia on chest X-rays in a public dataset. Similar experiences have been reported for the detection of vertebral fractures on plain spinal radiography (Murata et al., 2020), the diagnosis of tuberculosis (Lakhani and Sundaram, 2017), and the estimation of bone age (Dedouit et al., 2015). More generally, different DL methods have been applied to biomedical image analysis (Ierardi et al., 2016; Trimboli et al., 2018; Villanueva-Meyer et al., 2019) and successfully used with various imaging modalities, such as breast (Kallenberg et al., 2016; Zanotel et al., 2018; Geras et al., 2019; Hickman et al., 2021) and cardiac imaging (van Assen et al., 2020), MSCT (Lerouge et al., 2015; Kooi et al., 2017; Hu et al., 2020), MRI (Havaei et al., 2017; Ariji et al., 2019; Figure 2), as well as in interventional radiology (Gurgitano et al., 2021). AI can also be helpful to quantify lung involvement and predict prognosis in patients with COVID-19 pneumonia (Belfiore et al., 2020; Akram et al., 2021; Cappabianca et al., 2021), and Harmon et al. (2020) recently found that a series of DL algorithms trained in a diverse multinational cohort of 1280 patients can achieve up to 90.8% accuracy, with 84% sensitivity and 93% specificity in detecting COVID-19 pneumonia on chest CT examinations of 1337 patients. Other AI applications allow prioritizing the reporting of certain exams (e.g., urgent brain CT scans in patients with hemorrhagic stroke), thus optimizing the workflow and avoiding diagnostic delays, especially in situations in which members of the radiology department are busy with other tasks (Ngo et al., 2017; European Society of Radiology (ESR), 2019b). However, there are currently no commercial solutions available which can independently interpret images and generate a report.

FIGURE 2.

FIGURE 2

Evaluation of a CNN-based DL algorithm (PMetNet) for detection of occult peritoneal metastases (PM) from gastric cancer on preoperative CT images. (Upper) Following tumor segmentation on original CT images (tumor contours highlighted in red, tumor region of interest [ROI] shown within magnified tumor location [boxed in yellow]), the gradient-weighted class activation mapping (Grad-CAM) was used to highlight which areas of the CT image are important in generating a particular prediction. The numbers 0.99 and 0.01 represent the predicted probability of PM. (Lower) Area under the receiver operating characteristic curves (AUCs) derived from the PMetNet and clinicopathological factors for the diagnosis of occult PM in the training cohort and in two validation cohort. The discrimination performance of PMetNet was substantially higher than conventional clinicopathological factors, and in multivariable logistic regression analysis, PMetNet was an independent predictor of occult PM. Adapted from Jiang et al. (2021) under Creative Commons Attribution 4.0 (CC BY 4.0) license (https://creativecommons.org/licenses/by/4.0/).

The ability of extracting structured and categorized data from existing radiological archives [radiology information system (RIS) and picture archiving and communication system (PACS)] is an essential requirement for the development and dissemination of AI in radiological environments. In fact, the training of AI systems usually requires enormous amounts of data which should be as accurate and correctly categorized as possible. However, to date, most radiological reports are written in the form of an unstructured narrative text which greatly complicates the extraction of information, even if the aim is to create AI systems based exclusively on clinical data. The latter could be resolved with the adoption of structured reporting (SR), which, if properly implemented, would allow the exchange of information with a common lexicon and semantics. While there is extensive evidence that SR has several advantages over conventional unstructured reporting (including better clarity, improved communication with patients and referring physicians, higher productivity, and ease of data mining), it has seen a relatively slow diffusion so far due to radiologists’ fears, among others, that it could diminish their autonomy and professional reputation with respect to patients or non-radiologist specialists (Marcovici and Taylor, 2014; Faggioni et al., 2017; Coppola et al., 2021). In this context, constructive interaction between medical radiologists and other specialists, industries, and institutions would be desirable to promote a large-scale dissemination of the SR, offering a decisive stimulus for additional development of AI in the radiological field (Bosmans et al., 2015; Pinto Dos Santos and Baeßler, 2018; European Society of Radiology (ESR), 2019b).

A topic of interest for both the biomedical industry and research is the prospect of using AI to optimize biomedical image acquisition protocols, with potential advantages in terms of patient safety and health management costs. For example, some AI algorithms allow obtaining equivalent or even superior results as compared to commercial non-AI-based solutions for noise reduction in MSCT and positron emission tomography (PET) examinations, allowing the acquisition of diagnostic images with a significantly lower radiation exposure than conventional protocols (Zhu et al., 2018; Shan et al., 2019).

Radiomics is a field of research which has become very popular in the era of modern precision medicine. Radiomics refers to the established use of ML techniques applied to the analysis of radiological images. Radiomics is often defined as “the extraction of a large number of quantitative features from conventional biomedical images in order to obtain data that can be used in clinical decision support systems to improve diagnostic, prognostic, and predictive accuracy” (Choy et al., 2018). A radiomic model can reveal the value of biomarkers extracted from the images (which are quantifiable by means of applying mathematical and statistical models, even of considerable complexity), but it can also extend to a so-called hybrid model, including other data not from images (e.g., from clinical data or laboratory parameters). In any case, these models are able to provide information not obtainable with standard radiological semeiotics, such as those related to the early response to treatment, the prediction of the biological aggressiveness of a neoplasm, the existence of molecular targets for any targeted therapies, up to the prediction of the individual prognosis, and the personalization of therapies (Gillies et al., 2016; Lambin et al., 2017; Choy et al., 2018; Bi et al., 2019; Liu et al., 2019; Rogers et al., 2020; Zerunian et al., 2021). A notable feature of radiomics consists of the possibility of obtaining, in a repeatable and non-invasive way, information about a tissue in its entirety in contrast, for example, to what happens with a classic biopsy which is invasive and limited to a portion of tissue, with the risk of collecting a sample that is not representative of the heterogeneity of the lesion (Lambin et al., 2017; Abdollahi et al., 2019; Nazari et al., 2020; Zerunian et al., 2021). The radiomics approach can also be extended to the analysis of the genetic structure of a tissue (e.g., neoplastic), which is referred to as radiogenomics (King et al., 2013; Pinker et al., 2018; Story and Durante, 2018; Gabelloni et al., 2019; Lo Gullo et al., 2020). A radiomic biomarker is made up of a set of characteristics (or features) extracted from the image, it is represented by a mathematical equation, known as a “signature,” and its value can be calculated using dedicated programs, starting from images acquired using routine protocols.

However, even if many of these tools are easy for operators to use and allow extracting radiomic features in a relatively short time, their diffusion is currently still limited by several factors. These include the enormous amount of work often necessary for image segmentation and the difficulty in ensuring the adequate quality of the data entered to obtain consistent results, also considering the inevitable differences both between the image acquisition protocols, and different machines and imaging centers (Stoyanova et al., 2016).

Artificial Intelligence and Humans

Several authors have hypothesized that AI systems might shortly be able to replace medical radiologists in their professional activity (Rizzo et al., 2018). The key question is: will AI be able to replace radiologists in the observation, characterization, and quantification tasks that they currently accomplish using their cognitive skills? (Beregi et al., 2018; Tajmir and Alkasab, 2018; Mendelson, 2019; Miller and Brown, 2019; Rubin, 2019). The short answer is: NO. However, as argued by Dr. Curtis Langlotz at the European Congress of Radiology in 2018: “AI won’t replace radiologists, but radiologists who use AI will replace radiologists who don’t” (Krittanawong, 2018), and this concept could be generalized to all fields of healthcare (Meskó et al., 2018). In this context, it is important to point out that the final decision regarding patient diagnosis is still autonomous and the responsibility lies with the radiologist, not AI systems. What is most likely to change will be the use of information not only derived from morphological analysis in the formulation of the diagnosis, but also from the numerical values provided by AI. These refer directly to statistically significant distributions of the pixel values of the image which are not perceptible to the naked eye.

Both radiologists and AI systems must follow essential rules and principles for optimal patient management. Several issues are related to the proper use of AI in clinical practice and include (but are not limited to) the following:

  • Data (including generation, recording, maintenance, processing, dissemination, sharing, and use)

  • AI algorithms used to process patients’ data for a specific task

  • Practices (including responsible innovation, programming, security, formulation, and implementation of ethical solutions)

  • Communication (including the tools through which the information obtained from AI systems is provided to patients, as well as the management of psychological problems arising from them, which cannot be handed over to a computer system) (Meskó et al., 2018).

Various aspects of data ethics can be recognized, including informed consent, privacy and data protection, data ownership, objectivity in managing data, and the likelihood that a gap may exist between those who have the resources to manage and analyze large amounts of data and those who do not. In addition, the operation of AI systems integrated into big data networks raises ethical and legal issues related to patient-specific consent, data sharing, privacy and security protection, and the availability of multi-layered access to fully or partially anonymized health information.

Artificial Intelligence Overconfidence and Medico-Legal Issues

As stated by Kulikowski (2019), “whether a good ethical human can work with an AI and remain ethical is a major open problem for all of us that will have to be confronted not only scientifically, but also in a socially acceptable and humanistic way in clinical informatics.” Hence, ethics should always guide radiologists (and physicians in general) in deciding when to rely on AI, so as to avoid improper applications of it which may have a harmful impact on both healthcare operators and patients.

One of the main biases which can hamper the use of AI in diagnostic imaging is the automation bias, which can be defined as the propensity to favor a machine-generated diagnosis over evidence derived from scientific knowledge and the physician’s own expertise. This leads to the so-called omission and commission errors. Omission errors occur when the physician, deeming AI flawless, does not notice (or outright ignores) the fallacy of one of its tools. On the other hand, commission errors occur when a machine’s decision is accepted, even in the face of contrary evidence. The risks of automation bias can be amplified in realities which suffer from a lack of medical personnel, since there may not be any radiologist to double-check the AI results (Geis et al., 2019). It has also been observed that automation could engender overreliance by its users (due to its advantages in terms of increased efficiency), and in the long term, lead to the so-called deskilling, with physicians losing their ability to autonomously perform tasks which have become automated (Cabitza et al., 2017). Harada et al. (2021) performed a randomized controlled study aimed to explore the prevalence of AI diagnoses in physicians’ differential diagnoses when using an AI-driven diagnostic decision support system (DDSS) based on the information entered by the patient before the clinical encounter, showing that at least 15% of physicians’ differential diagnoses were affected by the differential diagnosis list in the AI-driven DDSS. While many clinicians hope that AI will free them to focus on patient interaction, research on the overreliance of technology in medicine has found that the increased use of electronic health records has led to a prioritization of physician–technology interactions over physician–patient interactions, leading to decreased patient satisfaction, a scenario that could foreshadow the role of AI in patient care (Lu, 2016; Ross and Spates, 2020).

There is a still highly unmet need for specific guidelines, policies, and recommendations offering an ethical framework that can guide the use and implementation of AI technologies in an increasingly broad spectrum of clinical applications, which are progressively emerging as an effect of technological evolution, but also carry substantial psychological and ethical implications. Some of such potential applications include, for instance, AI in assisted reproductive technologies for human embryo selection in vitro fertilization (Dirvanauskas et al., 2019), and optimization of clinical trials of innovative stem cell and gene therapies in pediatric patients by precise planning of treatments, simplifying patient recruitment and retention, and lowering their complexity and costs (Sniecinski and Seghatchian, 2018). However, despite efforts by scientists, healthcare professionals, administrative managers, and lawmakers, so far very few countries worldwide have adequate and critical governance frames allowing best understanding and steering AI innovation trajectories in healthcare (Dzobo et al., 2020).

Such scenario is further complicated by the sweeping speed at which AI techniques are being developed or sometimes used, even before the publication of appropriate policies and guidelines, which might leave users confused about how to best integrate this new technology into their practice. This implies that updated regulatory policies and continuing education of all users (including adequate information to patients about the purposes, rights, and legal terms related to the use of AI for their health management) should be promoted, as AI systems are poised to become more widely available, complex and powerful. To this purpose, it is noteworthy that the majority of Singaporean radiology residents joining a national multiprogram survey thought that since AI will drastically change radiology practice, AI/ML knowledge should be taught during residency (84.8% of survey participants), and this was as important as imaging physics and clinical skills/knowledge curricula (80.0 and 72.8%, respectively) (Ooi et al., 2021). From a psychological standpoint, it has been observed that openness to experience is associated with higher trust toward robots and AI, as well as having a degree in technology or engineering, exposure to robots online, and robot use self-efficacy (Oksanen et al., 2020), highlighting the importance of technology knowledge in addition to personal differences in building AI confidence.

A key medico-legal aspect regarding the use of AI in healthcare is the responsibility for the decision-making processes upon which the patient’s health depends. In the absence of specific regulations, there may be ethical and medico-legal issues where an AI system is involved in the process and may suggest solutions (right or wrong); however, the final decision (also right or wrong) is, and will always be, the responsibility of the physician who is legally responsible (Mittelstadt and Floridi, 2016; Price et al., 2019, 2021; Reddy et al., 2019; Neri et al., 2020). Price et al. (2019, 2021) provided an in-depth analysis of the potential legal outcomes related to the use of AI in healthcare under current law (Figure 3).

FIGURE 3.

FIGURE 3

Comparison of potential legal outcomes under current law according to analysis of Price et al. (2019) and empiric study findings of Tobia et al. (2021). ★ = agreement that physician decision was reasonable (highest is ★★★★; lowest is ★). Greater agreement indicates lower likelihood of liability; ☑ = study results confirming Price et al.’s analysis of current tort law; Inline graphic study results suggesting that jury outcome may also be liability; Inline graphic study results suggesting that jury might decide no liability. Reproduced from Price et al. (2021). © SNMMI.

Another problem is the opacity related to AI models being mostly a “black box” without a universal understanding of their inner workings, which are not acceptable for decision support solutions (especially in healthcare) and may lead to ethical and legal risks and liability issues, as well as undermine patients’ and physicians’ confidence into AI (Valiuškaitė et al., 2020; Tohka and van Gils, 2021). Putting this issue in the context of radiological practice, radiologists would be asked to monitor AI system outputs and validate AI interpretations, so they would risk carrying the ultimate responsibility of validating something they cannot understand (Neri et al., 2020). There is a clear difference between statistical and clinical validation, and hence achieving adequate informed consent is problematic when the algorithmic decision-making process is opaque to clinicians, patients, or courts (Martinez-Martin et al., 2018; Arnold, 2021). Actually, while the number of published articles on the applications of AI in medical imaging and other medical specialties is steadily increasing, so far only few AI applications have been validated for clinical use, partly due to the difficulty of using AI projects on a large scale in real-life clinical practice, poor adherence to scientific quality standards (Nagendran et al., 2020; Park et al., 2020), and clinical validation issues. While the ongoing development of AI has generated considerable hype and highly optimistic expectations in the scientific community, such enthusiasm is often curbed by the reality of proper performance assessment, which is not trivial (requiring an understanding of the problem and data) and is often costly (data needs to be reserved) and time consuming (Tohka and van Gils, 2021). According to some researchers, the overall problem is wide and ultimately originates from inappropriate experimental design and hypothesis testing procedures, including so-called Hypothesizing After the Results are Known (aka HARKing) practices (Gencoglu et al., 2019; Tohka and van Gils, 2021).

The process of technology and infrastructure development requires close multidisciplinary cooperation among governmental institutions, research centers, healthcare professionals, and industry. In this context, a potential solution could involve enrolling AI experts in radiology units to act as a link between AI systems and radiologists who, in turn, should be trained to use those systems independently. To ensure that the approach to an innovative and potentially destructive technology is properly managed, radiologists will have to develop strategies not based on prejudice, but specifically adapted to the peculiar characteristics of AI systems, their technical/scientific development, their implementation by the industry, and their actual diffusion.

Data Confidentiality and Regulation Policies

Using AI systems for diagnosing diseases (including life-threatening or invalidating diseases, with a potentially dramatic impact on the physical and psychological well-being of patients and their families) and finding the most appropriate therapeutic approach implies that these systems should have the highest grade of reliability and dependability. To this end, the following requirements should be met:

  • The largest possible amount of data (both imaging and non-imaging-related) should be shared.

  • The quality and integrity of these data should be as high as possible, avoiding errors due to poor image quality, mislabeling, and over- and underfitting.

  • The anonymity and depersonalization of data must be guaranteed so as to ensure that the individual(s) who has/have consented to their use can be traced.

In recent years, several authors have discussed the requisites for the reliable innovation of AI by means of attaining the most important ethical principles. These principles have been embodied in the laws of the various countries throughout the world, and while an overarching political vision and long-term strategy for the development of a “good AI” society are currently lacking, this process has been characterized in broad terms by the search for a tradeoff between AI technological innovation and regulation (Cath et al., 2018; Pesapane et al., 2018b; Monreale, 2020).

As mentioned by Cath et al. (2018), the political attitude of the United States toward the implementation of AI in healthcare can be summarized by the sentence: “Letting a thousand flowers bloom,” whereas that of the European Union (EU) can be described as: “European standards for robotics and AI,” and the UK approach (“Keep calm and commission on”) stands in an approximately intermediate position between the US and the EU policies. In the United States, a “Silicon Valley” model oriented toward the more liberal regulation of ethical issues and based on the “move fast, break things first, apologize later” approach has prevailed (Armitage et al., 2017) and, in 2020, the Trump administration published a guide for AI application which discouraged any action resulting in limiting innovation and technological progress (Vought, 2020). At the other extreme, the EU policy points to strictly codifying regulations based on ethical principles. While this policy has raised the objection that it could hinder AI innovation, the European Commission sees the codification of ethical principles for AI use as a competitive advantage which will promote consumer confidence in their products and harmonize their adoption across the EU (Monreale, 2020). Data protection in the EU is regulated by General Data Protection Regulation (GDPR) EU 2016/679 and other EU directives for confidential data protection, which is of paramount importance in case of the AI/ML systems being trained on personal healthcare data (Voigt and Von dem Bussche, 2017). In this respect, the principles of “privacy by design” described by Cavoukian (2009) and updated by Monreale et al. (2014) could be applied, in the perspective of promoting research and innovation while taking care and full responsibility of the protection of the personal data, rights, and freedom of EU citizens.

It is especially important that the requirements for privacy protection are fulfilled during the process of data extraction for the training of ML algorithms, avoiding the potential risks related to illegal access either to confidential data during training or to the ML model used for clinical patient management (Brundage et al., 2018; Pesapane et al., 2018b; Monreale, 2020). On the other hand, current data protection laws may pose a significant limitation for researchers who develop and use ML algorithms, resulting in a lack of generalization of training which has so far prevented a more widespread application of such algorithms into clinical practice by healthcare providers across the world. This holds especially true for rare diseases, for which the accuracy of ML algorithms could be limited due to the relatively small amount of data for algorithms to train on and data collection is inherently slow due to a low disease prevalence. Similarly, algorithms which predict outcomes from genetic findings may lack generalizability if there are only a limited number of studies in certain populations (Kelly et al., 2019; Abayomi-Alli et al., 2021; Tohka and van Gils, 2021).

Machine learning models are programmed based on de-identified data, i.e., those which do not directly allow identifying an individual in a univocal manner. Unfortunately, in some contexts, de-identification is not sufficient to protect a person’s privacy, since individuals could be indirectly reidentified by means of the correlation of the de-identified data with public data (Jones, 2018), prompting the adoption of more advanced solutions aimed at fully protecting patient anonymity, such as k-anonymity (Sweeney, 2002). From a more general ethical and legal viewpoint, while patient data stored in electronic health records may be de-identified and, through data linkage, generate beneficial research outcomes, this may create a tension between beneficence (for the public) and private confidentiality, overriding contemporary notions of privacy and confidentiality according to the duty of “easy rescue,” particularly in circumstances of minimal risk as defined by research regulators (Porsdam Mann et al., 2016; Arnold, 2021). Moreover, a study from the University of California, Berkeley, suggests that progress in AI has rendered the privacy standards set by the Health Insurance Portability and Accountability Act of 1996 (HIPAA) obsolete (Na et al., 2018). The important conclusion is that privacy standards associated with the current legal and regulatory framework should be revisited and reworked, such that the advances of AI and its impact on data privacy as it pertains to healthcare are factored in (Na et al., 2018; Ahuja, 2019; Kulkarni, 2021).

An additional risk regarding a breach of patient confidentiality could derive from so-called “membership inference attacks,” i.e., malicious attacks toward AI algorithms which are aimed at detecting the confidential data used to build the algorithm (Shokri et al., 2017). Actually, the implementation of AI systems means access to sensitive health data, which intrinsically always carries the risk of cyberattacks, posing a substantial risk on the privacy of patients (especially those with lower education and financial income; Bilen and Özer, 2021) and requiring a guaranteed level of robustness against such attacks (Catak et al., 2021; Zhou et al., 2021). Attacks on AI systems can undermine diagnostic accuracy, administer lethal drug doses, or sabotage critical moves in an operation, and in the area of diagnostic imaging, they can manipulate data entering AI systems (so-called “input attacks”), leading to false diagnosis and altered patient care and/or reimbursement (Finlayson et al., 2019; Kiener, 2020; Myers et al., 2020). The malware can obtain personal information by means of query and repersonalization of the data within the algorithm, and most strategies aimed at offering protection against such privacy violations rely on methods based on differential privacy, i.e., a privacy model based on the concept of data perturbation (Bugliesi et al., 2006). While several solutions have been proposed to forecast, prevent, and mitigate threats from malicious uses of AI technologies, a coordinated action of all involved stakeholders (including researchers, engineers, and AI users) has been advocated to manage what is expected to become a long-term equilibrium between AI attackers and defenders (Brundage et al., 2018).

Accessibility of Artificial Intelligence Services

An important issue that should receive special attention is related to the possibility that the access to AI systems may be not equal for all patients or healthcare professionals. In fact, smaller facilities and academic centers with fewer resources may lack the means to acquire (and skills to use) complex and more performing AI systems. Furthermore, if AI were to be developed and exclusively owned by large entities in the private sector, this would likely further restrict its spread to a wider public. Collaboration between academic institutions and the public and private sectors has been advocated to foster the development of both workforce and AI applications in healthcare (Mikhaylov et al., 2018; Ishii et al., 2020).

The so-called “digital divide” can be classified as global, social, and democratic (Srinuan and Bohlin, 2011); in any case, it invariably implies that affected subjects are excluded from the benefits of technological progress and innovation. A realist review in general practice by Huxley et al. (2015) showed that while digital communication technology offers increased opportunities for marginalized groups to access health care, it cannot remove all barriers to care for these groups, and actually they will likely remain disadvantaged relative to other population groups after their introduction. There is a risk that such phenomenon may occur in a previously unseen fashion with AI, with factors including age, gender, health condition, level of education, or financial income possibly leading to unequal access to AI systems.

An increasingly recognized issue is the potential for bias of AI systems with respect to certain population subgroups with a lack of diversity (e.g., in age, ethnicity, socioeconomic background, etc.) if algorithms have been developed on datasets which under- or over-represent them (Caliskan et al., 2017; Reddy et al., 2019; DeCamp and Lindvall, 2020; Hickman et al., 2021). Algorithmic bias may occur in ML systems for healthcare, perhaps predicting a greater likelihood of disease on the basis of gender or race when those are not actual causal factors (Davenport and Kalakota, 2019). Other issues could arise from discriminatory behaviors toward socially weaker individuals, from the need to gain the physicians and patients’ trust in a context where AI systems process biomedical data and play a crucial role in clinical management, or from the duty of providing concrete rights of access to services to each patient. The communication of medical information, rules for the use of data, and requirements for institutional review committees may need to include new possibilities for patient data management (Kohli and Geis, 2018).

Communication and Psychological Issues

As outlined previously, while AI is supposed to offer radiologists substantial aid in their professional activity, predictions that it will replace radiologists in a more or less distant future are unfounded since the professional role of radiologists involves many tasks which cannot be accomplished by computers alone, including carrying out interventional radiology procedures, performing a clinical-radiological correlation in image interpretation, interpreting complex findings, and communicating them to colleagues and/or patients (Russell and Bohannon, 2015; Price et al., 2019). However, radiologists will have to improve their relationship with patients in the AI era to avoid any patient discomfort due to a lack of empathy and of a human reference figure during all the steps of a radiological procedure. These range from the patient’s admission to the communication and discussion of imaging findings (the latter being a source of considerable psychological stress for patients and, hence, a task which could not be assigned to any, however perfect, AI algorithm). Moreover, reaching a diagnosis may often involve the use of multiple imaging techniques which are proposed by the radiologist (in combination with clinical and laboratory data, as well as with other non-radiological tests), and the overall interpretation of imaging findings is a complex task which requires a global assessment of the patient’s condition and as such cannot be demanded to a computer system. In this context, the ability of AI systems to not only improve the detection and characterization of diseases (e.g., cancer) but also guide treatment and predict individual patient outcomes and prognosis (Bi et al., 2019; Rogers et al., 2020) can create additional issues related to the complexity of communicating and discussing topics with a high emotional impact (Butow and Hoque, 2020).

While AI allows saving time regarding the diagnostic and therapeutic decision process, the latter could actually be delayed if the role of AI is not taken into account in the consultation between physician and patient (Pravettoni and Triberti, 2019) (who may wish, and has the right to, know the implications of using AI in his/her clinical management), or if AI conclusions need to be revised by human doctors (especially when important decisions are to be made based on such conclusions) (Triberti et al., 2020). Moreover, as mentioned above, the poor explainability of most current AI systems (which are undoubtedly characterized by a high degree of complexity) and their lack of transparency could engender anxiety, distrust, or outright hostility with respect to AI in patients and clinicians (Bi et al., 2019). The relationship between patient and physician is a complex and profound psychosocial interaction characterized by mutual knowledge, trust, loyalty, and regard, so that human interaction will remain essential for patient-centered care due to the uniqueness of “human touch,” consisting of peculiar features (such as empathy or the ability to be in tune with other people’s thoughts and feelings) (Honavar, 2018). To this regard, it is known that a better communication between patients and physicians is associated with lower patient anxiety, fewer malpractice claims, and improved quality of life (Levinson et al., 2010). As to patients’ trust in AI performance, Juravle et al. (2020) reported three online experiments showing that given the option of receiving their diagnosis from AI or human physicians, patients trusted those latter more for both first diagnoses and a second opinion for high risk diseases, and their trust in AI did not increase when they were told that AI outperformed the human doctor, but the trust in AI diagnosis increased significantly when participants could choose their doctor.

Owing to their pivotal role in the diagnostic process, radiologists are often the first healthcare professionals who are asked by patients about their imaging findings (and hence find themselves to deal with patients’ emotional reaction), and from whom patients expect a direct communication of their imaging findings (Berlin, 2007; Capaccio et al., 2010; Cox and Graham, 2020). The use of AI systems with the ability to provide additional information which may have a significant impact on patient management and overall life (e.g., eligibility to specific treatment options, prognosis, etc.) will entail for radiologists more stringent requirements in terms of communication skills and psychological balance, as well as a high degree of constructive interaction and feedback with other medical and non-medical specialists involved in patient care.

A detailed knowledge of the main features of AI (including its technical background, its current fields of application, and its psychological and legal implications) is a preliminary condition for its usage as a mature professional tool (Kobayashi et al., 2019; Savadjiev et al., 2019; Sogani et al., 2020). In a nationwide online survey among members of the Italian Society of Medical and Interventional Radiology (SIRM), most radiologists (77%) were favorable to the adoption of AI in their working practice, with a lower diagnostic error rate and work optimization being main perceived advantages, whereas the risk of a poorer professional reputation compared with non-radiologists was seen as one major downside (60% of survey respondents). However, about 90% of surveyed radiologists were not afraid of losing their job due to AI, and less than 20% of them were concerned that computers will replace radiologists for reporting of imaging examinations (Coppola et al., 2021). To this respect, it is worth mentioning that while most medical students surveyed by Gong et al. (2019) were discouraged from considering the radiology specialty out of anxiety that AI could potentially displace radiologists, in that same study prior significant exposure to radiology and high confidence in AI understanding were associated with a lower anxiety level, suggesting that professional education can have a significant impact on the psychological attitude of physicians toward AI.

Moreover, in the aforementioned SIRM survey and in a EuroAIM survey aimed at assessing the perceived impact of AI in radiology among European Society of Radiology (ESR) members, most respondents believed that if AI systems will allow radiologists to save time, such time should be used to interact with other clinicians or patients, thus improving personal interaction and communication (European Society of Radiology (ESR), 2019a; Coppola et al., 2021). Similar findings were reported in a French survey including 70 radiology residents and 200 senior radiologists, whose main expectations about AI included a lower risk of imaging-related medical errors and an increase in the time spent with patients (Waymel et al., 2019).

In light of the above, AI could alleviate radiologists’ traditional work burden by undertaking tasks that could better be performed by computers, while giving them the opportunity to invest time and resources for other tasks that are better or uniquely accomplished by humans, such as interpreting imaging findings in the full width and complexity of a real clinical context, enhancing communication with patients and clinicians, supervising the correct operation and usage of AI systems, and being actively engaged in research (including AI-assisted data mining for big data handling and management of large-scale clinical trials) and quality optimization of the whole healthcare process. Like pathologists (who also extract medical information from images), radiologists will have an inescapable opportunity to leave once for all the stigma of “invisibility” which has often overshadowed the perception of their professional role by patients and clinicians in the past (Glazer and Ruiz-Wibbelsmann, 2011), and to take on a pivotal role in patient care as information specialists, adapting incrementally to AI and retaining their own services for cognitively challenging tasks and interaction with patients and clinicians (Jha and Topol, 2016; Recht and Bryan, 2017). Likewise, also clinicians need not fear AI as a potential enemy who could harm their professional reputation in the patients’ eyes or their jobs in the future, but they should leverage its power to tackle computationally and labor-intensive tasks better than humans and to concentrate on those tasks which require human action (Ahuja, 2019). Therefore, an enhanced professional role could be envisaged for both radiologists and clinicians, requiring more advanced and specific skills (Recht and Bryan, 2017; Krittanawong, 2018; Ahuja, 2019; Waymel et al., 2019), despite fears that AI taking over professional tasks once performed by humans could, in the long run, lead to deskilling of human physicians (Bisschops et al., 2019; Campbell et al., 2020; Panesar et al., 2020). AI could actually help radiologists and clinicians make the most of their own specialty knowledge and competence in a medical science of rapidly increasing complexity (where “diseases do not respect boundaries” between medical specialties and require the cooperation of multiple specialists; Deo, 2021), avoiding misunderstandings and “turf wars” due to poor communication and confusion regarding their specialty-specific roles in patient management, and possibly fostering the adoption of AI-augmented multidisciplinary teams (including software engineers and data scientists among participants) for clinical decision making (Di Ieva, 2019; Lee and Lee, 2020; Martín-Noguerol et al., 2021).

Other potential issues of AI in the physician–patient relationship include misunderstanding (since a disagreement between the physician and AI can cause confusion, and the patient may not recognize who has the real authority in the care management) and alienation due to the physician or patient feeling excluded from the contribution of AI. To this regard, it should be considered that AI is deficient in emotional intelligence, whereas a physician has skills, beliefs, and subjective perceptions which can shape the communication with the patient, thus seeking an adequate patient’s understanding of the disease and its related treatment options as the main aim of the communication process (Oh et al., 2017; Keskinbora, 2019; Pravettoni and Triberti, 2019; Triberti et al., 2020).

It has been observed that once digital and objective data will have become accessible to both caregivers and patients, the so-called “digital health” (of which AI is a major component) will lead to an equal level of physician–patient relationship with shared decision-making and a democratization of care (Meskó et al., 2017). However, it is possible that while some patients could accept or even require AI as an additional tool for decision making in their own medical care, others would not accept its use in decision-making (Meskó et al., 2018), thus stressing the need for setting out shared policies aimed to a rational utilization of AI in patient management. A recent study on patients’ perception about the use of AI for skin cancer screening as assessed by means of a semistructured interview revealed that most of them were favorable to AI and believed that it may improve the quality of care, but only if implemented in a manner which preserves the integrity of the human physician–patient relationship (Nelson et al., 2020). Again, direct physician–patient communication must be considered as an integral part of care delivery which cannot be substituted by a machine, and as Krittanawong has pointed out: “AI cannot engage in high-level conversation or interaction with patients to gain their trust, reassure them, or express empathy, all important parts of the doctor–patient relationship” (Krittanawong, 2018).

eXplainable Artificial Intelligence and Causability: Forthcoming Steps for Artificial Intelligence to Enter Maturity?

In conclusion, while it is undeniable that AI will sooner or later affect healthcare and the professional role and work of healthcare providers, physicians should neither uncritically accept nor unreasonably resist developments in AI, but they must actively engage and contribute to an iterative discourse to preserve humanitarian concerns in future models of care (Arnold, 2021). In this context, it is clear that the sustainable use of AI involves keeping in mind its fields of applicability and limitations, thus envisaging a future where its capabilities and advantages integrate (rather than supplant) human intelligence.

The main future goal is to make AI capable of interacting with operators in a meaningful and easily accessible manner. In this context, eXplainable Artificial Intelligence (xAI) has emerged as a new discipline which tries to fulfill the need for causability in the medical domain; in the same way that usability encompasses measurements for the quality of use, causability encompasses measurements for the quality of explanations produced by xAI. Multi-modal causability is especially important in the medical domain, since results are often achieved by means of multiple different modalities. The key for future human–AI interfaces is to map explainability with causability, and to allow a domain expert to ask questions so as to understand why AI has come up with a result, and also to ask “what-if” questions (counterfactuals) to gain insight into the underlying independent explanatory factors of a result (Holzinger, 2021).

Author Contributions

FC, LF, and MG: conceptualization. MG, GV, AP, and SLM: methodology. FDV, VM, AC, MAC, and LVP: literature review. FC, LF, FDV, and VM: writing–original draft preparation. MG, MM, SM, and AB: writing–review and editing. EN and RG: supervision and guarantee of scientific integrity. All authors have read and agreed to the published version of the manuscript.

Conflict of Interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s Note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Acknowledgments

The authors wish to thank Lorenzo Tumminello for providing insights about the psychological impact of AI on physicians and patients and Gerard Goldsmith for kindly revising the English language of the manuscript.

References

  1. Abayomi-Alli O. O., Damaševičius R., Maskeliūnas R., Misra S. (2021). Few-shot learning with a novel voronoi tessellation-based image augmentation method for facial palsy detection. Electronics 10:978. 10.3390/electronics10080978 [DOI] [Google Scholar]
  2. Abdollahi H., Mofid B., Shiri I., Razzaghdoust A., Saadipoor A., Mahdavi A., et al. (2019). Machine learning-based radiomic models to predict intensity-modulated radiation therapy response, Gleason score and stage in prostate cancer. Radiol. Med. 124 555–567. 10.1007/s11547-018-0966-4 [DOI] [PubMed] [Google Scholar]
  3. Ahuja A. S. (2019). The impact of artificial intelligence in medicine on the future role of the physician. PeerJ 7:e7702. 10.7717/peerj.7702 [DOI] [PMC free article] [PubMed] [Google Scholar]
  4. Akram T., Attique M., Gul S., Shahzad A., Altaf M., Naqvi S. S. R., et al. (2021). A novel framework for rapid diagnosis of COVID-19 on computed tomography scans. Pattern Anal. Appl. 24 965–965. 10.1007/s10044-020-00950-0 [DOI] [PMC free article] [PubMed] [Google Scholar]
  5. American College of Radiology (ACR) (2021). Reporting and Data Systems. Available online at: https://www.acr.org/Clinical-Resources/Reporting-and-Data-Systems [Accessed July 22, 2021] [Google Scholar]
  6. Ariji Y., Fukuda M., Kise Y., Nozawa M., Yanashita Y., Fujita H., et al. (2019). Contrast-enhanced computed tomography image assessment of cervical lymph node metastasis in patients with oral cancer by using a deep learning system of artificial intelligence. Oral. Surg. Oral Med. Oral Pathol. Oral. Radiol. 127 458–463. 10.1016/j.oooo.2018.10.002 [DOI] [PubMed] [Google Scholar]
  7. Armitage A., Cordova A., Siegel R. (2017). Design-Thinking: The Answer To The Impasse Between Innovation And Regulation. UC Hastings Research Paper No. 250. Available online at: https://ssrn.com/abstract=3024176 or 10.2139/ssrn.3024176 (Accessed August 22, 2017). [DOI] [Google Scholar]
  8. Arnold M. H. (2021). Teasing out artificial intelligence in medicine: an ethical critique of artificial intelligence and machine learning in medicine. J. Bioeth. Inq. 18 121–139. 10.1007/s11673-020-10080-1 [DOI] [PMC free article] [PubMed] [Google Scholar]
  9. Belfiore M. P., Urraro F., Grassi R., Giacobbe G., Patelli G., Cappabianca S., et al. (2020). Artificial intelligence to codify lung CT in Covid-19 patients. Radiol. Med. 125 500–504. 10.1007/s11547-020-01195-x [DOI] [PMC free article] [PubMed] [Google Scholar]
  10. Beregi J.-P., Zins M., Masson J.-P., Cart P., Bartoli J.-M., Silberman B., et al. (2018). Radiology and artificial intelligence: an opportunity for our specialty. Diagn. Interv. Imaging 99 677–678. 10.1016/j.diii.2018.11.002 [DOI] [PubMed] [Google Scholar]
  11. Berlin L. (2007). Communicating results of all radiologic examinations directly to patients: has the time come? AJR Am. J. Roentgenol. 189 1275–1282. 10.2214/AJR.07.2740 [DOI] [PubMed] [Google Scholar]
  12. Bi W. L., Hosny A., Schabath M. B., Giger M. L., Birkbak N. J., Mehrtash A., et al. (2019). Artificial intelligence in cancer imaging: clinical challenges and applications. CA Cancer J. Clin. 69 127–157. 10.3322/caac.21552 [DOI] [PMC free article] [PubMed] [Google Scholar]
  13. Bilen A., Özer A. B. (2021). Cyber-attack method and perpetrator prediction using machine learning algorithms. PeerJ. Comput. Sci. 7:e475. 10.7717/peerj-cs.475 [DOI] [PMC free article] [PubMed] [Google Scholar]
  14. Bishop C. M. (2006). Pattern recognition And Machine Learning. Available online at: https://cds.cern.ch/record/998831 [Accessed July 22, 2021]. [Google Scholar]
  15. Bisschops R., East J. E., Hassan C., Hazewinkel Y., Kamiński M. F., Neumann H., et al. (2019). Advanced imaging for detection and differentiation of colorectal neoplasia: European Society of Gastrointestinal Endoscopy (ESGE) guideline – update 2019. Endoscopy 51 1155–1179. 10.1055/a-1031-7657 [DOI] [PubMed] [Google Scholar]
  16. Blackmore C. C., Medina L. S. (2006). Evidence-based radiology and the ACR appropriateness criteria. J. Am. Coll. Radiol. 3 505–509. 10.1016/j.jacr.2006.03.003 [DOI] [PubMed] [Google Scholar]
  17. Bosmans J. M. L., Neri E., Ratib O., Kahn C. E., Jr. (2015). Structured reporting: a fusion reactor hungry for fuel. Insights Imaging 6 129–132. 10.1007/s13244-014-0368-7 [DOI] [PMC free article] [PubMed] [Google Scholar]
  18. Brink J. A., Arenson R. L., Grist T. M., Lewin J. S., Enzmann D. (2017). Bits and bytes: the future of radiology lies in informatics and information technology. Eur. Radiol. 27 3647–3651. 10.1007/s00330-016-4688-5 [DOI] [PubMed] [Google Scholar]
  19. Brundage M., Avin S., Clark J., Toner H., Eckersley P., Garfinkel B., et al. (2018). The malicious use of artificial intelligence: forecasting, prevention, and mitigation. arXiv [CS.AI]. Available online at: http://arxiv.org/abs/1802.07228 (accessed July 22, 2021). [Google Scholar]
  20. Bugliesi M., Preneel B., Sassone V., Wegener I. (2006). “Automata, Languages and Programming,” in Proceedings of the 33rd International Colloquium, ICALP 2006, Venice, Italy, July 10-14, 2006 (Berlin: Springer; ). ISBN: 9783540359050 [Google Scholar]
  21. Butow P., Hoque E. (2020). Using artificial intelligence to analyse and teach communication in healthcare. Breast 50 49–55. 10.1016/j.breast.2020.01.008 [DOI] [PMC free article] [PubMed] [Google Scholar]
  22. Cabitza F., Rasoini R., Gensini G. F. (2017). Unintended consequences of machine learning in medicine. JAMA 318 517–518. 10.1001/jama.2017.7797 [DOI] [PubMed] [Google Scholar]
  23. Caliskan A., Bryson J. J., Narayanan A. (2017). Semantics derived automatically from language corpora contain human-like biases. Science 356 183–186. 10.1126/science.aal4230 [DOI] [PubMed] [Google Scholar]
  24. Campbell C. G., Ting D. S. W., Keane P. A., Foster P. J. (2020). The potential application of artificial intelligence for diagnosis and management of glaucoma in adults. Br. Med. Bull. 134 21–33. 10.1093/bmb/ldaa012 [DOI] [PubMed] [Google Scholar]
  25. Capaccio E., Podestà A., Morcaldi D., Sormani M. P., Derchi L. E. (2010). How often do patients ask for the results of their radiological studies? Insights Imaging 1 83–85. 10.1007/s13244-009-0003-1 [DOI] [PMC free article] [PubMed] [Google Scholar]
  26. Cappabianca S., Fusco R., De Lisio A., Paura C., Clemente A., Gagliardi G., et al. (2021). Clinical and laboratory data, radiological structured report findings and quantitative evaluation of lung involvement on baseline chest CT in COVID-19 patients to predict prognosis. Radiol. Med. 126 29–39. 10.1007/s11547-020-01293-w [DOI] [PMC free article] [PubMed] [Google Scholar]
  27. Catak F. O., Ahmed J., Sahinbas K., Khand Z. H. (2021). Data augmentation based malware detection using convolutional neural networks. PeerJ Comput. Sci. 7 e346. 10.7717/peerj-cs.346 [DOI] [PMC free article] [PubMed] [Google Scholar]
  28. Cath C., Wachter S., Mittelstadt B., Taddeo M., Floridi L. (2018). Artificial intelligence and the ‘Good Society’: the US, EU, and UK approach. Sci. Eng. Ethics 24 505–528. 10.1007/s11948-017-9901-7 [DOI] [PubMed] [Google Scholar]
  29. Cavoukian A. (2009). Privacy by design: the 7 foundational principles. Information and privacy commissioner of Ontario, Canada 5, 12. Available online at: http://dataprotection.industries/wp-content/uploads/2017/10/privacy-by-design.pdf [Accessed July 22, 2021]. [Google Scholar]
  30. Chartrand G., Cheng P. M., Vorontsov E., Drozdzal M., Turcotte S., Pal C. J., et al. (2017). Deep learning: a primer for radiologists. Radiographics 37 2113–2131. 10.1148/rg.2017170077 [DOI] [PubMed] [Google Scholar]
  31. Choy G., Khalilzadeh O., Michalski M., Do S., Samir A. E., Pianykh O. S., et al. (2018). Current applications and future impact of machine learning in radiology. Radiology 288 318–328. 10.1148/radiol.2018171820 [DOI] [PMC free article] [PubMed] [Google Scholar]
  32. Coppola F., Faggioni L., Regge D., Giovagnoni A., Golfieri R., Bibbolino C., et al. (2021). Artificial intelligence: radiologists’ expectations and opinions gleaned from a nationwide online survey. Radiol. Med. 126 63–71. 10.1007/s11547-020-01205-y [DOI] [PubMed] [Google Scholar]
  33. Cox J., Graham Y. (2020). Radiology and patient communication: if not now, then when? Eur. Radiol. 30 501–503. 10.1007/s00330-019-06349-8 [DOI] [PubMed] [Google Scholar]
  34. Curtis C., Liu C., Bollerman T. J., Pianykh O. S. (2018). Machine learning for predicting patient wait times and appointment delays. J. Am. Coll. Radiol. 15 1310–1316. 10.1016/j.jacr.2017.08.021 [DOI] [PubMed] [Google Scholar]
  35. Davenport T., Kalakota R. (2019). The potential for artificial intelligence in healthcare. Future Healthc. J. 6 94–98. 10.7861/futurehosp.6-2-94 [DOI] [PMC free article] [PubMed] [Google Scholar]
  36. DeCamp M., Lindvall C. (2020). Latent bias and the implementation of artificial intelligence in medicine. J Am. Med. Inform. Assoc. 27 2020–2023. 10.1093/jamia/ocaa094 [DOI] [PMC free article] [PubMed] [Google Scholar]
  37. Dedouit F., Saint-Martin P., Mokrane F.-Z., Savall F., Rousseau H., Crubézy E., et al. (2015). Virtual anthropology: useful radiological tools for age assessment in clinical forensic medicine and thanatology. Radiol. Med. 120 874–886. 10.1007/s11547-015-0525-1 [DOI] [PubMed] [Google Scholar]
  38. Deo A. (2021). Will AI deskill doctors?. Available online at: https://healthcare2020plus.com/?p=98 [Accessed July 22, 2021]. [Google Scholar]
  39. Di Ieva A. (2019). AI-augmented multidisciplinary teams: hype or hope? Lancet 394:1801. 10.1016/S0140-6736(19)32626-1 [DOI] [PubMed] [Google Scholar]
  40. Dirvanauskas D., Maskeliūnas R., Raudonis V., Damaševičius R., Scherer R. (2019). HEMIGEN: human embryo image generator based on generative adversarial networks. Sensors (Basel) 19:3578. 10.3390/s19163578 [DOI] [PMC free article] [PubMed] [Google Scholar]
  41. Dodd J. D. (2007). Evidence-based Practice in Radiology: Steps 3 and 4—Appraise and apply diagnostic radiology literature. Radiology 242 342–354. 10.1148/radiol.2422051679 [DOI] [PubMed] [Google Scholar]
  42. Domingos P. (2012). A few useful things to know about machine learning. Commun. ACM 55 78–87. 10.1145/2347736.2347755 [DOI] [Google Scholar]
  43. Duda R. O., Hart P. E., Stork D. G. (2001). Pattern classification. New Delhi: Wiley-Interscience. [Google Scholar]
  44. Dzobo K., Adotey S., Thomford N. E., Dzobo W. (2020). Integrating artificial and human intelligence: a partnership for responsible innovation in biomedical engineering and medicine. OMICS 24 247–263. 10.1089/omi.2019.0038 [DOI] [PubMed] [Google Scholar]
  45. European Society of Radiology (ESR) (2019a). Impact of artificial intelligence on radiology: a EuroAIM survey among members of the European Society of Radiology. Insights Imaging 10:105. 10.1186/s13244-019-0798-3 [DOI] [PMC free article] [PubMed] [Google Scholar]
  46. European Society of Radiology (ESR) (2019b). What the radiologist should know about artificial intelligence - an ESR white paper. Insights Imaging 10:44. 10.1186/s13244-019-0738-2 [DOI] [PMC free article] [PubMed] [Google Scholar]
  47. Faggioni L., Coppola F., Ferrari R., Neri E., Regge D. (2017). Usage of structured reporting in radiological practice: results from an Italian online survey. Eur. Radiol. 27 1934–1943. 10.1007/s00330-016-4553-6 [DOI] [PubMed] [Google Scholar]
  48. Finlayson S. G., Bowers J. D., Ito J., Zittrain J. L., Beam A. L., Kohane I. S. (2019). Adversarial attacks on medical machine learning. Science 363 1287–1289. 10.1126/science.aaw4399 [DOI] [PMC free article] [PubMed] [Google Scholar]
  49. Fukushima K. (1980). Neocognitron: A self-organizing neural network model for a mechanism of pattern recognition unaffected by shift in position. Biol. Cybernet. 36 193–202. 10.1007/bf00344251 [DOI] [PubMed] [Google Scholar]
  50. Gabelloni M., Faggioni L., Neri E. (2019). Imaging biomarkers in upper gastrointestinal cancers. BJR Open 1:20190001. 10.1259/bjro.20190001 [DOI] [PMC free article] [PubMed] [Google Scholar]
  51. Geis J. R., Brady A. P., Wu C. C., Spencer J., Ranschaert E., Jaremko J. L., et al. (2019). Ethics of artificial intelligence in radiology: summary of the joint European and North American multisociety statement. J. Am. Coll. Radiol. 16 1516–1521. 10.1016/j.jacr.2019.07.028 [DOI] [PubMed] [Google Scholar]
  52. Gencoglu O., van Gils M., Guldogan E., Morikawa C., Süzen M., Gruber M., et al. (2019). HARK Side Of Deep Learning – From Grad Student Descent To Automated Machine Learning. Available online at: https://arxiv.org/abs/1904.07633v1 (Accessed July 22, 2021). [Google Scholar]
  53. Geras K. J., Mann R. M., Moy L. (2019). Artificial intelligence for mammography and digital breast tomosynthesis: current concepts and future perspectives. Radiology 293 246–259. 10.1148/radiol.2019182627 [DOI] [PMC free article] [PubMed] [Google Scholar]
  54. Gillies R. J., Kinahan P. E., Hricak H. (2016). Radiomics: images are more than pictures, they are data. Radiology 278 563–577. 10.1148/radiol.2015151169 [DOI] [PMC free article] [PubMed] [Google Scholar]
  55. Glazer G. M., Ruiz-Wibbelsmann J. A. (2011). The invisible radiologist. Radiology 258 18–22. 10.1148/radiol.10101447 [DOI] [PubMed] [Google Scholar]
  56. Gong B., Nugent J. P., Guest W., Parker W., Chang P. J., Khosa F., et al. (2019). Influence of artificial intelligence on Canadian medical students’ preference for radiology specialty: a national survey study. Acad. Radiol. 26 566–577. 10.1016/j.acra.2018.10.007 [DOI] [PubMed] [Google Scholar]
  57. Grassi R., Miele V., Giovagnoni A. (2019). Artificial intelligence: a challenge for third millennium radiologist. Radiol. Med. 124 241–242. 10.1007/s11547-019-00990-5 [DOI] [PubMed] [Google Scholar]
  58. Gurgitano M., Angileri S. A., Rodà G. M., Liguori A., Pandolfi M., Ierardi A. M., et al. (2021). Interventional radiology ex-machina: impact of artificial intelligence on practice. Radiol. Med. 2021 1–9. 10.1007/s11547-021-01351-x [DOI] [PMC free article] [PubMed] [Google Scholar]
  59. Harada Y., Katsukura S., Kawamura R., Shimizu T. (2021). effects of a differential diagnosis list of artificial intelligence on differential diagnoses by physicians: an exploratory analysis of data from a randomized controlled study. Int. J. Environ. Res. Public Health 18:5562. 10.3390/ijerph18115562 [DOI] [PMC free article] [PubMed] [Google Scholar]
  60. Harmon S. A., Sanford T. H., Xu S., Turkbey E. B., Roth H., Xu Z., et al. (2020). Artificial intelligence for the detection of COVID-19 pneumonia on chest CT using multinational datasets. Nat. Commun. 11:4080. 10.1038/s41467-020-17971-2 [DOI] [PMC free article] [PubMed] [Google Scholar]
  61. Havaei M., Davy A., Warde-Farley D., Biard A., Courville A., Bengio Y., et al. (2017). Brain tumor segmentation with Deep Neural Networks. Med. Image Anal. 35 18–31. 10.1016/j.media.2016.05.004 [DOI] [PubMed] [Google Scholar]
  62. Hickman S. E., Baxter G. C., Gilbert F. J. (2021). Adoption of artificial intelligence in breast imaging: evaluation, ethical constraints and limitations. Br. J. Cancer 125 15–22. 10.1038/s41416-021-01333-w [DOI] [PMC free article] [PubMed] [Google Scholar]
  63. Hinton G. E., Salakhutdinov R. R. (2006). Reducing the dimensionality of data with neural networks. Science 313 504–507. 10.1126/science.1127647 [DOI] [PubMed] [Google Scholar]
  64. Holzinger A. (2021). Explainable AI and multi-modal causability in medicine. I-Com 19 171–179. 10.1515/icom-2020-0024 [DOI] [PMC free article] [PubMed] [Google Scholar]
  65. Honavar S. G. (2018). Patient–physician relationship – communication is the key. Indian J. Ophthalmol. 66 1527–1528. 10.4103/ijo.IJO_1760_18 [DOI] [PMC free article] [PubMed] [Google Scholar]
  66. Hosny A., Parmar C., Quackenbush J., Schwartz L. H., Aerts H. J. W. L. (2018). Artificial intelligence in radiology. Nat. Rev. Cancer 18 500–510. 10.1038/s41568-018-0016-5 [DOI] [PMC free article] [PubMed] [Google Scholar]
  67. Hu H. T., Shan Q. Y., Chen S. L., Li B., Feng S. T., Xu E. J., et al. (2020). CT-based radiomics for preoperative prediction of early recurrent hepatocellular carcinoma: technical reproducibility of acquisition and scanners. Radiol. Med. 125 697–705. 10.1007/s11547-020-01174-2 [DOI] [PubMed] [Google Scholar]
  68. Huxley C. J., Atherton H., Watkins J. A., Griffiths F. (2015). Digital communication between clinician and patient and the impact on marginalised groups: a realist review in general practice. Br. J. Gen. Pract. 65 e813–e821. 10.3399/bjgp15X687853 [DOI] [PMC free article] [PubMed] [Google Scholar]
  69. Ierardi A. M., Fontana F., Giorlando F., De Marchi G., Pinto A., Radaelli A., et al. (2016). Evaluation of tablet ultrasound for routine abdominal interventional procedures. Radiol. Med. 121 675–680. 10.1007/s11547-016-0641-6 [DOI] [PubMed] [Google Scholar]
  70. Ishii E., Ebner D. K., Kimura S., Agha-Mir-Salim L., Uchimido R., Celi L. A. (2020). The advent of medical artificial intelligence: lessons from the Japanese approach. J. Intensive Care Med. 8:35. 10.1186/s40560-020-00452-5 [DOI] [PMC free article] [PubMed] [Google Scholar]
  71. Jha S., Topol E. J. (2016). Adapting to artificial intelligence: radiologists and pathologists as information specialists. JAMA 316 2353–2354. 10.1001/jama.2016.17438 [DOI] [PubMed] [Google Scholar]
  72. Jiang Y., Liang X., Wang W., Chen C., Yuan Q., Zhang X., et al. (2021). Noninvasive prediction of occult peritoneal metastasis in gastric cancer using deep learning. JAMA Netw. Open 4:e2032269. 10.1001/jamanetworkopen.2020.32269 [DOI] [PMC free article] [PubMed] [Google Scholar]
  73. Jones H. (2018). Geoff hinton dismissed the need for explainable AI: 8 experts explain why he’s wrong. Forbes Magazine. Available online at: https://www.forbes.com/sites/cognitiveworld/2018/12/20/geoff-hinton-dismissed-the-need-for-explainable-ai-8-experts-explain-why-hes-wrong/ [Accessed July 22, 2021]. [Google Scholar]
  74. Juravle G., Boudouraki A., Terziyska M., Rezlescu C. (2020). Trust in artificial intelligence for medical diagnoses. Prog. Brain. Res. 253 263–282. 10.1016/bs.pbr.2020.06.006 [DOI] [PubMed] [Google Scholar]
  75. Kallenberg M., Petersen K., Nielsen M., Ng A. Y., Diao P., Igel C., et al. (2016). Unsupervised deep learning applied to breast density segmentation and mammographic risk scoring. IEEE Trans. Med. Imaging 35 1322–1331. 10.1109/TMI.2016.2532122 [DOI] [PubMed] [Google Scholar]
  76. Kelly C. J., Karthikesalingam A., Suleyman M., Corrado G., King D. (2019). Key challenges for delivering clinical impact with artificial intelligence. BMC Med. 17:195. 10.1186/s12916-019-1426-2 [DOI] [PMC free article] [PubMed] [Google Scholar]
  77. Kermany D. S., Goldbaum M., Cai W., Valentim C. C. S., Liang H., Baxter S. L., et al. (2018). Identifying medical diagnoses and treatable diseases by image-based deep learning. Cell 172 1122–1131.e9. 10.1016/j.cell.2018.02.010 [DOI] [PubMed] [Google Scholar]
  78. Keskinbora K. H. (2019). Medical ethics considerations on artificial intelligence. J. Clin. Neurosci. 64 277–282. 10.1016/j.jocn.2019.03.001 [DOI] [PubMed] [Google Scholar]
  79. Kiener M. (2020). Artificial intelligence in medicine and the disclosure of risks. AI Soc. 10.1007/s00146-020-01085-w [DOI] [PMC free article] [PubMed] [Google Scholar]
  80. King A. D., Chow K.-K., Yu K.-H., Mo F. K. F., Yeung D. K. W., Yuan J., et al. (2013). Head and neck squamous cell carcinoma: diagnostic performance of diffusion-weighted MR imaging for the prediction of treatment response. Radiology 266 531–538. 10.1148/radiol.12120167 [DOI] [PubMed] [Google Scholar]
  81. Kobayashi Y., Ishibashi M., Kobayashi H. (2019). How will “democratization of artificial intelligence” change the future of radiologists? Jpn. J. Radiol. 37 9–14. 10.1007/s11604-018-0793-5 [DOI] [PubMed] [Google Scholar]
  82. Kohli M., Geis R. (2018). Ethics, artificial intelligence, and radiology. J. Am. Coll. Radiol. 15 1317–1319. 10.1016/j.jacr.2018.05.020 [DOI] [PubMed] [Google Scholar]
  83. Kooi T., Litjens G., van Ginneken B., Gubern-Mérida A., Sánchez C. I., Mann R., et al. (2017). Large scale deep learning for computer aided detection of mammographic lesions. Med. Image Anal. 35 303–312. 10.1016/j.media.2016.07.007 [DOI] [PubMed] [Google Scholar]
  84. Krittanawong C. (2018). The rise of artificial intelligence and the uncertain future for physicians. Eur. J. Intern. Med. 48 e13–e14. 10.1016/j.ejim.2017.06.017 [DOI] [PubMed] [Google Scholar]
  85. Kulikowski C. A. (2019). Beginnings of artificial intelligence in medicine (AIM): computational artifice assisting scientific inquiry and clinical art - with reflections on present AIM challenges. Yearb. Med. Inform. 28 249–256. 10.1055/s-0039-1677895 [DOI] [PMC free article] [PubMed] [Google Scholar]
  86. Kulkarni A. (2021). AI In Healthcare: Data Privacy And Ethics Concerns. Available online at: https://www.lexalytics.com/lexablog/ai-healthcare-data-privacy-ethics-issues [Accessed July 22, 2021]. [Google Scholar]
  87. Lakhani P., Sundaram B. (2017). Deep learning at chest radiography: automated classification of pulmonary tuberculosis by using convolutional neural networks. Radiology 284 574–582. 10.1148/radiol.2017162326 [DOI] [PubMed] [Google Scholar]
  88. Lambin P., Leijenaar R. T. H., Deist T. M., Peerlings J., de Jong E. E. C., van Timmeren J., et al. (2017). Radiomics: the bridge between medical imaging and personalized medicine. Nat. Rev. Clin. Oncol. 14 749–762. 10.1038/nrclinonc.2017.141 [DOI] [PubMed] [Google Scholar]
  89. LeCun Y., Bengio Y., Hinton G. (2015). Deep learning. Nature 521 436–444. 10.1038/nature14539 [DOI] [PubMed] [Google Scholar]
  90. LeCun Y., Bottou L., Bengio Y., Haffner P. (1998). Gradient-based learning applied to document recognition. Proc. IEEE 86 2278–2324. 10.1109/5.726791 [DOI] [Google Scholar]
  91. Lee K., Lee S. H. (2020). Artificial intelligence-driven oncology clinical decision support system for multidisciplinary teams. Sensors (Basel) 20:4693. 10.3390/s20174693 [DOI] [PMC free article] [PubMed] [Google Scholar]
  92. Lerouge J., Herault R., Chatelain C., Jardin F., Modzelewski R. (2015). IODA: an input/output deep architecture for image labeling. Pattern Recognit. 48 2847–2858. 10.1016/j.patcog.2015.03.017 [DOI] [Google Scholar]
  93. Levinson W., Lesser C. S., Epstein R. M. (2010). Developing physician communication skills for patient-centered care. Health Aff. (Millwood) 29 1310–1318. 10.1377/hlthaff.2009.0450 [DOI] [PubMed] [Google Scholar]
  94. Litjens G., Kooi T., Bejnordi B. E., Setio A. A. A., Ciompi F., Ghafoorian M., et al. (2017). A survey on deep learning in medical image analysis. Med. Image Anal. 42 60–88. 10.1016/j.media.2017.07.005 [DOI] [PubMed] [Google Scholar]
  95. Liu Z., Wang S., Dong D., Wei J., Fang C., Zhou X., et al. (2019). The applications of radiomics in precision diagnosis and treatment of oncology: opportunities and challenges. Theranostics 9 1303–1322. 10.7150/thno.30309 [DOI] [PMC free article] [PubMed] [Google Scholar]
  96. Lo Gullo R., Daimiel I., Morris E. A., Pinker K. (2020). Combining molecular and imaging metrics in cancer: radiogenomics. Insights Imaging 11 1. 10.1186/s13244-019-0795-6 [DOI] [PMC free article] [PubMed] [Google Scholar]
  97. Lu J. (2016). Will medical technology deskill doctors? Int. Educ. Stud. 9 130–134. 10.5539/ies.v9n7p130 [DOI] [Google Scholar]
  98. Lustberg T., van Soest J., Gooding M., Peressutti D., Aljabar P., van der Stoep J., et al. (2018). Clinical evaluation of atlas and deep learning based automatic contouring for lung cancer. Radiother. Oncol. 126 312–317. 10.1016/j.radonc.2017.11.012 [DOI] [PubMed] [Google Scholar]
  99. Marcovici P. A., Taylor G. A. (2014). Journal Club: structured radiology reports are more complete and more effective than unstructured reports. AJR Am. J. Roentgenol. 203 1265–1271. 10.2214/AJR.14.12636 [DOI] [PubMed] [Google Scholar]
  100. Martinez-Martin N., Dunn L. B., Roberts L. W. (2018). Is it ethical to use prognostic estimates from machine learning to treat psychosis? AMA J. Ethics 20 E804–E811. 10.1001/amajethics.2018.804 [DOI] [PMC free article] [PubMed] [Google Scholar]
  101. Martín-Noguerol T., Paulano-Godino F., López-Ortega R., Górriz J. M., Riascos R. F., Luna A. (2021). Artificial intelligence in radiology: relevance of collaborative work between radiologists and engineers for building a multidisciplinary team. Clin. Radiol. 76 317–324. 10.1016/j.crad.2020.11.113 [DOI] [PubMed] [Google Scholar]
  102. Mendelson E. B. (2019). Artificial intelligence in breast imaging: potentials and limitations. AJR Am. J. Roentgenol. 212 293–299. 10.2214/AJR.18.20532 [DOI] [PubMed] [Google Scholar]
  103. Meskó B., Drobni Z., Bényei É, Gergely B., Győrffy Z. (2017). Digital health is a cultural transformation of traditional healthcare. mHealth 3:38. 10.21037/mhealth.2017.08.07 [DOI] [PMC free article] [PubMed] [Google Scholar]
  104. Meskó B., Hetényi G., Győrffy Z. (2018). Will artificial intelligence solve the human resource crisis in healthcare? BMC Health Serv. Res. 18:545. 10.1186/s12913-018-3359-4 [DOI] [PMC free article] [PubMed] [Google Scholar]
  105. Mikhaylov S. J., Esteve M., Campion A. (2018). Artificial intelligence for the public sector: opportunities and challenges of cross-sector collaboration. Philos. Trans. A Math. Phys. Eng. Sci. 376:20170357. 10.1098/rsta.2017.0357 [DOI] [PMC free article] [PubMed] [Google Scholar]
  106. Miller D. D., Brown E. W. (2019). How cognitive machines can augment medical imaging. AJR Am. J. Roentgenol. 212 9–14. 10.2214/AJR.18.19914 [DOI] [PubMed] [Google Scholar]
  107. Mittelstadt B. D., Floridi L. (2016). The ethics of big data: current and foreseeable issues in biomedical contexts. Sci. Eng. Ethics 22 303–341. 10.1007/s11948-015-9652-2 [DOI] [PubMed] [Google Scholar]
  108. Monreale A. (2020). Rischi Etico-Legali Dell’intelligenza Artificiale. DPCE Online 44. Available online at: http://www.dpceonline.it/index.php/dpceonline/article/view/1083 [Accessed July 22, 2021]. [Google Scholar]
  109. Monreale A., Rinzivillo S., Pratesi F., Giannotti F., Pedreschi D. (2014). Privacy-by-design in big data analytics and social mining. EPJ Data Sci. 3 10. 10.1140/epjds/s13688-014-0010-4 [DOI] [Google Scholar]
  110. Murata K., Endo K., Aihara T., Suzuki H., Sawaji Y., Matsuoka Y., et al. (2020). Artificial intelligence for the detection of vertebral fractures on plain spinal radiography. Sci. Rep. 10:20031. 10.1038/s41598-020-76866-w [DOI] [PMC free article] [PubMed] [Google Scholar]
  111. Myers T. G., Ramkumar P. N., Ricciardi B. F., Urish K. L., Kipper J., Ketonis C. (2020). Artificial intelligence and orthopaedics: an introduction for clinicians. J. Bone Joint. Surg. Am. 102 830–840. 10.2106/JBJS.19.01128 [DOI] [PMC free article] [PubMed] [Google Scholar]
  112. Na L., Yang C., Lo C. C., Zhao F., Fukuoka Y., Aswani A. (2018). Feasibility of reidentifying individuals in large national physical activity data sets from which protected health information has been removed with use of machine learning. JAMA Netw. Open 1:e186040. 10.1001/jamanetworkopen.2018.6040 [DOI] [PMC free article] [PubMed] [Google Scholar]
  113. Nagendran M., Chen Y., Lovejoy C. A., Gordon A. C., Komorowski M., Harvey H., et al. (2020). Artificial intelligence versus clinicians: systematic review of design, reporting standards, and claims of deep learning studies. BMJ 368:m689. 10.1136/bmj.m689 [DOI] [PMC free article] [PubMed] [Google Scholar]
  114. Nance J. W., Jr., Meenan C., Nagy P. G. (2013). The future of the radiology information system. AJR Am. J. Roentgenol. 200 1064–1070. 10.2214/AJR.12.10326 [DOI] [PubMed] [Google Scholar]
  115. Napel S., Mu W., Jardim-Perassi B. V., Aerts H. J. W. L., Gillies R. J. (2018). Quantitative imaging of cancer in the postgenomic era: radio(geno)mics, deep learning, and habitats. Cancer 124 4633–4649. 10.1002/cncr.31630 [DOI] [PMC free article] [PubMed] [Google Scholar]
  116. Nazari M., Shiri I., Hajianfar G., Oveisi N., Abdollahi H., Deevband M. R., et al. (2020). Noninvasive Fuhrman grading of clear cell renal cell carcinoma using computed tomography radiomic features and machine learning. Radiol Med. 125 754–762. 10.1007/s11547-020-01169-z [DOI] [PubMed] [Google Scholar]
  117. Nelson C. A., Pérez-Chada L. M., Creadore A., Li S. J., Lo K., Manjaly P., et al. (2020). Patient perspectives on the use of artificial intelligence for skin cancer screening: a qualitative study. JAMA Dermatol. 156 501–512. 10.1001/jamadermatol.2019.5014 [DOI] [PMC free article] [PubMed] [Google Scholar]
  118. Neri E., Coppola F., Miele V., Bibbolino C., Grassi R. (2020). Artificial intelligence: who is responsible for the diagnosis? Radiol. Med. 125 517–521. 10.1007/s11547-020-01135-9 [DOI] [PubMed] [Google Scholar]
  119. Ngo T. A., Lu Z., Carneiro G. (2017). Combining deep learning and level set for the automated segmentation of the left ventricle of the heart from cardiac cine magnetic resonance. Med. Image Anal. 35 159–171. 10.1016/j.media.2016.05.009 [DOI] [PubMed] [Google Scholar]
  120. Nguyen G. K., Shetty A. S. (2018). Artificial intelligence and machine learning: opportunities for radiologists in training. J. Am. Coll. Radiol. 15 1320–1321. 10.1016/j.jacr.2018.05.024 [DOI] [PubMed] [Google Scholar]
  121. Oh C., Lee T., Kim Y., Park S., Kwon S., Suh B. (2017). “Us vs. them: understanding artificial intelligence technophobia over the Google DeepMind challenge match,” in Proceedings of the 2017 CHI Conference on Human Factors in Computing Systems CHI ’17 (New York, NY: Association for Computing Machinery; ), 2523–2534. 10.1145/3025453.3025539 [DOI] [Google Scholar]
  122. Oksanen A., Savela N., Latikka R., Koivula A. (2020). Trust toward robots and artificial intelligence: an experimental approach to human-technology interactions online. Front. Psychol. 11:568256. 10.3389/fpsyg.2020.568256 [DOI] [PMC free article] [PubMed] [Google Scholar]
  123. Ooi S. K. G., Makmur A., Soon A. Y. Q., Fook-Chong S., Liew C., Sia S. Y., et al. (2021). Attitudes toward artificial intelligence in radiology with learner needs assessment within radiology residency programmes: a national multi-programme survey. Singapore Med. J. 62 126–134. 10.11622/smedj.2019141 [DOI] [PMC free article] [PubMed] [Google Scholar]
  124. Panesar S. S., Kliot M., Parrish R., Fernandez-Miranda J., Cagle Y., Britz G. W. (2020). Promises and perils of artificial intelligence in neurosurgery. Neurosurgery 87 33–44. 10.1093/neuros/nyz471 [DOI] [PubMed] [Google Scholar]
  125. Park J. E., Kim D., Kim H. S., Park S. Y., Kim J. Y., Cho S. J., et al. (2020). Quality of science and reporting of radiomics in oncologic studies: room for improvement according to radiomics quality score and TRIPOD statement. Eur. Radiol. 30 523–536. 10.1007/s00330-019-06360-z [DOI] [PubMed] [Google Scholar]
  126. Park W. J., Park J.-B. (2018). History and application of artificial neural networks in dentistry. Eur. J. Dent. 12 594–601. 10.4103/ejd.ejd_325_18 [DOI] [PMC free article] [PubMed] [Google Scholar]
  127. Pesapane F., Codari M., Sardanelli F. (2018a). Artificial intelligence in medical imaging: threat or opportunity? Radiologists again at the forefront of innovation in medicine. Eur. Radiol. Exp. 2:35. 10.1186/s41747-018-0061-6 [DOI] [PMC free article] [PubMed] [Google Scholar]
  128. Pesapane F., Volonté C., Codari M., Sardanelli F. (2018b). Artificial intelligence as a medical device in radiology: ethical and regulatory issues in Europe and the United States. Insights Imaging 9 745–753. 10.1007/s13244-018-0645-y [DOI] [PMC free article] [PubMed] [Google Scholar]
  129. Philbrick K. A., Yoshida K., Inoue D., Akkus Z., Kline T. L., Weston A. D., et al. (2018). What does deep learning see? Insights from a classifier trained to predict contrast enhancement phase from CT images. AJR Am. J. Roentgenol. 211 1184–1193. 10.2214/AJR.18.20331 [DOI] [PubMed] [Google Scholar]
  130. Pinker K., Shitano F., Sala E., Do R. K., Young R. J., Wibmer A. G., et al. (2018). Background, current role, and potential applications of radiogenomics. J. Magn. Reson. Imaging 47 604–620. 10.1002/jmri.25870 [DOI] [PMC free article] [PubMed] [Google Scholar]
  131. Pinto Dos Santos D., Baeßler B. (2018). Big data, artificial intelligence, and structured reporting. Eur. Radiol. Exp. 2:42. 10.1186/s41747-018-0071-4 [DOI] [PMC free article] [PubMed] [Google Scholar]
  132. Porsdam Mann S., Savulescu J., Sahakian B. J. (2016). Facilitating the ethical use of health data for the benefit of society: electronic health records, consent and the duty of easy rescue. Philos. Trans. A. Math. Phys. Eng. Sci. 374:20160130. 10.1098/rsta.2016.0130 [DOI] [PMC free article] [PubMed] [Google Scholar]
  133. Pravettoni G., Triberti S. (2019). Il Medico 4.0: Come Cambia La Relazione Medico-Paziente Nell’era Delle Nuove Tecnologie. Palm Beach Gardens, FL: EDRA. [Google Scholar]
  134. Price W. N., II, Gerke S., Cohen I. G. (2019). Potential liability for physicians using artificial intelligence. JAMA 322 1765–1766. 10.1001/jama.2019.15064 [DOI] [PubMed] [Google Scholar]
  135. Price W. N., II, Gerke S., Cohen I. G. (2021). How much can potential jurors tell us about liability for medical artificial intelligence? J. Nucl. Med. 62 15–16. 10.2967/jnumed.120.257196 [DOI] [PubMed] [Google Scholar]
  136. Rajpurkar P., Irvin J., Ball R. L., Zhu K., Yang B., Mehta H., et al. (2018). Deep learning for chest radiograph diagnosis: a retrospective comparison of the CheXNeXt algorithm to practicing radiologists. PLoS Med. 15:e1002686. 10.1371/journal.pmed.1002686 [DOI] [PMC free article] [PubMed] [Google Scholar]
  137. Recht M., Bryan R. N. (2017). Artificial intelligence: threat or boon to radiologists? J. Am. Coll. Radiol. 14 1476–1480. 10.1016/j.jacr.2017.07.007 [DOI] [PubMed] [Google Scholar]
  138. Reddy S., Fox J., Purohit M. P. (2019). Artificial intelligence-enabled healthcare delivery. J. R. Soc. Med. 112 22–28. 10.1177/0141076818815510 [DOI] [PMC free article] [PubMed] [Google Scholar]
  139. Rizzo S., Botta F., Raimondi S., Origgi D., Fanciullo C., Morganti A. G., et al. (2018). Radiomics: the facts and the challenges of image analysis. Eur. Radiol. Exp. 2 36. 10.1186/s41747-018-0068-z [DOI] [PMC free article] [PubMed] [Google Scholar]
  140. Rogers W., Thulasi Seetha S., Refaee T. A. G., Lieverse R. I. Y., Granzier R. W. Y., Ibrahim A., et al. (2020). Radiomics: from qualitative to quantitative imaging. Br. J. Radiol. 93 20190948. 10.1259/bjr.20190948 [DOI] [PMC free article] [PubMed] [Google Scholar]
  141. Ross P., Spates K. (2020). Considering the safety and quality of artificial intelligence in health care. Jt. Comm. J. Qual. Patient Saf. 46 596–599. 10.1016/j.jcjq.2020.08.002 [DOI] [PMC free article] [PubMed] [Google Scholar]
  142. Rubin D. L. (2019). Artificial intelligence in imaging: the radiologist’s role. J. Am. Coll. Radiol. 16 1309–1317. 10.1016/j.jacr.2019.05.036 [DOI] [PMC free article] [PubMed] [Google Scholar]
  143. Russell S., Bohannon J. (2015). Artificial intelligence. Fears of an AI pioneer. Science 349:252. 10.1126/science.349.6245.252 [DOI] [PubMed] [Google Scholar]
  144. Sardanelli F., Hunink M. G., Gilbert F. J., Di Leo G., Krestin G. P. (2010). Evidence-based radiology: why and how? Eur. Radiol. 20 1–15. 10.1007/s00330-009-1574-4 [DOI] [PubMed] [Google Scholar]
  145. Savadjiev P., Chong J., Dohan A., Vakalopoulou M., Reinhold C., Paragios N., et al. (2019). Demystification of AI-driven medical image interpretation: past, present and future. Eur. Radiol. 29 1616–1624. 10.1007/s00330-018-5674-x [DOI] [PubMed] [Google Scholar]
  146. Shan H., Padole A., Homayounieh F., Kruger U., Khera R. D., Nitiwarangkul C., et al. (2019). Competitive performance of a modularized deep neural network compared to commercial algorithms for low-dose CT image reconstruction. Nat. Mach. Intell. 1 269–276. 10.1038/s42256-019-0057-9 [DOI] [PMC free article] [PubMed] [Google Scholar]
  147. Shokri R., Stronati M., Song C., Shmatikov V. (2017). “Membership inference attacks against machine learning models,” in Proceedings of the 2017 IEEE Symposium on Security and Privacy (SP), San Jose, CA, 3–18. 10.1109/SP.2017.41 [DOI] [Google Scholar]
  148. Sniecinski I., Seghatchian J. (2018). Artificial intelligence: a joint narrative on potential use in pediatric stem and immune cell therapies and regenerative medicine. Transfus. Apher. Sci. 57 422–424. 10.1016/j.transci.2018.05.004 [DOI] [PubMed] [Google Scholar]
  149. Sogani J., Allen B., Jr., Dreyer K., McGinty G. (2020). Artificial intelligence in radiology: the ecosystem essential to improving patient care. Clin. Imaging 59 A3–A6. 10.1016/j.clinimag.2019.08.001 [DOI] [PubMed] [Google Scholar]
  150. Srinuan C., Bohlin E. (2011). Understanding The Digital Divide: A Literature Survey And Ways Forward. Available online at: https://www.econstor.eu/handle/10419/52191 [Accessed July 22, 2021]. [Google Scholar]
  151. Story M. D., Durante M. (2018). Radiogenomics. Med. Phys. 45 e1111–e1122. 10.1002/mp.13064 [DOI] [PubMed] [Google Scholar]
  152. Stoyanova R., Pollack A., Takhar M., Lynne C., Parra N., Lam L. L. C., et al. (2016). Association of multiparametric MRI quantitative imaging features with prostate cancer gene expression in MRI-targeted prostate biopsies. Oncotarget 7 53362–53376. 10.18632/oncotarget.10523 [DOI] [PMC free article] [PubMed] [Google Scholar]
  153. Sweeney L. (2002). k-anonymity: a model for protecting privacy. Int. J. Uncertain. Fuzziness Knowledge Based Syst. 10 557–570. 10.1142/S0218488502001648 [DOI] [Google Scholar]
  154. Tajmir S. H., Alkasab T. K. (2018). Toward augmented radiologists: changes in radiology education in the era of machine learning and artificial intelligence. Acad. Radiol. 25 747–750. 10.1016/j.acra.2018.03.007 [DOI] [PubMed] [Google Scholar]
  155. Tang X. (2020). The role of artificial intelligence in medical imaging research. BJR Open 2:20190031. 10.1259/bjro.20190031 [DOI] [PMC free article] [PubMed] [Google Scholar]
  156. Tobia K., Nielsen A., Stremitzer A. (2021). When does physician use of AI increase liability? J. Nucl. Med. 62 17–21. 10.2967/jnumed.120.256032 [DOI] [PMC free article] [PubMed] [Google Scholar]
  157. Tohka J., van Gils M. (2021). Evaluation of machine learning algorithms for health and wellness applications: a tutorial. Comput. Biol Med. 132:104324. 10.1016/j.compbiomed.2021.104324 [DOI] [PubMed] [Google Scholar]
  158. Trebeschi S., van Griethuysen J. J. M., Lambregts D. M. J., Lahaye M. J., Parmar C., Bakers F. C. H., et al. (2017). Deep learning for fully-automated localization and segmentation of rectal cancer on multiparametric MR. Sci. Rep. 7:5301. 10.1038/s41598-017-05728-9 [DOI] [PMC free article] [PubMed] [Google Scholar]
  159. Triberti S., Durosini I., Pravettoni G. (2020). A “third wheel” effect in health decision making involving artificial entities: a psychological perspective. Front. Public Health 8:117. 10.3389/fpubh.2020.00117 [DOI] [PMC free article] [PubMed] [Google Scholar]
  160. Trimboli R. M., Codari M., Bert A., Carbonaro L. A., Maccagnoni S., Raciti D., et al. (2018). Breast arterial calcifications on mammography: intra- and inter-observer reproducibility of a semi-automatic quantification tool. Radiol. Med. 123 168–173. 10.1007/s11547-017-0827-6 [DOI] [PubMed] [Google Scholar]
  161. Valiuškaitė V., Raudonis V., Maskeliūnas R., Damaševičius R., Krilavičius T. (2020). Deep learning based evaluation of spermatozoid motility for artificial Insemination. Sensors (Basel) 21:72. 10.3390/s21010072 [DOI] [PMC free article] [PubMed] [Google Scholar]
  162. van Assen M., Muscogiuri G., Caruso D., Lee S. J., Laghi A., De Cecco C. N. (2020). Artificial intelligence in cardiac radiology. Radiol. Med. 125 1186–1199. 10.1007/s11547-020-01277-w [DOI] [PubMed] [Google Scholar]
  163. van Hoek J., Huber A., Leichtle A., Härmä K., Hilt D., von Tengg-Kobligk H., et al. (2019). A survey on the future of radiology among radiologists, medical students and surgeons: students and surgeons tend to be more skeptical about artificial intelligence and radiologists may fear that other disciplines take over. Eur. J. Radiol. 121:108742. 10.1016/j.ejrad.2019.108742 [DOI] [PubMed] [Google Scholar]
  164. Villanueva-Meyer J. E., Chang P., Lupo J. M., Hess C. P., Flanders A. E., Kohli M. (2019). Machine learning in neurooncology imaging: from study request to diagnosis and treatment. AJR Am. J. Roentgenol. 212 52–56. 10.2214/AJR.18.20328 [DOI] [PubMed] [Google Scholar]
  165. Voigt P., Von dem Bussche A. (2017). The EU General Data Protection Regulation (GDPR): A Practical Guide, 1st Edn. Cham: Springer. [Google Scholar]
  166. Vought R. T. (2020). Re: Guidance For Regulation Of Artificial Intelligence Applications. Available online at: https://www.acr.org/-/media/ACR/Files/Advocacy/Regulatory-Issues/acr_comments_draft-OMB-memo_3-12-2020.pdf [Accessed July 22, 2021]. [Google Scholar]
  167. Wang S., Liu Z., Rong Y., Zhou B., Bai Y., Wei W., et al. (2019). Deep learning provides a new computed tomography-based prognostic biomarker for recurrence prediction in high-grade serous ovarian cancer. Radiother. Oncol. 132 171–177. 10.1016/j.radonc.2018.10.019 [DOI] [PubMed] [Google Scholar]
  168. Wang S., Zhou M., Liu Z., Liu Z., Gu D., Zang Y., et al. (2017). Central focused convolutional neural networks: developing a data-driven model for lung nodule segmentation. Med. Image Anal. 40 172–183. 10.1016/j.media.2017.06.014 [DOI] [PMC free article] [PubMed] [Google Scholar]
  169. Waymel Q., Badr S., Demondion X., Cotten A., Jacques T. (2019). Impact of the rise of artificial intelligence in radiology: What do radiologists think? Diagn. Interv. Imaging 100 327–336. 10.1016/j.diii.2019.03.015 [DOI] [PubMed] [Google Scholar]
  170. Yu K.-H., Beam A. L., Kohane I. S. (2018). Artificial intelligence in healthcare. Nat. Biomed. Eng. 2 719–731. 10.1038/s41551-018-0305-z [DOI] [PubMed] [Google Scholar]
  171. Zackova E. (2015). “Intelligence explosion quest for humankind,” in Beyond artificial intelligence: the disappearing human-machine divide, eds Romportl J., Zackova E., Kelemen J. (Cham: Springer International Publishing; ), 31–43. 10.1007/978-3-319-09668-1_3 [DOI] [Google Scholar]
  172. Zanotel M., Bednarova I., Londero V., Linda A., Lorenzon M., Girometti R., et al. (2018). Automated breast ultrasound: basic principles and emerging clinical applications. Radiol. Med. 123 1–12. 10.1007/s11547-017-0805-z [DOI] [PubMed] [Google Scholar]
  173. Zerunian M., Caruso D., Zucchelli A., Polici M., Capalbo C., Filetti M., et al. (2021). CT based radiomic approach on first line pembrolizumab in lung cancer. Sci. Rep. 11:6633. 10.1038/s41598-021-86113-5 [DOI] [PMC free article] [PubMed] [Google Scholar]
  174. Zhou X., Ma Y., Zhang Q., Mohammed M. A., Damaševičius R. (2021). A reversible watermarking system for medical color images: balancing capacity, imperceptibility, and robustness. Electronics 10:1024. 10.3390/electronics10091024 [DOI] [Google Scholar]
  175. Zhu B., Liu J. Z., Cauley S. F., Rosen B. R., Rosen M. S. (2018). Image reconstruction by domain-transform manifold learning. Nature 555 487–492. 10.1038/nature25988 [DOI] [PubMed] [Google Scholar]

Articles from Frontiers in Psychology are provided here courtesy of Frontiers Media SA

RESOURCES