Skip to main content
Elsevier Sponsored Documents logoLink to Elsevier Sponsored Documents
. 2022 Jan;77:29–52. doi: 10.1016/j.inffus.2021.07.016

Unbox the black-box for the medical explainable AI via multi-modal and multi-centre data fusion: A mini-review, two showcases and beyond

Guang Yang a,b,c,, Qinghao Ye d,e, Jun Xia f,
PMCID: PMC8459787  PMID: 34980946

Abstract

Explainable Artificial Intelligence (XAI) is an emerging research topic of machine learning aimed at unboxing how AI systems’ black-box choices are made. This research field inspects the measures and models involved in decision-making and seeks solutions to explain them explicitly. Many of the machine learning algorithms cannot manifest how and why a decision has been cast. This is particularly true of the most popular deep neural network approaches currently in use. Consequently, our confidence in AI systems can be hindered by the lack of explainability in these black-box models. The XAI becomes more and more crucial for deep learning powered applications, especially for medical and healthcare studies, although in general these deep neural networks can return an arresting dividend in performance. The insufficient explainability and transparency in most existing AI systems can be one of the major reasons that successful implementation and integration of AI tools into routine clinical practice are uncommon. In this study, we first surveyed the current progress of XAI and in particular its advances in healthcare applications. We then introduced our solutions for XAI leveraging multi-modal and multi-centre data fusion, and subsequently validated in two showcases following real clinical scenarios. Comprehensive quantitative and qualitative analyses can prove the efficacy of our proposed XAI solutions, from which we can envisage successful applications in a broader range of clinical questions.

Keywords: Explainable AI, Information fusion, Multi-domain information fusion, Weakly supervised learning, Medical image analysis

Highlights

  • We performed a mini-review for XAI in medicine and digital healthcare.

  • Our mini-review is comprehensive on most recent studies of XAI in the medical field.

  • We proposed two XAI methods and constructed two representative showcases.

  • One of our XAI models is based on weakly supervised learning for COVID-19 classification.

  • One of our XAI models is developed for ventricle segmentation in hydrocephalus patients.

1. Introduction

Recent years have seen significant advances in the capacity of Artificial Intelligence (AI), which is growing in sophistication, complexity and autonomy. A continuously veritable and explosive growth of data with a rapid iteration of computing hardware advancement provides a turbo boost for the development of AI.

AI is a generic concept and an umbrella term that implies the use of a machine with limited human interference to model intelligent actions. It covers a broad range of research studies from machine intelligence for computer vision, robotics, natural language processing to more theoretical machine learning algorithms design and recently re-branded and thrived deep learning development (Fig. 1).

Fig. 1.

Fig. 1

Left: Terminology and historical timeline of AI, machine learning and deep learning. Right: We are still at the stage of narrow AI, a concept used to describe AI systems that are capable of handling a single or limited task. General AI is the hypothetical wisdom of AI systems capable of comprehending or learning any intelligent activity a human being might perform. Super AI is an AI that exceeds human intelligence and skills.

1.1. Born of AI

AI changes almost every sector globally, e.g., enhancing (digital) healthcare (e.g., making diagnosis more accurate, allowing improved disease prevention), accelerating drug/vaccine development and repurposing, raising agricultural productivity, leading to mitigation and adaptation in climate change, improving the efficiency of manufacturing processes by predictive maintenance, supporting the development of autonomous vehicles and programming more efficient transport networks, and in many other successful applications, which make significant positive socio-economic impact. Besides, AI systems are being deployed in highly-sensitive policy fields, such as facial recognition in the police or recidivism prediction in the criminal justice system, and in areas where diverse social and political forces are presented. Therefore, nowadays, AI systems are incorporated into a wide variety of decision-making processes. As AI systems become integrated into all kinds of decision-making processes, the degree to which people who develop AI, or are subject to an AI-enabled decision, can understand how the resulting decision-making mechanism operates and why a specific decision is reached, has been increasingly debated in science and policy communities.

A collection of innovations, which are typically correlated with human or animal intelligence, is defined as the term “artificial intelligence”. John McCarthy, who coined this term in 1955, described it as “the scientific and technical expertise in the manufacture of intelligent machines”, and since then many different definitions have been endowed.

1.2. Growth of machine learning

Machine learning is a subdivision of AI that helps computer systems to intelligently execute complex tasks. Traditional AI methods, which specify step by step how to address a problem, are normally based on hard-coded rules. Machine learning framework, by contrast, leverages the power of a large amount of data (as examples and not examples) for the identification of characteristics to accomplish a pre-defined task. The framework then learns how the target output will be better obtained. Three primary subdivisions of machine learning algorithms exist:

  • A machine learning framework, which is trained using labelled data, is generally categorised as supervised machine learning. The labels of the data are grouped into one or more classes at each data point, such as “cats” or ”dogs”. The supervised machine learning framework exploits the nature from these labelled data (i.e., training data), and forecasts the categories of the new or so called test data.

  • Learning without labels is referred to as unsupervised learning. The aim is to identify the mutual patterns among data points, such as the formation of clusters and allotting data points to these clusters.

  • Reinforcement learning on the other hand is about knowledge learning, i.e., learning from experience. In standard reinforcement learning settings, an agent communicates with its environment, and is given a reward function that it tries to optimise. The purpose of the agent is to understand the effect of its decisions, and discover the best strategies for maximising its rewards during the training and learning procedure.

It is of note that some hybrid methods, e.g., semi-supervised learning (using partially labelled data) and weakly supervised (using indirect labels), are also under development.

Although not achieving the human-level intelligence often associated with the definition of the general AI, the capacity to learn from knowledge increases the amount and sophistication of tasks that can be tackled by machine learning systems (Fig. 1). A wide variety of technologies, many of which people face on a daily basis, are nowadays enabled by rapid developments in machine learning, contributing to current advancements and dispute about the influence of AI in society. Many of the concepts that frame the existing machine learning systems are not new. The mathematical underpinnings of the field date back many decades, and since the 1950s, researchers have developed machine learning algorithms with varying degrees of complexity. In order to forecast results, machine learning requires computers to process a vast volume of data. How systems equipped with machine learning can handle probabilities or uncertainty in decision-making is normally informed by statistical approaches. Statistics, however, often cover areas of research that are not associated with the development of algorithms that can learn to make forecasts or decisions from results. Although several key principles of machine learning are rooted in data science and statistical analysis, some of the complex computational models do not converge with these disciplines naturally. Symbolic approaches, compared to statistical methods, are also used for AI. In order to create interpretations of a problem and to reach a solution, these methods use logic and inference.

1.3. Boom of deep learning

Deep learning is a relatively recent congregation of approaches that have radically transformed machine learning. Deep learning is not an algorithm per se, but a range of algorithms that implements neural networks with deep layers. These neural networks are so deep that they can only be implemented on computer node clusters – modern methods of computing – such as graphics processing units (GPUs), are needed to train them successfully. Deep learning functions very well for vast quantities of data, and it is never too difficult to engineer the functionality even if a problem is complex (for example, due to the unstructured data). When it comes to image detection, natural language processing, and voice recognition, deep learning can always outperform the other types of algorithms. Deep learning assisted disease screening and clinical outcome prediction or automated driving, which were not feasible using previous methods, are well manifested now. Actually, the deeper the neural network with more data loaded for training, the higher accuracy a neural network can produce. The deep learning is very strong, but there are a few disadvantages to it. The reasoning of how deep learning algorithms reach to a certain solution is almost impossible to reveal clearly. Although several tools are now available that can increase insights into the inner workings of the deep learning model, this black-box problem still exists. Deep learning often involves long training cycles, a lot of data and complex hardware specifications, and it is not easy to obtain the specific skills necessary to create a new deep learning approach to tackle a new problem.

Although acknowledging that AI includes a wide variety of scientific areas, this paper uses the umbrella word ‘AI’ and much of the recent interest in AI has been motivated by developments in machine learning and deep learning. More importantly, we should realise that there is not one algorithm, though, that will adapt or solve all issues. Success normally depends on the exact problem that needs to be solved and the knowledge available. A hybrid solution is often required to solve the problem, where various algorithms are combined to provide a concrete solution. Each issue involves a detailed analysis into what constitutes the best-fit algorithm. Transparency of the input size, capabilities of the deep neural network and time efficiency should also be taken into consideration, since certain algorithms take a long time to train.

1.4. Stunt by the black-box and promotion of the explainable AI

Any of today’s deep learning tools are capable of generating extremely reliable outcomes, but they are often highly opaque, if not fully invisible, making it difficult to understand their behaviours. For even skilled experts to completely comprehend these so-called ’black-box’ models may be still difficult. As these deep learning tools are applied on a wide scale, researchers and policymakers can challenge whether the precision of a given task outweighs more essential factors in the decision-making procedure.

As part of attempts to integrate ethical standards into the design and implementation of AI-enabled technologies, policy discussions around the world increasingly involve demands for some form of Trustable AI, which includes Valid AI, Responsible AI, Privacy-Preserving AI, and Explainable AI (XAI), in which the XAI want to address the fundamental question about the rationale of the decision making process including both human level XAI and machine level XAI (Fig. 2). For example, in the UK, such calls came from the AI Committee of the House of Lords, which argued that the development of intelligible AI systems is a fundamental requirement if AI will be integrated as a trustworthy tool for our society. In the EU, the High-Level Group on AI has initiated more studies on the pathway towards XAI (Fig. 2). Similarly, in the USA, the Defence Advanced Research Projects Agency funds a new research effort aiming at the development of AI with more explainability. These discussions will become more urgent as AI approaches are used to solve problems in a wide variety of complicated policy making areas, as experts increasingly work alongside AI-enabled decision-making tools, for example in clinical studies, and as people more regularly experience AI systems in real life when decisions have a major impact. Meanwhile, research studies in AI continue to progress at a steady pace. XAI is a vigorous area with many on-going studies emerging and several new strategies evolving that make a huge impact on AI development in various ways.

Fig. 2.

Fig. 2

Left: Trustable AI or Trustworthy AI includes Valid AI, Responsible AI, Privacy-Preserving AI, and Explainable AI (XAI). Right: EU General Data Protection Regulation (GDPR) highlights the Fairness, Privacy, Transparency and Explainability of the AI.

While the usage of the term is inconsistent, “XAI” refers to a class of systems that have insight into how an AI system makes decisions and predictions. XAI explores the reasoning for the decision-making process, presents the positives and drawbacks of the system, and offers a glimpse of how the system will act in the future. By offering accessible explanations of how AI systems perform their study, XAI can allow researchers to understand the insights that come from research results. For example, in Fig. 3, an additional explainable surrogate module can be added to the learnt model to achieve a more transparent and trustworthy model. In other words, for a conventional machine or deep learning model, only generalisation error has been considered while adding an explainable surrogate, both generalisation error and human experience can be considered and a verified prediction can be achieved. In contrast, a learnt black-box model without an explainable surrogate module will cause concerns for the end-users although the performance of the learnt model can be high. Such a black-box model can always cause confusions like “Why did you do that?”, “Why did you not do that?”, “When do you succeed or fail?”, “How do I correct an error?”, and “Can I trust the prediction?”. The XAI powered model, on the other hand, can provide clear and transparent predictions to reassure “I understand why.”, “I understand why not.”, “I know why you succeed or fail.”, “I know how to correct an error.”, and “I understand, therefore I trust”. A typical feedback loop of the XAI development can be found in Fig. 4, which includes seven steps from training, quality assurance (QA), deployment, prediction, split testing (A/B test), monitoring, and debugging.

Fig. 3.

Fig. 3

Schema of the added explainable surrogate module for the normal machine or deep learning procedure that can achieve a more transparent and trustworthy model.

Fig. 4.

Fig. 4

A typical feedback loop of the XAI development that includes seven steps from training, quality assurance (QA), deployment, prediction, split testing (A/B test), monitoring, and debugging.

A variety of terms are used to define certain desired characteristics of an XAI system in research, public, and policy debates, including:

  • Interpretability: it means a sense of knowing how the AI technology functions.

  • Explainability: it provides an explanation for a wider range of users that how a decision has been drawn.

  • Transparency: it measures the level of accessibility to the data or model.

  • Justifiability: it indicates an understanding of the case to support a particular outcome.

  • Contestability: it implies how the users can argue against a decision.

Comprehensive surveys on general XAI can be found elsewhere, e.g., [1], [2], [3], [4]; therefore, here we provide an overview of most important concepts of the XAI. Broadly speaking, XAI can be categorised into model-specific or model-agnostic based approaches. Besides, these methods can be classified into local or global methods that can be either intrinsic or post-hoc [1]. Essentially, there are many machine learning models that are intrinsically explainable, e.g., linear models, rule-based models and decision trees, which are also known as transparent models or white-box models. However, these relatively simple models may have a relatively lower performance (Fig. 5). For more complex models, e.g., support vector machines (SVM), convolutional neural networks (CNN), recurrent neural networks (RNN) and ensemble models, we can design model-specific and post-hoc XAI strategies for each of them. For example, commonly used strategies include explanation by simplification, architecture modification, feature relevance explanation, and visual explanation [1]. In general, these more complex models may achieve better performance while the explainability becomes lower (Fig. 5). It is of note that due to the nature of the problems, less complicated models may still perform well compared to deep learning based models while preserving the explainability. Therefore, Fig. 5 just depicts a preliminary representation inspired by [1], in which XAI demonstrates its ability to enhance the trade-off between model interpretability and efficiency.

Fig. 5.

Fig. 5

Model explainability vs. model performance for widely used machine learning and deep learning algorithms. The ideal solution should have both high explainability and high performance. However, existing linear models, rule-based models and decision trees are more transparent, but with lower performance in general. In contrast, complex models, e.g., deep learning and ensembles, manifest higher performance while less explainability can be obtained. HBN: Hierarchical Bayesian Networks; SLR: Simple Linear Regression; CRF: Conditional Random Fields; MLN: Markov Logic Network; SVM: Support Vector Machine; AOG: Stochastic And-Or-Graphs; XGB: XGBoost; CNN: Convolutional Neural Network; RNN: Recurrent Neural Network; and GAN: Generative Adversarial Network.

Recently, model-agnostic based approaches attract great attention that rely on a simplified surrogate function to explain the predictions [3]. Model-agnostic approaches are not attached to a specific machine learning model. This class of techniques, in other words, distinguishes prediction from explanation. Model-agnostic representations are usually post-hoc that are generally used to explain deep neural networks with interpretable surrogates that can be local or global [2]. Below is some summary for XAI in more complex deep learning based models.

1.4.1. Model-specific global XAI

By integrating interpretability constraints into the procedure of deep learning, these model-specific global XAI strategies can improve the understandability of the models. Structural restrictions may include sparsity and monotonicity, where fewer input features are leveraged or the correlation between features and predictions is confined as monotonic). Semantic prior knowledge can also be impelled to restrict the higher-level abstractions derived from the data. For instance, in a CNN based brain tumour detection/classification model using multimodal MRI data fusion, constraints can be imposed by forcing disengaged representations that are recognisable to each MRI modality (e.g., T1, T1 post-contrast and FLAIR), respectively. In doing so, the model can identify crucial information from each MRI modality and distinguish brain tumours and sub-regions into necrotic, more or less infiltrative that can provide vital diagnosis and prognosis information. On the contrary, simple aggregation based information fusion (combining all the multimodal MRI data like a sandwich) would not provide such explainability.

1.4.2. Model-specific local XAI

In a deep learning model, a model-specific local XAI technique offers an interpretation for a particular instance. Recently, novel attention mechanisms have been proposed to emphasise the importance of different features of the high-dimensional input data to provide an explanation of a representative instance. Consider a deep learning algorithm that encodes an X-ray image into a vector using a CNN and then use an RNN to produce a clinical description for the X-ray image by using the encoded vector. For the RNN, an attention module can be applied to explain to the user what image fragments the model focuses on to produce each substantive term for the clinical description. For example, the attention mechanism will represent the appropriate segments of the image corresponding to the clinical key words derived by the deep learning model when a clinician is baffled to link the clinical key words to the regions of interest in the X-ray image.

1.4.3. Model-agnostic global XAI

In model-agnostic global XAI, a surrogate representation is developed to approximate an interpretable module for the black-box model. For instance, an interpretable decision tree based model can be used to approximate a more complex deep learning model on how clinical symptoms impact treatment response. A clarification of the relative importance of variables in affecting treatment response to clinical symptoms can be given by the IF-THEN logic of the decision tree. Clinical experts can analyse these variables and are likely to believe the model to the extent that particular symptomatic factors are known to be rational and confounding noises can be accurately removed. Diagnostic methods can also be useful to produce insights into the significance of individual characteristics in the predictions of the model. Partial dependence plots can be leveraged to determine the marginal effects of the chosen characteristics vs. the performance of the forecast, whereas individual conditional expectation can be employed to obtain a granular explanation of how a specific feature affects particular instances and to explore variation in impacts throughout instances. For example, a partial dependency plot can elucidate the role of clinical symptoms in reacting favourably to a particular treatment strategy, as observed by a computer-aided diagnosis system. On the other hand, individual conditional expectation can reveal variability in the treatment response among subgroups of patients.

1.4.4. Model-agnostic local XAI

For this type of XAI approaches, the aim is to produce model-agnostic explanations for a particular instance or the vicinity of a particular instance. Local Interpretable Model-Agnostic Explanation (LIME) [5], a well-validated tool, can provide an explanation for a complex deep learning model in the neighbourhood of an instance. Consider a deep learning algorithm that classifies a physiological attribute as a high-risk factor for certain diseases or cause of death, for which the clinician requires a post-hoc clarification. The interpretable modules are perturbed to determine how the predictions made by the change of those physiological attributes. For this perturbed dataset, a linear model is learnt with higher weights given to the perturbed instances in the vicinity of the physiological attribute. The most important components of the linear model can indicate the influence of a particular physiological attribute that can suggest a high-risk factor or the contrary can be implied. This can provide comprehensible means for the clinicians to interpret the classifier.

2. Related studies in AI for healthcare and XAI for healthcare

2.1. AI in healthcare

AI attempts to emulate the neural processes of humans, and it introduces a paradigm change to healthcare, driven by growing healthcare data access and rapid development in analytical techniques. We survey briefly the present state of healthcare AI applications and explore their prospects. For a detailed up to date review, the readers can refer to Jiang et al. [6], Panch et al. [7], and Yu et al. [8] on general AI techniques for healthcare and Shen et al. [9], Litjens et al. [10] and Ker et al. [11] on medical image analysis.

In the medical literature, the effects of AI have been widely debated [12], [13], [14]. Sophisticated algorithms can be developed using AI to ‘read’ features from a vast amount of healthcare data and then use the knowledge learnt to help clinical practice. To increase its accuracy based on feedback, AI can also be fitted with learning and self-correcting capabilities. By presenting up-to-date medical knowledge from journals, manuals and professional procedures to advise effective patient care, an AI-powered device [15] will support clinical decision making. Besides, in human clinical practice, an AI system may help to reduce medical and therapeutic mistakes that are unavoidable (i.e., more objective and reproducible) [12], [13], [15], [16], [17], [18], [19]. In addition, to help render real-time inferences for health risk warning and health outcome estimation, an AI system can handle valuable knowledge collected from a large patient population [20].

As AI has recently re-emerged into the scientific and public consciousness, AI in healthcare has new breakthroughs and clinical environments are imbued with novel AI-powered technologies at a breakneck pace. Nevertheless, healthcare was described as one of the most exciting application fields for AI. Researchers have suggested and built several systems for clinical decision support since the mid-twentieth century [21], [22]. Since the 1970s, rule-based methods had many achievements and have been seen to interpret ECGs [23], identify diseases [24], choose optimal therapies [25], offer scientific logic explanations [26] and assist doctors in developing diagnostic hypotheses and theories in challenging cases of patients [27]. Rule-based systems, however, are expensive to develop and can be unstable, since they require clear expressions of decision rules and, like any textbook, require human-authored modifications. Besides, higher-order interactions between various pieces of information written by different specialists are difficult to encode and the efficiency of the structures is constrained by the comprehensiveness of prior medical knowledge [28]. To narrow down the appropriate psychological context, prioritise medical theories, and prescribe treatment, it was also difficult to incorporate a method that combines deterministic and probabilistic reasoning procedures [29], [30].

Recent AI research has leveraged machine learning approaches, which can account for complicated interactions [31], to recognise patterns from the clinical results, in comparison to the first generation of AI programmes, which focused only on the curation of medical information by experts and the formulation of rigorous decision laws. The machine learning algorithm learns to create the correct output for a given input in new instances by evaluating the patterns extracted from all the labelled input–output pairs [32]. Supervised machine learning algorithms are programmed to determine the optimal parameters in the models in order to minimise the differences between their training case predictions and the effects observed in these cases, with the hope that the correlations found are generalisable to cases not included in the dataset of training. The model generalisability can be then calculated using the test dataset. For supervised machine learning models, grouping, regression and characterisation of the similarity between instances with similar outcome labels are among the most commonly used tasks. For the unlabelled dataset, unsupervised learning infers the underlying patterns for discovering sub-clusters of the original dataset, for detecting outliers in the data, or for generating low-dimensional data representations. However, it is of note that in a supervised manner, the recognition of low-dimensional representations for labelled dataset may be done more effectively. Machine-learning approaches allow the development of AI applications that promote the exploration of previously unrecognised data patterns without the need to define decision-making rules for each particular task or to account for complicated interactions between input features. Machine learning has therefore been the preferred method for developing AI utilities [31], [33], [34].

The recent rebirth of AI has primarily been motivated by the active implementation of deep learning – which includes training a multi-layer artificial neural network (i.e., a deep neural network) on massive datasets – to wide sources of labelled data [35]. Existing neural networks are getting deeper and typically have >100 layers. Multi-layer neural networks may model complex interactions between input and output, but may also require more data, processing time, or advanced architecture designs to achieve better performance. Modern neural networks commonly have tens of millions to hundreds of millions of parameters and require significant computing resources to perform the model training [8]. Fortunately, recent developments in computer-processor architecture have empowered the computing resources required for deep learning [36]. However, in labelled instances, deep-learning algorithms are incredibly ’data hungry.’ Huge repositories of medical databases that can be integrated into these algorithms have only recently become readily available, due to the establishment of a range of large-scale research (in particular the Cancer Genome Atlas [37] and the UK Biobank [38]), data collection platforms (e.g., Broad Bioimage Benchmark Collection [39] and the Image Data Resources [40]) and the Health Information Technology for Economic and Clinical Health (HITECH) Act, which has promised to provide financial incentives for the use of electronic health records (EHRs) [41], [42]. In general, deep learning based AI algorithms have been developed for image-based classification [43], diagnosis [44], [45], [46] and prognosis [47], [48], genome interpretation [49], biomarker discovery [50], [51], monitoring by wearable life-logging devices [52], and automated robotic surgery [53] to enhance the digital healthcare [54]. The rapid explosion of AI has given rise to the possibilities of using aggregated health data to generate powerful models that can automate diagnosis and also allow an increasingly precise approach to medicine by tailoring therapies and targeting services with optimal efficacy in a timely and dynamic manner. A non-exhaustive map of possible applications is showing in Fig. 6.

Fig. 6.

Fig. 6

A non-exhaustive map of the AI in healthcare applications.

While AI is promising to revolutionise medical practice, several technological obstacles lie ahead. Because deep learning based approaches rely heavily on the availability of vast volumes of high-quality training data, caution must be taken to collect data that is representative of the target patient population. For example, data from various healthcare settings, which include different forms of bias and noise, may cause a model trained in the data of one hospital to fail to generalise to another [55]. Where the diagnostic role has an incomplete inter-expert agreement, it has been shown that consensus diagnostics could greatly boost the efficiency of the training of the deep learning based models [56]. In order to manage heterogeneous data, adequate data curation is important. However, achieving a good quality gold standard for identifying the clinical status of the patients requires physicians to review their clinical results independently and maybe repeatedly, which is prohibitively costly at a population scale. A silver standard [57] that used natural-language processing methods and diagnostic codes to determine the true status of patients has recently been proposed [58]. Sophisticated algorithms that can handle the idiosyncrasies and noises of different datasets can improve the efficiency and safety of prediction models in life-and-death decisions.

Most of the recent advancement in neural networks has been limited to well-defined activities that do not require data integration across several modalities. Approaches for the application of deep neural networks to general diagnostics (such as analysis of signs and symptoms, prior medical history, laboratory findings and clinical course) and treatment planning are less simple. While deep learning has been effective in image detection [59], translation [60], speech recognition [61], [62], sound synthesis [63] and even automated neural architecture search [64], clinical diagnosis and treatment tasks often need more care (e.g., patient interests, beliefs, social support and medical history) than the limited tasks that deep learning can be normally adept. Moreover, it is unknown if transfer learning approaches will be able to translate models learnt from broad non-medical datasets into algorithms for the study of multi-modality clinical datasets. This suggests that more comprehensive data-collection and data-annotation activities are needed to build end-to-end clinical AI programmes. It is of note that one most recent study advocated for the use of Graph Neural Networks as a tool of choice for multi-modal causability knowledge fusion [65].

The design of a computing system for the processing, storage and exchange of EHRs and other critical health data remains a problem [104]. Privacy-preserving approaches, e.g., via federated learning, can allow safe sharing of data or models across cloud providers [105]. However, the creation of interoperable systems that follow the requirement for the representation of clinical knowledge is important for the broad adoption of such technology [106]. Deep and seamless incorporation of data across healthcare applications and locations remains questionable and can be inefficient. However, new software interfaces for clinical data are starting to show substantial adoption through several EHR providers, such as Substitutable Medical Applications and Reusable Technologies on the Fast Health Interoperability Resources platform [107], [108]. Most of the previously developed AI in healthcare applications were conducted on retrospective data for the proof of concept [109]. Prospective research and clinical trials to assess the efficiency of the developed AI systems in clinical environments are necessary to verify the real-world usefulness of these medical AI systems [110]. Prospective studies will help recognise the fragility of the AI models in real-world heterogeneous and noisy clinical settings and identify approaches to incorporate medical AI for existing clinical workflows.

AI in medicine would eventually result in safety, legal and ethical challenges [111] with respect to medical negligence attributed to complicated decision-making support structures, and have to face the regulation hurdles [112]. If malpractice lawsuits involving medical AI applications occur, the judicial system will continue to provide specific instructions as to which agency is responsible. Health providers with malpractice insurance have to be clear on coverage as health care decisions are taken in part by the AI scheme [8]. With the deployment of automatic AI for particular clinical activities, the criteria for diagnostic, surgical, supporting and paramedical tasks will need to be revised and the functions of healthcare practitioners will begin to change as different AI modules are implemented into the quality of treatment, and the bias needs to be minimised while the patient satisfaction must be maximised [113], [114].

2.2. XAI in healthcare

Despite deep learning based AI technologies will usher in a new era of digital healthcare, challenges exist. XAI can play a crucial role, as an auxiliary development (Fig. 6), for potentially solving the small sample learning by filter out clinically meaningless features. Moreover, many high-performance deep learning models produce findings that are impossible for unaided humans to understand. While these models can produce better-than-human efficiency, it is not easy to express intuitive interpretations that can justify model findings, define model uncertainties, or derive additional clinical insights from these computational ’black-boxes.’ With potentially millions of parameters in the deep learning model, it can be tricky to understand what the model sees in the clinical data, e.g., radiological images [115]. For example, research investigation has explicitly stated that being a black box is a “strong limitation” for AI in dermatology since it is not capable of doing a personalised evaluation by a qualified dermatologist that can be used to clarify clinical facts [116]. This black-box design poses an obstacle for the validation of the developed AI algorithms. It is necessary to demonstrate that a high-performance deep learning model actually identifies the appropriate area of the image and does not over-emphasise unimportant findings. Recent approaches have been developed to describe AI models including the visualisation methods. Some widely used levers include occlusion maps [117], salience maps [118], class activation maps [119], and attention maps [120]. Localisation and segmentation algorithms can be more readily interpreted since the output is an image. Model understanding, however, remains much more difficult for deep neural network models trained on non-imaging data other than images that is a current open question for ongoing research efforts [5].

Deep learning-based AI methods have gained popularity in the medical field, with a wide range of work in automatic triage, diagnosis, prognosis, treatment planning and patient management [6]. We can find many open questions in the medical field that have galvanised clinical trials leveraging deep learning and AI approaches (e.g., from grand-challenge.org). Nevertheless, in the medical field, the issue of interpretability is far from theoretical development. More precisely, it is noted that interpretabilities in the clinical sectors include considerations not recognised in other areas, including risk and responsibilities [7], [121]. Life may be at risk as medical responses are made, and leaving those crucial decisions to AI algorithms that without explainabilities and accountabilities will be irresponsible [122]. Apart from legal concerns, this is a serious vulnerability that could become disastrous if used with malicious intent.

As a result, several recent studies [120], [123], [124], [125], [126], [127] have been devoted to the exploration of explainability in medical AI. More specifically, specific analyses have been investigated, e.g., chest radiography [128], emotion analysis in medicine [129], COVID-19 detection and classification [43], and the research encourages understanding of the importance of interpretability in the medical field [130]. Besides, the exposition argues [131] that a certain degree of opaqueness is appropriate, that is, it would be more important for us to deliver empirically checked reliable findings than to dwell too hard on how to unravel the black-box. It is advised that readers consider these studies first, at least for an overview of interpretability in medical AI.

An obvious XAI approach has been taken by many researchers is to provide their predictive models with interpretability. These methods depend primarily on maintaining the interpretability of less complicated AI models while improving their performance by techniques of refinement and optimisation. For example, as Fig. 5 shows, decision tree based methods are normally interpretable, research studies have been done using automated pruning of decision trees for various classifications of illnesses [132] and accurate decision trees focused on boosting patient stratification [133]. However, such model optimisation is not always straightforward and it is not a trivial task.

Previous survey studies on XAI in healthcare can be found elsewhere, e.g., Tjoa and Guan [134] in medical XAI and Payrovnaziri et al. [135] in XAI for EHR. For specific applications, e.g., digital pathology, the readers can refer to Pocevivciute et al. [136] and Tosun et al. [137]. The research studies in XAI and medical XAI have been increased exponentially especially after 2018 alongside increasingly development of multimodal clinical information fusion (Fig. 7). In this mini-review, we only surveyed the most recent studies that were not covered by previous more comprehensive review studies. In this mini-review, we classified XAI in medicine and healthcare into five categories, which synthesised the approach by Payrovnaziri et al. [135], including (1) XAI via dimension reduction, (2) XAI via feature importance, (3) XAI via attention mechanism, (4) XAI via knowledge distillation, and (5) XAI via surrogate representations (Table 1).

Fig. 7.

Fig. 7

Publication per year for XAI and medical XAI (top) and percentage for two categories of research (bottom). Data retrieved from Scopus® (Jan 8th, 2021) by using these commands when querying this database—XAI: (ALL(“Explainable AI”) OR ALL(“Interpretable AI”) OR ALL(“Explainable Artificial Intelligence”) OR ALL(“Interpretable Artificial Intelligence”) OR ALL(“XAI”)) AND PUBYEAR = 20XX; Medical XAI: (ALL(“Explainable AI”) OR ALL(“Interpretable AI”) OR ALL(“Explainable Artificial Intelligence”) OR ALL(“Interpretable Artificial Intelligence”) OR ALL(“XAI”)) AND (ALL(“medical”) OR ALL(“medicine”)) AND PUBYEAR = 20XX, in which XX represents the actual year.

Table 1.

Summary of various XAI methods in digital healthcare and medicine including their category (XAI via dimension reduction, feature importance, attention mechanism, knowledge distillation, and surrogate representations), reference, key idea, type (Intrinsic or Post-hoc, Local or Global, and Model-specific or Model-agnostic) and specific clinical applications.

XAI Category Reference Method Intrinsic/
Post-hoc
Local/Global Model-specific/
Model-agnostic
Application
Dimension Reduction Zhang et al. [66] Optimal feature selection Intrinsic Global Model-specific Drug side effect estimation
Yang et al. [67] Laplacian Eigenmaps Intrinsic Global Model-specific Brain tumour classification using MRS
Zhao and Bolouri [68] Cluster analysis and LASSO Intrinsic Global Model-agnostic Lung cancer patients stratification
Kim et al. [69] Optimal feature selection Intrinsic Global Model-agnostic Cell-type specific enhancers prediction
Hao et al. [70] Sparse deep learning Intrinsic Global Model-agnostic Long-term survival prediction for glioblastoma multiforme
Bernardini et al. [71] Sparse-balanced SVM Intrinsic Global Model-agnostic Early diagnosis of type 2 diabetes

Feature Importance Eck et al. [72] Feature marginalisation Post-hoc Global, Local Model-agnostic Gut and skin microbiota/inflammatory bowel diseases diagnosis
Ge et al. [73] Feature weighting Post-hoc Global Model-agnostic ICU mortality prediction (all-cause)
Zuallaert et al. [74] DeepLIFT Post-hoc Global Model-agnostic Splice site detection
Suh et al. [75] Shapley value Post-hoc Global, Local Model-agnostic Decision-supporting for prostate cancer
Singh et al. [76] DeepLIFT and others Post-hoc Global, Local Model-agnostic Ophthalmic diagnosis

Attention Mechanism Kwon et al. [77] Attention Intrinsic Global, Local Model-specific Clinical risk prediction (cardiac failure/cataract)
Zhang et al. [78] Attention Intrinsic Local Model-specific EHR based future hospitalisation prediction
Choi et al. [79] Attention Intrinsic Local Model-specific Heart failure prediction
Kaji et al. [80] Attention Intrinsic Global, Local Model-specific Predictions of clinical events in ICU
Shickel et al. [81] Attention Intrinsic Global, Local Model-specific Sequential organ failure assessment/in-hospital mortality
Hu et al. [82]
Attention
Intrinsic
Local
Model-specific
Prediction of HIV genome integration site
Izadyyazdanabadi et al. [83] MLCAM Intrinsic Local Model-specific Brain tumour localisation
Zhao et al. [84] Respond-CAM Intrinsic Local Model-specific Macromolecular complexes
Couture et al. [85] Super-pixel maps Intrinsic Local Model-specific Histologic tumour subtype classification
Lee et al. [86] CAM Intrinsic Local Model-specific Acute intracranial haemorrhage detection
Kim et al. [87] CAM Intrinsic Local Model-specific Breast neoplasm ultrasonography analysis
Rajpurkar et al. [88] Grad-CAM Intrinsic Local Model-specific Diagnosis of appendicitis
Porumb et al. [89] Grad-CAM Intrinsic Local Model-specific ECG based hypoglycaemia detection
Hu et al. [43] Multiscale CAM Intrinsic Local Model-specific COVID-19 classification

Knowledge Distillation Caruana et al. [90] Rule-based system Intrinsic Global Model-specific Prediction of pneumonia risk and 30-day readmission forecast
Letham et al. [91] Bayesian rule lists Intrinsic Global Model-specific Stroke prediction
Che et al. [92] Mimic learning Post-hoc Global, Local Model-specific ICU outcome prediction (acute lung injury)
Ming et al. [93] Visualisation of rules Post-hoc Global Model-specific Clinical diagnosis and classification (breast cancer, diabetes)
Xiao et al. [94] Complex relationships distilling Post-hoc Global Model-specific Prediction of the heart failure caused hospital readmission
Davoodi and Moradi [95] Fuzzy rules Intrinsic Global Model-specific In-hospital mortality prediction (all-cause)
Lee et al. [96] Visual/textual justification Post-hoc Global, Local Model-specific Breast mass classification
Prentzas et al. [97] Decision rules Intrinsic Global Model-specific Stroke Prediction

Surrogate Models Pan et al. [98] LIME Post-hoc Local Model-agnostic Forecast of central precocious puberty
Ghafouri-Fard et al. [99] LIME Post-hoc Local Model-agnostic Autism spectrum disorder diagnosis
Kovalev et al. [100] LIME Post-hoc Local Model-agnostic Survival models construction
Meldo et al. [101] LIME Post-hoc Local Model-agnostic Lung lesion segmentation
Panigutti et al. [102] LIME like with rule-based XAI Post-hoc Local Model-agnostic Prediction of patient readmission, diagnosis and medications
Lauritsen et al. [103] Layer-wise relevance propagation Post-hoc Local Model-agnostic Prediction of acute critical illness from EHR

2.2.1. XAI via dimension reduction

Dimension reduction methods, e.g., using principal component analysis (PCA) [138], independent component analysis (ICA) [139], and Laplacian Eigenmaps [140] and other more advanced techniques, are commonly and conventionally used approaches to decipher AI models by representing the most important features. For example, by integrating multi-label k-nearest neighbour and genetic algorithm techniques, Zhang et al. [66] developed a model for drug side effect estimation based on the optimal dimensions of the input features. Yang et al. [67] proposed a nonlinear dimension reduction method to improve unsupervised classification of the 1H MRS brain tumour data and extract the most prominent features using Laplacian Eigenmaps. Zhao and Bolouri [68] stratified stage-one lung cancer patients by defining the most insightful examples via a supervised learning scheme. In order to recognise a group of “exemplars” to construct a ”dense data matrix”, they introduced a hybrid method for dimension reduction by combining pattern recognition with regression analytics. Then they used examples in the final model that are the most predictive for the outcome. Based on domain knowledge, Kim et al. [69] developed a deep learning method to extract and rank the most important features based on their weights in the model, and visualised the outcome for predicting cell-type-specific enhancers. To explore the gene pathways and their associations in patients with the brain tumour, Hao et al. [70] proposed a pathway-associated sparse deep learning method. Bernardini et al. [71] used the least absolute shrinkage and selection operator (LASSO) to prompt sparsity for SVMs for the early diagnosis of type 2 diabetes.

Simplifying the information down to a small subset using dimension reduction methods can make the underlying behaviour of the model understandable. Besides, with potentially more stable regularised models, they are less prone to overfitting, which may also be beneficial in general. Nevertheless, the possibility of losing crucial features, which may still be relevant for clinical predictions on a case-by-case basis, can be common and these important features may be neglected unintentionally by the dimensional reduced models.

2.2.2. XAI via feature importance

Researchers have leveraged the feature importance to explain the characteristics and significance of the extracted features and the correlations among features and between features and the outcomes for providing interpretability for AI models [2], [141], [142]. Ge et al. [73] used feature weights to rank the top ten extracted features to predict mortality of the intensive care unit. Suh et al. [75] developed a risk calculator model for prostate cancer (PCa) and clinically significant PCa with XAI modules that used Shapley value to determine the feature importance [143]. Sensitivity analysis of the extracted features can represent the feature importance, and essentially the more important features are those for which the output is more sensitive [144]. Eck et al. [72] defined the most significant features of a microbiota-based diagnosis task by roughly marginalising the features and testing the effect on the model performance.

Shrikumar et al. [145] implemented the Deep Learning Important FeaTures (DeepLIFT)—a backpropagation based approach to realise interpretability. Backpropagation approaches measure the output gradient for input through the backpropagation algorithm to report the significance of the feature. Zuallaert et al. [74] developed the DeepLIFT based method to create interpretable deep models for splice site prediction by measuring the contribution score for each nucleotide. A recent comparative study of different models of XAI, including DeepLIFT [145], Guided backpropagation (GBP) [146], Layer wise relevance propagation (LRP) [147], SHapley Additive exPlanations (SHAP) [148] and others, was conducted for ophthalmic diagnosis [76].

XAI, by the extraction of feature importance, can not only explain essential feature characteristics, but may also reflect their relative importance to clinical interpretation; however, numerical weights are either not easy to understand or maybe misinterpreted.

2.2.3. XAI via attention mechanism

The core concept behind the attention mechanism [149] is that the model “pays attention” only to the parts of the input where the most important information is available. It was originally proposed for tackling the relation extraction task in machine translation and other natural language processing problems. Because certain words are more relevant than others in the relation extraction task, the attention mechanism can assess the importance of the words for the purpose of classification, generating a meaning representation vector. There are various types of attention mechanisms, including global attention, which uses all words to build the context, local attention, which depends only on a subset of words, or self-attention, in which several attention mechanisms are implemented simultaneously, attempting to discover every relation between pairs of words [150]. The attention mechanism has also been shown to contribute to the enhancement of interpretability as well as to technical advances in the field of visualisation [151].

Kaji et al. [80] demonstrated particular occasions when the input features have mostly influenced the predictions of clinical events in ICU patients using the attention mechanism. Shickel et al. [81] presented an interpretable acuity score framework using deep learning and attention-based sequential organ failure assessment that can assess the severity of patients during an ICU stay. Hu et al. [82] provided “mechanistic explanations” for the accurate prediction of HIV genome integration sites. Zhang et al. [78] also built a method to learn how to represent EHR data that could document the relationship between clinical outcomes within each patient. Choi et al. [79] implemented the Reverse Time Attention Model (RETAIN), which incorporated two sets of attention weights, one for visit level to capture the effect of each visit and the other at the variable-level. RETAIN was a reverse attention mechanism intended to maintain interpretability, to replicate the actions of clinicians, and to integrate sequential knowledge. Kwon et al. [77] proposed a visually interpretable cardiac failure and cataract risk prediction model based on RETAIN (RetainVis). The general intention of these research studies is to improve the interpretability of deep learning models by highlighting particular position(s) within a sequence (e.g., time, visits, DNA) in which those input features can affect the prediction outcome.

Class activation mapping (CAM) [152] method and its variations have been investigated for XAI since 2016, and have been subsequently used for digital healthcare, especially the medical image analysis areas. Lee et al. [86] developed an XAI algorithm for the detection of acute intracranial haemorrhage from small datasets that is one of the most famous studies using CAM. Kim et al. [87] summarised AI based breast ultrasonography analysis with CAM based XAI. Zhao et al. [84] reported a Respond-CAM method that offered a heatmap-based saliency on 3D images obtained from cryo-tomography of cellular electrons. The region where macromolecular complexes were present was marked by the high intensity in the heatmap. Izadyyazdanabadi et al. [83] developed a multilayer CAM (MLCAM), which was used for brain tumour localisation. Coupling with CNN, Couture et al. [85] proposed a multi-instance aggregation approach to classify breast tumour tissue microarray for various clinical tasks, e.g., histologic subtype classification, and the derived super-pixel maps could highlight the area where the tumour cells were and each mark corresponded to a tumour class. Rajpurkar et al. [88] used Grad-CAM for the diagnosis of appendicitis from a small dataset of CT exams using video pretraining. Porumb et al. [89] combined CNN and RNN for electrocardiogram (ECG) analysis and applied Grad-CAM for the identification of the most relevant heartbeat segments for the hypoglycaemia detection. In Hu et al. [43], a COVID-19 classification system was implemented with multiscale CAM to highlight the infected areas. By the means of visual interpretability, these saliency maps are recommended. The clinician analysts who examine the AI output can realise that the target is correctly identified by the AI model, rather than mistaking the combination of the object with the surrounding as the object itself.

Attention based XAI methods do not advise the clinical end user specifically of the response, but highlight the areas of greater concern to facilitate easier decision-making. Clinical users can, therefore, be more tolerant of imperfect precision. However, it might not be beneficial to actually offer this knowledge to a clinical end user because of the major concerns, including information overload and warning fatigue. It can potentially be much more frustrating to have areas of attention without clarification about what to do with the findings if the end user is unaware of what the rationale of a highlighted segment is, and therefore the end user can be prone to ignore non-highlighted areas that could also be critical.

2.2.4. XAI via knowledge distillation and rule extraction

Knowledge distillation is one form of the model-specific XAI, which is about eliciting knowledge from a complicated model to a simplified model—enables to train a student model, which is usually explainable, with a teacher model, which is hard to interpret. For example, this can be accomplished by model compression [153] or tree regularisation [154] or through a coupling approach of model compression and dimension reduction [141]. Research studies have investigated this kind of technique for several years, e.g., Hinton et al. [155], but has recently been uplifted along with the development of AI interpretability [156], [157], [158]. Rule extraction is another widely used XAI method that is closely associated with knowledge distillation and can have a straightforward application for digital healthcare, for example, decision sets or rule sets have been studied for interpretability [159] and Model Understanding through Subspace Explanations (MUSE) method [160] has been developed to describe the projections of the global model by considering the various subgroups of instances defined by user interesting characteristics that also produces explanation in the form of decision sets.

Che et al. [92] introduced an interpretable mimic-learning approach, which is a straightforward knowledge-distillation method that uses gradient-boosting trees to learn interpretable structures and make the baseline model understandable. The approach used the information distilled to construct an interpretable prediction model for the outcome of the ICU, e.g., death, ventilator usage, etc. A rule-based framework that could include an explainable statement of death risk estimation due to pneumonia was introduced by Caruana et al. [90]. Letham et al. [91] also proposed an XAI model named Bayesian rule lists, which offered certain stroke prediction claims. Ming et al. [93] developed a visualisation approach to derive rules by approximating a complicated model via model induction at different tasks such as diagnosis of breast cancer and the classification of diabetes. Xiao et al. [94] built a deep learning model to break the dynamic associations between readmission to hospital and possible risk factors for patients by translating EHR incidents into embedded clinical principles to characterise the general situation of the patients. Classification rules were derived as a way of providing clinicians interpretable representations of the predictive models. Davoodi and Moradi [95] developed a rule extraction based XAI technique to predict mortality in ICUs and Das et al. [161] used a similar XAI method for the diagnosis of Alzheimer’s disease. In the LSTM-based breast mass classification, Lee et al. [96] incorporated the textual reasoning for interpretability. For the characterisation of stroke and risk prediction, Prentzas et al. [97] implemented the argumentation theory for their XAI algorithm training process by extracting decision rules.

XAI approaches, which rely on knowledge distillation and rule extraction, are theoretically more stable models. The summarised representations of complicated clinical data can provide clinical end-users with the interpretable results intuitively. However, if the interpretation of these XAI results could not be intuitively understood by clinical end-users, then the representations are likely to make it much harder for the end-users to comprehend.

2.2.5. XAI via surrogate representation

An effective application of XAI in the medical field is the recognition of individual health-related factors that lead to disease prediction using the local interpretable model-agnostic explanation (LIME) method [5] that offers explanations for any classifier by approximating the reference model with a surrogate interpretable and “locally faithful” representation. LIME disrupts an instance, produces neighbourhood data, and learns linear models in the neighbourhood to produce explanations [162].

Pan et al. [98] used LIME to analyse the contribution of new instances to forecast central precocious puberty in children. Ghafouri-Fard et al. [99] have applied a similar approach to diagnose autism spectrum disorder. Kovalev et al. [100] proposed a method named SurvLIME to explain AI base survival models. Meldo et al. [101] used a local post-hoc explanation model, i.e., LIME, to select important features from a special feature representation of the segmented lung suspicious objects. Panigutti et al. [102] developed the “Doctor XAI” system that could predict the readmission, diagnosis and medications order for the patient. Similar to LIME, the implemented system trained a local surrogate model to mimic the black-box behaviour with a rule-based explanation, which can then be mined using a multi-label decision tree. Lauritsen et al. [103] tested an XAI method using Layer-wise Relevance Propagation [163] for the prediction of acute critical illness from EHR.

Surrogate representation is a widely used scheme for XAI; however, the white-box approximation must accurately describe the black-box model to gain trustworthy explanation. If the surrogate models are too complicated or too abstract, the clinician comprehension might be affected.

3. Proposed method

3.1. Problem formulation

In this study, we have demonstrated two typical but important applications of using XAI, which have been developed for classification and segmentation—two mostly widely discussed problems in medical image analysis and AI-powered digital healthcare. Our developed XAI techniques have been manifested using CT images classification for COVID-19 patients and segmentation for hydrocephalus patients using CT and MRI datasets.

3.2. XAI for classification

In this subsection, we provide a practical XAI solution for explainable COVID-19 classification that is capable of alleviating the domain shift problem caused by multicentre data collected for distinguishing COVID-19 patients from other lung diseases using CT images. The main challenge for multicentre data is that hospitals are likely to use different scanning protocols and parameters for CT scanners when collecting data from patients leading to distinct data distribution. Moreover, it can be observed that images obtained from various hospitals are visually different although they are imaging the same organ. If a machine learning model is trained on data from one hospital and tested on the data from another hospital (i.e., another centre), the performance of the model often degrades drastically. Besides, another challenge is that only patient-level annotations are available commonly but image-level labels are not since it would take a large amount of time for radiologists to annotate them [164]. Therefore, we propose a weakly supervised learning based classification model to cope with these two problems. Besides, an explainable diagnosis module in the proposed model can also offer the auxiliary diagnostic information visually for radiologists. The overview of our proposed model is illustrated in Fig. 8.

Fig. 8.

Fig. 8

The overview of our proposed model. P(c|Si) denotes the probability of the Section Si, and P(c|P) represents the probability of the patient who is COVID-19 infected or not. QR2×2×C indicates the noise transaction from the probability of the true label P(yc|I) to the noise label P(zc|I). Besides, ϕ() is a non-linear feature transformation function, which projects the feature into embedding space.

3.2.1. Explainable diagnosis module (EDM)

As the predicting process of deep learning models is in a black-box, it is desirable to develop an explainable technique in medical image diagnosis, which provides an explainable auxiliary tool for radiologists. For common practice, CAM can generate the localisation maps for the prediction through the weighted sum of feature maps from the backbone networks such as ResNet [165]. Suppose FkRH×W is the kth feature map with the shape of H×W, and WfcRK×C, where K is the number of feature maps. Therefore, the class score for class c can be computed as

sc=k=1KWk,cfc1HWi=1Hj=1WFi,jk. (1)

Therefore, the activation map Acfc for class c can be defined by

(Acfc)i,j=k=1KWk,cfcFi,jk. (2)

However, generating CAMs is not an end-to-end process, in which the network should be firstly trained on the dataset and utilises the weights of the last fully connected layer to compute the CAMs, bringing extra computation. To tackle this drawback, in our explainable diagnosis module (EDM), we replace the fully connected layer with a 1 × 1 convolutional layer of which the weight Wconv shares the same mathematical form as Wfc. So we can reformulate Eq. (1) as

sc=1HWi=1Hj=1Wk=1KWk,cconvFi,jk=1HWi=1Hj=1W(Acconv)i,j, (3)

where Acconv is the activation map for class c that can be learnt adaptively during the training procedure. The activation map produced by the EDM can not only accurately indicate the importance of the region from CT images and locate the infected parts of the patients, but can also offer the explainable results which are able to account for the prediction.

3.2.2. Slice integration module (SIM)

Intuitively, each COVID-19 patient case has a different severity. Some patients are severely infected with large lesions, while most of the positive cases can be mild of which only a small portion of the CT volume is infected. Therefore, if we directly apply the patient level annotations as the labels for the image slices, the data would be extremely noisy leading to poor performance as the consequence. To overcome this problem, instead of relying on single images, we propose a slice integration module (SIM) and use the joint distribution of the image slices to model the probability of the patient being infected or not. In our SIM, we assume that the lesions are consecutive and the distribution of the lesion positions is consistent. Therefore, we adopt a section based strategy to handle this problem and fit this into a Multiple Instance Learning (MIL) framework [166]. In the MIL, each sample is regarded as a bag, which is composed of a set of instances. A positive bag contains at least one positive instance, while the negative bag solely consists of negative instances. In our scenario, only patient annotations (bag labels) are provided, and the sections can be regarded as instances in the bags.

Given a patient P=[I1,I2,,In] with n CT slices, we divide them into disjoint sections P={Si}i=1|S|, where |S| is the total amount of sections for patient P, that is

|S|=max1,nls. (4)

Here ls is the section length, which is a designed parameter. Then we integrate the probability of each section as the probability of the patient, that is

P(c|P)=P(c|{Si}i=1|S|)=11+i=1|S|(1P(c|Si)1), (5)

where P(c|Si) is the probability of the ith section Si that belongs to class c. By taking the k-max probability of the images for each class to compute the section probability, we can mitigate the problem that some slices may contain few infections, which can hinder the prediction for the section. The k-max selection method can be formulated as

P(c|Si)=σ1kmaxs(j)Mj=1ksc(j),
s.t.MSi,|M|=k. (6)

where sc(j) is the top jth class score of the slice in the ith section for the class c, and σ() represents the Sigmoid function. Then we apply the patient annotations y to compute the classification loss, which can be formulated as

Lcls=c=01yclogP(c|P)+(1yc)log(1P(c|P)). (7)

3.2.3. Noisy correction module (NCM)

In real-world applications, radiologists would only diagnose the disease from one image. Therefore, it is also significant for improving the prediction accuracy on single images. However, the image-level labels are extremely noisy since only patient-level annotations are available. To further alleviate the negative impact of patient-level annotations, we propose a noisy correction module (NCM). Inspired by [167], we model the noise transaction distribution P(zc=i|yc=j,I), which transforms the true posterior distribution P(yc|I) to the noisy label distribution P(zc|I) by

P(zc=i|I)=jP(zc=i|yc=j,I)P(yc=j|I). (8)

In practice, we estimate the noise transaction distribution Qijc=P(zc=i|yc=j,I) for the class c via

Qijc=P(zc=i|yc=j,I)=exp(wijcϕ(I)+bijc)iexp(wijcϕ(I)+bijc), (9)

where i,j{0,1}; ϕ() is a nonlinear mapping function implemented by convolution layers; wijc and bijc are the trainable parameters. The noise transaction score Tijc=wijcϕ(I)+bijc represents the confidence score of the transaction from the true label i to the noise label j for the class c. Therefore, Eq. (8) can be reformulated as

P(zc=i|I)=jQijcP(yc=j|I). (10)

By estimating the noisy label distribution P(zc|I) for patient P, the noisy classification loss can be computed by

Lnoisy=1Nn=1Nc=01[ycnlogP(zc=1|In)+(1ycn)logP(zc=0|In)]. (11)

By combining Eqs. (7), (11), we can obtain the total loss function for our XAI solution of an explainable COVID-19 classification, that is

L=Lcls+λLnoisy, (12)

where λ is a hyper-parameter to balance the classification loss Lcls and the noisy classification loss Lnoisy.

3.3. XAI for segmentation

In this subsection, we introduce an XAI model that is applicable for the explainable brain ventricle segmentation using multimodal MRI data acquired from the hydrocephalus patients. Previous methods [168], [169] have conducted experiments using images with a slice thickness of less than 3 mm. This is because the smaller of the image thickness, the more images could be obtained, which helps improve the representation power of the model. However, in a real-world scenario, it is not practical for clinicians to use these models because labelling these image slices is extremely labour-intensive and time-consuming. Therefore, it is more common for the annotations of images with larger slice thicknesses, which are easily available while those images with smaller slice thickness are not. Besides, models trained only on thick-slice images have poor generalisation on thin-slice images. To alleviate these problems, we proposed a thickness agnostic image segmentation model, which can be applicable for both thick-slice and thin-slice images, but only requires the annotations of thick-slice images during the training procedure.

Suppose we have a set of thick-slice images DS={(xs,ys)|xsRH×W×3,ysRH×W} and a set of thin-slice images DT={xt|xtRH×W×3}. The main idea of our model is to utilise the unlabelled thin-slice images DT to minimise the model performance gap between thick-slice and thin-slice images while a post-hoc XAI can also be developed.

3.3.1. Segmentation network

With the wide applications of deep learning methods, the encoder–decoder based architectures are usually adopted in automated high accuracy medical image segmentation. The workflow of our proposed segmentation network is illustrated in Fig. 9. Inspired by the U-Net [170] model, we replace the original encoder with ResNet-50 [165] pre-trained on ImageNet dataset [171] since it can provide better feature representation for the input images. In addition, the decoder of the U-Net has at least a couple of drawbacks: 1) the increase of low-resolution feature maps can bring a large amount of computational complexity, and 2) interpolation methods [172] such as bilinear interpolation and bicubic interpolation do not bring extra information to improve the segmentation. Instead, the decoder of our model adopts sub-pixel convolution for constructing segmentation results. The sub-pixel convolution can be represented as

FL=SP(WLFL1+bL), (13)

where SP() operator transforms and arranges a tensor shaped in H×W×C×r2 into a tensor with the shape of rH×rW×C, and r is the scaling factor. FL1 and FL are the input feature maps and output feature maps. WL and bL are the parameters of the sub-pixel convolution operators for the layer L.

Fig. 9.

Fig. 9

Overview of our proposed XAI model for explainable segmentation. Here ResBlock represents the residual block proposed in the ResNet [165].

3.3.2. Multimodal training

As aforementioned, the thick-slice images with annotations are available. Therefore, in order to minimise the performance gap between thick-slice images and thin-slice images. We apply a multimodal training procedure to jointly optimise for both types of images. Overall, the objective function of our proposed multimodal training can be computed as

L(xt,xs)=LS(ps,ys)+βLT(pt), (14)

where β is a hyper-parameter for weighting the impact of LS and LT. ps and pt are the prediction of the segmentation probability maps shaped in H×W×C for thick-slice images and thin-slices images, respectively. In particular, LS is the cross-entropy loss defined as follows

LS(ps,ys)=1HWCn=1HWc=1Cysn,clogpsn,c. (15)

For the unlabelled thin-slice images, we assume that LT can push the features away from the decision boundary of the feature distributions of the thick-slice images, thus achieving distribution alignment. Besides, according to [173], minimising the distance between the prediction distribution p and the uniform distribution U=1C can diminish the uncertainty of the prediction. To measure the distance of these two distributions, the objective function LT can be modelled by the f-divergence, that is

LT(pt)=1HWCn=1HWc=1CDf(ptn,cU)=1HWCn=1HWc=1Cf(Cptn,c). (16)

Most existing methods [173], [174] tend to choose f(x)=xlogx, which is alternatively named as KL-divergence. However, one of the main obstacle is that when adopting f(x)=xlogx, the gradient of LT would be extremely imbalanced. To be more specific, it can assign a large gradient to the easily classified samples, while assigning a small gradient to hardly classified samples. Therefore, in order to mitigate the unbalancing problem during the optimisation, we incorporate Pearson χ2-divergence (i.e., f(x)=x21) rather than using the KL-divergence for LT, that is

LT(pt)=CHWn=1HWc=1C(ptn,c)2. (17)

After applying the Pearson χ2-divergence, the gradient imbalanced issue can be mitigated since the slope of the gradient is constant, which can be verified by taking the second order derivative of LT.

During the training procedure, L(xt,xs) is optimised alternatively for both thick-slice and thin-slice images.

3.3.3. Latent space explanation

Once the model is trained using multimodal datasets, the performance of the network can be quantitatively evaluated by volumetric or regional overlapping metrics, e.g., Dice scores. However, the relation between network performance and input samples remains unclear. In order to provide information about the characteristics of data and their effect on model performance, through which users can set their expectations accordingly, we investigate the feature space and their correlation with the model performance. For feature space visualisation, we extract the outputs of the encoder module of our model, and then decompose them into a two-dimensional space via Principal Component Analysis (PCA). For estimating the whole space, we use a multi-layer perceptron to fit the decomposed samples and their corresponding Dice scores, which can provide an understanding of Dice scores for particular regions of interests in the latent space where there is no data available. Therefore, through analysing the characteristics of the samples in the latent space, we can retrieve the information about the relationships between samples and their prediction power.

3.4. Implementation details

For both our classification and segmentation tasks, we used ResNet-50 [165] as the backbone network pre-trained on ImageNet [171]. For classification, we resized these images into a spatial resolution of 224 × 224. During the training procedure, we set λ=1×104, the dropout rate as 0.7, and the L2 weight decay coefficient as 1×105. Besides, ls was set to 16, and k was set to 8 for the sake of computing the patient-level probability. For segmentation, we set β=1×102 for balancing the impact of supervised loss and unsupervised loss. During the training, Adam [175] optimiser was utilised with a learning rate 1×103. The training procedure is terminated after 4,000 iterations with batch size 8. All of the experiments were conducted on a workstation with 4 NVIDIA RTX GPUs using PyTorch framework with version 1.5.

4. Experimental settings and results

4.1. Showcase I: Classification for COVID-19

4.1.1. Datasets

We collected CT data from four different local hospitals in China and removed the personal information to ensure data privacy. The information of our collected data is summarised in Fig. 10. In total, there were 380 CT volumes of the patients who tested COVID-19 positive (reverse transcription polymerase chain reaction test confirmed) and 424 COVID-19 negative CT volumes. For a fair comparison, we trained the model on the cross-centre datasets collected from hospital A, B, C, and D. For an unbiased independent testing, CC-CCII data [176], a publicly available dataset, which contained 2,034 CT volumes with 130,511 images, was adopted to verify the effectiveness of the trained models.

Fig. 10.

Fig. 10

Class distribution of the collected CT data. The numbers in the sub-figures (a) and (b) represent the counts for the patient-level statistics and image-level statistics, respectively. The data collected from several clinical centres can result in great challenges in learning discriminative features from those class-imbalanced centres.

4.1.2. Data standardisation, pre-processing and augmentation

Following the protocol described in [176], we used the U-Net segmentation network [170] to segment the CT images. Then, we randomly cropped a rectangular region whose aspect ratio was randomly sampled in [34,43], the area was randomly sampled in [90%,100%], and the region was then resized into 224 × 224. Meanwhile, we randomly flipped the input volumes horizontally with 0.5 probability. The input data would be a set of CT volumes, which were composed of consecutive CT image slices.

Table 2.

Comparison results of our method vs. state-of-the-art methods performed on the CC-CCII dataset.

Annotation Method Patient Acc. (%) Precision (%) Sensitivity (%) Specificity (%) AUC (%)
Patient-level ResNet-50 [165] 53.44 64.45 63.03 35.71 53.24
COVID-Net [177] 57.13 62.53 84.70 6.16 49.58
COVNet [178] 69.96 70.20 93.33 26.75 81.61
VB-Net [179] 76.11 75.84 92.73 45.38 88.34
Ours 89.97 92.99 91.44 87.25 95.53

Image-level ResNet-50 [165] 52.56 61.60 71.27 18.06 50.19
COVID-Net [177] 60.03 64.81 83.91 15.98 58.39
COVNet [178] 75.55 79.90 83.24 61.37 79.48
Ours 80.41 88.56 80.15 80.89 86.06

4.1.3. Quantitative results

We compared our proposed classification model with several state-of-the-art COVID-19 CT classification models [165], [177], [178], [179]. Table 2 summarises the experimental results of COVID-19 classification on the CC-CCII data. For image-level annotations, ResNet-50 [165] and COVID-Net [177] simply treated patient-level labels as the image labels. Different from methods proposed by [165], [177], [178], VBNet [179] utilised the 3D residual convolutional neural network to train with patient-level annotations on the whole CT volumes rather than single slices. Besides, COVNet [178] extracted prediction scores from each slice in the CT volumes with ResNet and aggregated the prediction scores via a max-pooling operator to get the patient-level probability.

In Table 2, we can find that our method achieved the best performance among these SOTA methods. In particular, our method obtained a better performance by 7.2% on AUC compared to VB-Net [179] on the patient-level indicating that our method can be applicable for the real-world scenario. This also verified the benefit of modelling section information in the CT volumes via our proposed SIM, which we believe is also vital to the improvement of the classification performance. Besides, our method significantly outperformed other methods by at least 40% with respect to the specificity while maintaining high sensitivity, which is also a crucial indication for diagnosing COVID-19. In addition, models trained on patient-level annotations could achieve better performance compared to those trained on image-level labels. This is because the noise in the image labels could have a negative impact during the training, which might degrade the representation ability of the model. According to [180], models trained on images may rely on learning the textures of images that were highly discriminative among multiple centres. Therefore, these trained models might be overfitted and biased to the texture features of the images collected from different centres, which could explain the phenomenon that these methods (i.e., [165], [177]) were poorly performed on the unseen centres.

In another aspect, for CT volumes, the sequential ordering of CT image slices is also informative. COVID-Net [178] took the most discriminative slice as the representation of the whole CT volume, which ignored the encoding of adjacent slices. This would enforce the model only detect the most discriminative slice, leading to the bias towards positive cases, which could impede the detecting of negative cases that resulted in a low specificity. On the contrary, VBNet proposed by Ouyang et al. [179] preserved the sequential information by training on the whole CT volumes. In contrast, we partitioned the CT volume into several sections in order to preserve the sequential information to some extent. Besides, VB-Net was trained with stronger supervision that it utilised additional masks for its supervised training. For our method, we only used patient-level annotations that were much more efficient. More importantly, our method achieved better performance on both AUC and accuracy compared to VBNet [179] and COVNet [178].

In addition, we also provided the Precision–Recall (PR) and Receiver Operating Characteristic (ROC) curves to compare different methods on patient-level annotations and image-level annotations (Fig. 11, Fig. 12). From the figure, we can observe that models trained on image-level annotations (e.g., ResNet [165] and COVID-Net [177]) were poorly performed since their AUCs were close to 50% which indicated a random guess. In contrast, models trained on patient-level was more reliable since their AUCs were greater than 50%. In particular, we found that overall our proposed method remained the best-performed algorithm with an AUC of 95.53% at the patient-level and 86.06% at the image-level. These results verified our assumption that for mild COVID-19 cases, most of the image slices are disease-free. It is of note that the AUC of the ROC or the AUC of the precision recall curve could be too general because they contain unrealistic decision thresholds, while metrics like accuracy, sensitivity or the F1 score are measured at a single threshold that reflects an individual single probability or predicted risk, rather than a range of individuals or risk. Further study using deep ROC analysis will be investigated [181].

Fig. 11.

Fig. 11

The Receiver Operating Characteristic (ROC) curves of different compared methods.

Fig. 12.

Fig. 12

The Precision–Recall (PR) curves of different compared methods.

4.1.4. Qualitative results

In order to make the prediction to be more explainable, we used the trained model to visualise the CAMs and bounding boxes generated by our EDM as described above. Fig. 13 shows the visualisation results of the derived CAMs (i.e., Aconv). In this figure, we can clearly observe that our method tended to pay more attention to the discriminative part of the images so as to make the predictions. For example, in the first column, the lower left part of the lung was seriously infected and had a large area of lesions. Therefore, our method would make the predictions that the image was classified as COVID-19 positive, demonstrating the capability of our XAI model to make explainable predictions.

Fig. 13.

Fig. 13

Examples of the CAMs Aconv generated by our proposed EDM for classifying COVID-19 positive patients. The first row contains the original CT-scan image slices, and the second row illustrates the heatmaps of CAMs Aconv with bounding boxes confined to the infected areas.

In addition, based on the results of the derived CAMs, we also extracted the lesion bounding boxes from the CAMs. It can be found that our method was capable of yielding accurate bounding boxes from the salient part of the CAMs, as illustrated in Fig. 13, which further confirmed that our XAI method was applicable to be an auxiliary diagnosis tool for the clinicians.

To further illustrate the learnt features from our proposed method, we extracted the feature from the backbone network of our architecture, and used T-SNE [182] visualisation technique to transform the features extracted from the backbone network of our proposed model, and visualised the distribution of the classified images as shown in Fig. 14. In this figure, we can find the distinctive visual characteristics of the CT images from different hospitals (i.e., Hospital A, B, C, and D). Besides, it can be observed that the COVID-19 positive images were mostly clustered together, and negative images were mainly distributed in another cluster. More interestingly, in the cluster of the negative images, we can find several positive images in this cluster since these images were scanned from patients who were tested COVID-19 positive. In our intuition, we assume that for some mild cases, lesions were not presented in all of the CT slices. Therefore, there were indeed disease-free CT slices that could be falsely labelled as COVID-19 positive, which verified our assumption.

Fig. 14.

Fig. 14

T-SNE visualisation [182] of the learnt features from CT images. Original images are sampled from four different hospitals and represented in the figure. Besides, a falsely annotated image is drawn from the negative cluster.

Additionally, in order to explain each individual prediction, we adopted the LIME method [5] to investigate the contribution of each pixel for the prediction. Instead of using the individual pixel, we divided an image into super-pixels, which were composed of interconnected pixels with similar imaging patterns. Fig. 15 shows the explanations via LIME for COVID-19 positive images. In each pair of images, we visualised the super-pixels that contributed to the COVID-19 positive prediction results. We can observe that the lesion parts would explain for the positive prediction, which is reasonable to our deep learning model.

Fig. 15.

Fig. 15

Visualisation of the super-pixels that are positively contributed to the predictions via the LIME method [5].

However, the LIME method could only quantitatively estimate the importance according to how close the combination of super-pixels was to the original instance. It discarded the global view of the individual feature contributed by the super-pixels. To overcome this drawback, we further leveraged Kernel SHapley Additive exPlanations (Kernel SHAP) method [183] to estimate the contribution of each super-pixel quantitatively by the SHAP value. Samples explained by the Kernel SHAP are demonstrated in Fig. 16. We can observe that the super-pixels contained lesion areas positively contributed to the positive prediction, while those super-pixels related to the backgrounds or disease-free areas would reflect the contribution to negative prediction.

Fig. 16.

Fig. 16

The SHAP values for different super-pixels of the sampled images. We computed the SHAP values through the Kernel SHAP method [183]. The super-pixel with positive SHAP value indicates the positive impact to the positive prediction, while the negative value means that the super-pixel contributes to the negative prediction.

4.2. Showcase II: Segmentation for hydrocephalus

4.2.1. Datasets

The studied cohort included 20 normal elderly people, 20 patients with cerebral atrophy, 64 patients with normal pressure hydrocephalus, and 51 patients with acquired hydrocephalus (caused by subarachnoid haemorrhage, brain trauma or brain tumour). CT scans of the head were performed using two CT instruments, one of which was the SOMATOM Definition Flash from Siemens, Germany, and the other was the SOMATOM Emotion 16 from Siemens, Germany. Secondly, MRI examinations were conducted using a 1.5T MR scanner(Avanto, Siemens, Erlangen, Germany) and a 3.0T MRI scanner(Prisma, Siemens, Erlangen, Germany). The slice thickness of the CT images includes: 0.5 mm, 1.0 mm, 1.5 mm, 2.0 mm, 4.8 mm, 5.0 mm. The slice thickness of the MRI images includes: 1.0 mm, 7.8 mm, 8.0 mm. For experiments, we randomly split the thick-slice and thin-slice images into training, validation and testing sets. The details of the dataset are summarised in Table 3.

Table 3.

The number of thick-slice and thin-slice images used in our study.

Modality Training set
Validation set
Test set
Thick-slice Thin-slice Thick-slice Thin-slice Thick-slice Thin-slice
MRI 810 1,303 203 326 189 982
CT 2,088 2,076 523 519 309 492

4.2.2. Data standardisation, pre-processing and augmentation

For the pre-processing of these data, we normalised images using the z-score normalisation scheme, which was done by subtracting its mean then divided by its standard deviation. For anomaly pixels, we clipped them within the range of 1-quantile and 99-quantile. For data augmentation, we resized the images using a bicubic interpolation method and resized masks with the nearest interpolation. Then we flipped the images horizontally with 0.5 probability, and scaled the hue, saturation, and brightness with coefficients uniformly drawn from [0.8,1.2].

4.2.3. Quantitative results

Table 4 shows the segmentation performance of various compared models on different modalities. All of the models were trained on the thick-slice images with annotations and the unlabelled thin-slice images. We can observe that our proposed method outperformed all of the compared state-of-the-art methods by a large margin on the mixed datasets (i.e., the mixture of thick-slice and thin-slice images) with at least 4.4% of the Dice scores. It is of note that all three models achieved similar segmentation performance on thick-slice images. However, our proposed method gained a significant improvement on the thin-slice images for both MRI and CT scans. The primary reason is that our model could diminish the uncertainty of these thin-slice images while achieving the distribution alignment between thick-slice and thin-slice images, which could enhance the representation and generalisation capabilities of our model. Besides, we also investigated the effectiveness of LS and LT in Table 5. In the table, we can find that when only training on thick-slice images, the model performed perfectly on thick-slice images while performing poorly on thin-slice images, since the distribution of these two kinds of slices could vary. Moreover, the performance of the models trained only on unlabelled thin-slice images degraded sharply because of the lack of annotations to guide the segmentation. In the Exp.3 as shown in Table 5, our model could gain significant improvement on the thin-slice images while preserving good performance on the thick-slice images, which demonstrated that our trained model was applicable for both types of images for both CT and MRI modalities.

Table 4.

Comparison results (Dice scores) of our method vs. other state-of-the-art methods. Mixed represents the test set containing both thick-slice and thin-slice images.

Method MRI
CT
Thick Thin Mixed Thick Thin Mixed
U-Net [170] 0.9226 0.7665 0.8353 0.9351 0.7987 0.8513
U-Net++ [184] 0.9159 0.8495 0.8602 0.9421 0.7797 0.8424
Ours 0.9323 0.9056 0.9099 0.9365 0.8697 0.8954
Table 5.

Dice scores comparison for verifying the effectiveness of each loss term. Mixed represents the test set containing both thick-slice and thin-slice images.

Exp. LS LT MRI
CT
Thick Thin Mixed Thick Thin Mixed
1 0.9390 0.8199 0.8391 0.9438 0.8345 0.8767
2 0.0034 0.0108 0.0110 0.0109 0.0006 0.0069
3 0.9323 0.9056 0.9099 0.9365 0.8697 0.8954

Besides, in order to interpret the black-box segmentation model, we extracted the lowest bottom features and projected them into a 2D latent space using the PCA technique. We then computed the Dice score for each sample and visualised it in Fig. 17. In this figure, we can observe that slices sampled from the orange circle all contained a small region of ventricle where the model could not perform well. However, images from the green and yellow circle had multiple ventricles, which took a large proportion of the images. Therefore, these images could be well-predicted by our model.

Fig. 17.

Fig. 17

The visualisation of the Dice scores of the projected images. The plane was computed by smoothing the Dice scores. It is of note that images sharing similar characteristics were clustered together. On the left-hand side and right-hand side, samples from different regions of the plane are presented.

4.2.4. Qualitative results

To qualitatively examine the performance of our model and other state-of-the-art models, we presented some visualisation results of the CT and MRI images with thin-slices in Fig. 18, and computed the Dice scores for each segmentation result. For MRI images, our model and U-Net++ [184] were able to segment four ventricles in the brain. In particular, our model could predict the third ventricle in the brain more completely compared to the prediction generated by the U-Net++ [184] due to the informative feature representation by the pre-trained encoder. However, for CT images, the performance varied among different models. The primary reason is that original CT volumes contained the skull which could cause the brain to be visually unclear, after removing the skull, the contrast of the images could be largely distinct. More concretely, for those images with low contrast, (e.g., the row 1 and row 5 in Fig. 18), all of the three compared methods were capable of predicting the left lateral and right lateral ventricles. However, for those images with high contrast (e.g., the row 2 and row 4 in Fig. 18), our proposed method could predict most of the ventricle part in the brain while U-Net and U-Net++ failed.

Fig. 18.

Fig. 18

The visualisation of the 3D brain ventricles segmentation results using different compared models. The right lateral ventricle is coloured in red; the left lateral ventricle is coloured in green; the yellow coloured region represents the third ventricle; and the blue region represents the fourth ventricle. (For interpretation of the references to colour in this figure legend, the reader is referred to the web version of this article.)

In addition, we used the segmentation results generated by compared models to reconstruct the 3D images of each ventricle. The example is illustrated in Fig. 19. We can observe that U-Net [170] could hardly predict the ventricles on thin-slice images, while U-Net++ [184] was able to segment the left lateral and right lateral ventricles by taking advantage of dense connections of the intermediate feature maps. In contrast, our proposed method could not only predict the two ventricles mentioned above, but could also segment the third ventricle and the fourth ventricle well. One limitation of our model is that it could not predict the connection region between the third ventricle and fourth ventricle because the area is too small to be distinguished.

Fig. 19.

Fig. 19

Three-dimensional visualisation of the predictions on thin-slice MRI images for each ventricle segmented by different comparison models. The 3D segmentation results were visualised from the axial plane, the coronal plane, and the sagittal plane. Colouring scheme is consistent with Fig. 18.

4.3. Discussions

In missions increasingly vital to human healthcare, AI is being deployed. Automated decisions should be explainable in order to create trust in AI and prevent an algorithm-based totalitarian society. This is not just a human right, for example, enshrined in the European GDPR, but an ultimate goal for algorithm developers who want to know if the necessary clinical characteristics are captured by the decision support systems. XAI should be possible to provide explanations in a systematic manner in order to make the explainability scalable. To construct a surrogate while-box model for the black-box model used to make a prediction, a typical solution is to use simpler, more intuitive decision algorithms. There is a chance, though, that the surrogate model is too complicated or too abstract for it to be truly understandable for humans.

In this study, we have firstly provided a mini-review for XAI methods and their specific applications in medicine and digital healthcare that is followed by two example showcases that we have developed. From our two showcases, we have explored the classification model and segmentation model in terms of sensitivity (i.e., LIME [5] and Kernel SHAP [183]) and decomposition (i.e., T-SNE [182] and CAMs). For LIME and Kernel SHAP methods, the individual sample can be analysed and interpreted with each super-pixel, which is useful for individual diagnosis. These methods can provide a straightforward view of how local explanations affect the final predictions. In this study, although we have not proposed novel XAI methods, the two showcases have demonstrated new pipelines of using XAI techniques for COVID-19 classification and segmentation for hydrocephalus. It is of note that both classification and segmentation are the two most widely investigated problems in medical data analysis with the arising of machine learning and deep learning. For example, machine and deep learning algorithms have been widely used for chest X-ray and CT images analysis for COVID-19 patients; however, there are still some common pitfalls in detection, diagnosis and prognosis of using these algorithms, e.g., data duplication, image quality issues and labelling accuracy and reproducibility, and most importantly, the lack of model interpretability [185], [186], [187].

On the other hand, T-SNE provides us with an insight into the strength and weakness of our proposed models. For example, in Fig. 17, the distribution of the decomposed image features has an association with the prediction performance, which indicates the weakness of the black-box segmentation models. Meanwhile, the distribution of decomposed image features also reveals the clustered characteristics of the raw inputs (Fig. 14), which can help us to find the reason why a model would make such predictions.

In consequence, these methods can also be classified into two categories named as perceptive interpretability and mathematical interpretability. When visual evidence is not useful or erroneous, the mathematical evidence can be used as the complement for interpretability. Therefore, various methods should be applied simultaneously for the sake of providing reliable interpretability.

Nevertheless, a significant drawback of the current studies on XAI is that the interpretations are focused on the intuition of experts rather than from the demands of the end-users [188]. Current local explanations are typically provided in a feature-importance vector format, which is a full causal attribution and a low-level interpretation. This format would be satisfactory if the description viewers were the developers and analysts, since they could use the mathematical study of the distribution of features to debug the models. However, this type of XAI is less accommodating if the description receivers are lay-users of the AI. XAI can explain the complete judgement logic of the model, which includes a large amount of repetitive knowledge which can confuse the lay-users. The presentation of the XAI algorithms should be further improved to increase customer satisfaction.

The poor abstraction level of explanations is another drawback. For example, despite XAI derived heatmaps can indicate that individual pixels are important, there is normally no correlation computed between these significance regions to more abstract principles such as the anatomical or pathological regions shown in the images. More importantly, the explanations ought to be understood by humans to make sense of them and to grasp the understandable actions of the model. It is indeed desirable to provide meta-explanations that can integrate evidence from these low-level heatmaps to describe the behaviour of the model at a more abstract, more humanly understandable level. However, this level of understanding can be hard and erroneous. Previously proposed methods have recently been suggested to aggregate low-level explanations and measure the semantics of neural representations. Thus, a constructive topic for future study is the development of more advanced meta-explanations that leverages multimodal information fusion.

Because the audiences of XAI results are essentially human users, an important future research direction is the use of XAI in human–machine interaction; therefore, research studies in XAI need to explore human factors. A prerequisite for good human–machine interaction is to construct explanations for the right user focus, for instance, develop XAI to ask the correct questions in the proper manner, which is crucial in the clinical environment. Optimisation of the reasoning procedure for optimal human use, however, is still a problem that demands more research. Eventually, a broad open gap in XAI is the use of interpretabilities beyond using visualisation techniques. Future studies will demonstrate how to incorporate XAI into a broader optimisation mechanism in order to, e.g., boost the efficiency of the model and reduce the model complexity.

5. Conclusion

The recent confluence of large-scale annotated clinical databases, the innovation of deep learning approaches, open-source software packages, and inexpensive and rapidly increasing computing capacity and cloud storage has fuelled the recent exponential growth in AI. This foretells to change the landscape of medical practice in the near future. AI systems have specialised success in certain clinical activities that are more able to assess patient prognosis compared to doctors, and can help in surgical procedures. If deep learning models continue to advance, there is a growing chance that AI could revolutionise medical practice and redefine the role of clinicians in the process. Our mini-review has demonstrated the research trends towards the trustable AI or trustworthy AI, which promotes the XAI globally, and XAI methods in medicine and digital healthcare are highly in demand. Additionally, our two showcases have shown promising XAI results for the two most widely investigated classification and segmentation problems in medical image analysis. We can envisage further development of XAI in medicine and digital healthcare by integrating information fusion from cross-modalities imaging and non-imaging clinical data can be a stepping stone toward a more general acceptance of AI in clinical practice. Ultimately, the trustable AI will promote confidence and openness of its deployment in the clinical arena and also make it easier to comply with the legislation of the GDPR and regulations of the NHSX in the UK, CE-mark in the EU, FDA in the USA, and NMPA in China.

CRediT authorship contribution statement

Guang Yang: Conceived and designed the study, Data analysis, Data interpretation, Writing of the report, Literature search. Qinghao Ye: Conceived and designed the study, Data analysis, Data interpretation, Writing of the report, Data curation, Contributed to the tables and figures. Jun Xia: Conceived and designed the study, Data analysis, Data interpretation, Writing of the report, Data collection.

Declaration of Competing Interest

One or more of the authors of this paper have disclosed potential or pertinent conflicts of interest, which may include receipt of payment, either direct or indirect, institutional support, or association with an entity in the biomedical field which may be perceived to have potential conflict of interest with this work. For full disclosure statements refer to https://doi.org/10.1016/j.inffus.2021.07.016.QY is employed by Hangzhou Ocean’s Smart Boya Co., Ltd., China. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Acknowledgements

This work was supported in part by the Hangzhou Economic and Technological Development Area Strategical Grant [Imperial Institute of Advanced Technology], in part by the Project of Shenzhen International Cooperation Foundation (GJHZ20180926165402083), in part by the Clinical Research Project of Shenzhen Health and Family Planning Commission (SZLY2018018), in part by the European Research Council Innovative Medicines Initiative on Development of Therapeutics and Diagnostics Combatting Coronavirus Infections Award ‘DRAGON: rapiD and secuRe AI imaging based diaGnosis, stratification, fOllow-up, and preparedness for coronavirus paNdemics’ [H2020-JTI-IMI2 101005122], in part by the AI for Health Imaging Award ‘CHAIMELEON: Accelerating the Lab to Market Transition of AI Tools for Cancer Management’ [H2020-SC1-FA-DTS-2019-1 952172], in part by the British Heart Foundation [TG/18/5/34111, PG/16/78/32402], and in part by the UK Research and Innovation [MR/V023799/1]. All authors approved the submitted version.

Contributor Information

Guang Yang, Email: g.yang@imperial.ac.uk.

Qinghao Ye, Email: q7ye@ucsd.edu.

Jun Xia, Email: xiajun@email.szu.edu.cn.

References

  • 1.Arrieta A.B., Díaz-Rodríguez N., Del Ser J., Bennetot A., Tabik S., Barbado A., García S., Gil-López S., Molina D., Benjamins R. Explainable artificial intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI. Inf. Fusion. 2020;58:82–115. [Google Scholar]
  • 2.Adadi A., Berrada M. Peeking inside the black-box: A survey on explainable artificial intelligence (XAI) IEEE Access. 2018;6:52138–52160. [Google Scholar]
  • 3.Samek W., Müller K.-R. Explainable AI: Interpreting, Explaining and Visualizing Deep Learning. Springer; 2019. Towards explainable artificial intelligence; pp. 5–22. [Google Scholar]
  • 4.Rai A. Explainable AI: From black box to glass box. J. Acad. Mark. Sci. 2020;48(1):137–141. [Google Scholar]
  • 5.M.T. Ribeiro, S. Singh, C. Guestrin, “Why should I trust you?” Explaining the predictions of any classifier, in: Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, 2016, pp. 1135–1144.
  • 6.Jiang F., Jiang Y., Zhi H., Dong Y., Li H., Ma S., Wang Y., Dong Q., Shen H., Wang Y. Artificial intelligence in healthcare: past, present and future. Stroke Vasc. Neurol. 2017;2(4):230–243. doi: 10.1136/svn-2017-000101. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7.Panch T., Mattie H., Celi L.A. The “inconvenient truth” about AI in healthcare. NPJ Digit. Med. 2019;2(1):1–3. doi: 10.1038/s41746-019-0155-4. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.Yu K.-H., Beam A.L., Kohane I.S. Artificial intelligence in healthcare. Nat. Biomed. Eng. 2018;2(10):719–731. doi: 10.1038/s41551-018-0305-z. [DOI] [PubMed] [Google Scholar]
  • 9.Shen D., Wu G., Suk H.-I. Deep learning in medical image analysis. Annu. Rev. Biomed. Eng. 2017;19:221–248. doi: 10.1146/annurev-bioeng-071516-044442. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10.Litjens G., Kooi T., Bejnordi B.E., Setio A.A.A., Ciompi F., Ghafoorian M., Van Der Laak J.A., Van Ginneken B., Sánchez C.I. A survey on deep learning in medical image analysis. Med. Image Anal. 2017;42:60–88. doi: 10.1016/j.media.2017.07.005. [DOI] [PubMed] [Google Scholar]
  • 11.Ker J., Wang L., Rao J., Lim T. Deep learning applications in medical image analysis. IEEE Access. 2017;6:9375–9389. [Google Scholar]
  • 12.Dilsizian S.E., Siegel E.L. Artificial intelligence in medicine and cardiac imaging: harnessing big data and advanced computing to provide personalized medical diagnosis and treatment. Curr. Cardiol. Rep. 2014;16(1):441. doi: 10.1007/s11886-013-0441-8. [DOI] [PubMed] [Google Scholar]
  • 13.Patel V.L., Shortliffe E.H., Stefanelli M., Szolovits P., Berthold M.R., Bellazzi R., Abu-Hanna A. The coming of age of artificial intelligence in medicine. Artif. Intell. Med. 2009;46(1):5–17. doi: 10.1016/j.artmed.2008.07.017. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14.Jha S., Topol E.J. Adapting to artificial intelligence: radiologists and pathologists as information specialists. JAMA. 2016;316(22):2353–2354. doi: 10.1001/jama.2016.17438. [DOI] [PubMed] [Google Scholar]
  • 15.Strickland E. IBM Watson, heal thyself: How IBM overpromised and underdelivered on ai health care. IEEE Spectr. 2019;56(4):24–31. [Google Scholar]
  • 16.Weingart N.S., Wilson R.M., Gibberd R.W., Harrison B. Epidemiology of medical error. Bmj. 2000;320(7237):774–777. doi: 10.1136/bmj.320.7237.774. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17.Graber M.L., Franklin N., Gordon R. Diagnostic error in internal medicine. Arch. Intern. Med. 2005;165(13):1493–1499. doi: 10.1001/archinte.165.13.1493. [DOI] [PubMed] [Google Scholar]
  • 18.Winters B., Custer J., Galvagno S.M., Colantuoni E., Kapoor S.G., Lee H., Goode V., Robinson K., Nakhasi A., Pronovost P. Diagnostic errors in the intensive care unit: a systematic review of autopsy studies. BMJ Qual. Saf. 2012;21(11):894–902. doi: 10.1136/bmjqs-2012-000803. [DOI] [PubMed] [Google Scholar]
  • 19.Lee C.S., Nagy P.G., Weaver S.J., Newman-Toker D.E. Cognitive and system factors contributing to diagnostic errors in radiology. Am. J. Roentgenol. 2013;201(3):611–617. doi: 10.2214/AJR.12.10375. [DOI] [PubMed] [Google Scholar]
  • 20.Neill D.B. Using artificial intelligence to improve hospital inpatient care. IEEE Intell. Syst. 2013;28(2):92–95. [Google Scholar]
  • 21.Miller R.A. Medical diagnostic decision support systems—past, present, and future: a threaded bibliography and brief commentary. J. Am. Med. Inform. Assoc. 1994;1(1):8–27. doi: 10.1136/jamia.1994.95236141. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22.Musen M.A., Middleton B., Greenes R.A. Biomedical Informatics. Springer; 2014. Clinical decision-support systems; pp. 643–674. [Google Scholar]
  • 23.Kundu M., Nasipuri M., Basu D.K. Knowledge-based ECG interpretation: a critical review. Pattern Recognit. 2000;33(3):351–373. [Google Scholar]
  • 24.De Dombal F., Leaper D., Staniland J.R., McCann A., Horrocks J.C. Computer-aided diagnosis of acute abdominal pain. Br. Med. J. 1972;2(5804):9–13. doi: 10.1136/bmj.2.5804.9. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25.Shortliffe E.H., Davis R., Axline S.G., Buchanan B.G., Green C.C., Cohen S.N. Computer-based consultations in clinical therapeutics: explanation and rule acquisition capabilities of the MYCIN system. Comput. Biomed. Res. 1975;8(4):303–320. doi: 10.1016/0010-4809(75)90009-9. [DOI] [PubMed] [Google Scholar]
  • 26.Barnett G.O., Cimino J.J., Hupp J.A., Hoffer E.P. DXplain: an evolving diagnostic decision-support system. JAMA. 1987;258(1):67–74. doi: 10.1001/jama.258.1.67. [DOI] [PubMed] [Google Scholar]
  • 27.Miller R.A., McNeil M.A., Challinor S.M., Masarie Jr F.E., Myers J.D. The internist-1/quick medical reference project—Status report. West. J. Med. 1986;145(6):816. [PMC free article] [PubMed] [Google Scholar]
  • 28.Berner E.S., Webster G.D., Shugerman A.A., Jackson J.R., Algina J., Baker A.L., Ball E.V., Cobbs C.G., Dennis V.W., Frenkel E.P. Performance of four computer-based diagnostic systems. N. Engl. J. Med. 1994;330(25):1792–1796. doi: 10.1056/NEJM199406233302506. [DOI] [PubMed] [Google Scholar]
  • 29.Szolovits P., Pauker S.G. Categorical and probabilistic reasoning in medical diagnosis. Artificial Intelligence. 1978;11(1–2):115–144. [Google Scholar]
  • 30.Szolovits P., Pauker S.G. Artificial Intelligence in Perspective. MIT Press; Cambridge, MA: 1994. Categorical and probabilistic reasoning in medicine revisited. [Google Scholar]
  • 31.Deo R.C. Machine learning in medicine. Circulation. 2015;132(20):1920–1930. doi: 10.1161/CIRCULATIONAHA.115.001593. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 32.Yu K.-H., Snyder M. Omics profiling in precision oncology. Mol. Cell. Proteom. 2016;15(8):2525–2536. doi: 10.1074/mcp.O116.059253. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 33.Roberts K., Boland M.R., Pruinelli L., Dcruz J., Berry A., Georgsson M., Hazen R., Sarmiento R.F., Backonja U., Yu K.-H. Biomedical informatics advancing the national health agenda: the AMIA 2015 year-in-review in clinical and consumer informatics. J. Am. Med. Inform. Assoc. 2017;24(e1):e185–e190. doi: 10.1093/jamia/ocw103. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 34.Rogers W., Thulasi Seetha S., Refaee T.A., Lieverse R.I., Granzier R.W., Ibrahim A., Keek S.A., Sanduleanu S., Primakov S.P., Beuque M.P. Radiomics: from qualitative to quantitative imaging. Br. J. Radiol. 2020;93(1108) doi: 10.1259/bjr.20190948. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 35.Goodfellow I., Bengio Y., Courville A., Bengio Y. MIT press Cambridge; 2016. Deep Learning, vol. 1. [Google Scholar]
  • 36.Wang Y.E., Wei G.-Y., Brooks D. 2019. Benchmarking TPU, GPU, and CPU platforms for deep learning. arXiv Preprint arXiv:1907.10701. [Google Scholar]
  • 37.Tomczak K., Czerwińska P., Wiznerowicz M. The cancer genome atlas (TCGA): an immeasurable source of knowledge. Contemp. Oncol. 2015;19(1A):A68. doi: 10.5114/wo.2014.47136. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 38.Sudlow C., Gallacher J., Allen N., Beral V., Burton P., Danesh J., Downey P., Elliott P., Green J., Landray M. UK Biobank: an open access resource for identifying the causes of a wide range of complex diseases of middle and old age. Plos Med. 2015;12(3) doi: 10.1371/journal.pmed.1001779. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 39.Ljosa V., Sokolnicki K.L., Carpenter A.E. Annotated high-throughput microscopy image sets for validation. Nature Methods. 2012;9(7):637. doi: 10.1038/nmeth.2083. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 40.Williams E., Moore J., Li S.W., Rustici G., Tarkowska A., Chessel A., Leo S., Antal B., Ferguson R.K., Sarkans U. Image data resource: a bioimage data integration and publication platform. Nature Methods. 2017;14(8):775–781. doi: 10.1038/nmeth.4326. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 41.DesRoches C.M., Campbell E.G., Rao S.R., Donelan K., Ferris T.G., Jha A., Kaushal R., Levy D.E., Rosenbaum S., Shields A.E. Electronic health records in ambulatory care—a national survey of physicians. N. Engl. J. Med. 2008;359(1):50–60. doi: 10.1056/NEJMsa0802005. [DOI] [PubMed] [Google Scholar]
  • 42.Hsiao C.-J., Jha A.K., King J., Patel V., Furukawa M.F., Mostashari F. Office-based physicians are responding to incentives and assistance by adopting and using electronic health records. Health Aff. 2013;32(8):1470–1477. doi: 10.1377/hlthaff.2013.0323. [DOI] [PubMed] [Google Scholar]
  • 43.Hu S., Gao Y., Niu Z., Jiang Y., Li L., Xiao X., Wang M., Fang E.F., Menpes-Smith W., Xia J. Weakly supervised deep learning for COVID-19 infection detection and classification from CT images. IEEE Access. 2020 [Google Scholar]
  • 44.Litjens G., Sánchez C.I., Timofeeva N., Hermsen M., Nagtegaal I., Kovacs I., Hulsbergen-Van De Kaa C., Bult P., Van Ginneken B., Van Der Laak J. Deep learning as a tool for increased accuracy and efficiency of histopathological diagnosis. Sci. Rep. 2016;6:26286. doi: 10.1038/srep26286. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 45.Zhang N., Yang G., Gao Z., Xu C., Zhang Y., Shi R., Keegan J., Xu L., Zhang H., Fan Z. Deep learning for diagnosis of chronic myocardial infarction on nonenhanced cardiac cine MRI. Radiology. 2019;291(3):606–617. doi: 10.1148/radiol.2019182304. [DOI] [PubMed] [Google Scholar]
  • 46.Cao Y., Wang Z., Liu Z., Li Y., Xiao X., Sun L., Zhang Y., Hou H., Zhang P., Yang G. Multiparameter synchronous measurement with IVUS images for intelligently diagnosing coronary cardiac disease. IEEE Trans. Instrum. Meas. 2020 [Google Scholar]
  • 47.Cheerla A., Gevaert O. Deep learning with multimodal representation for pancancer prognosis prediction. Bioinformatics. 2019;35(14):i446–i454. doi: 10.1093/bioinformatics/btz342. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 48.Roberts M., Driggs D., Thorpe M., Gilbey J., Yeung M., Ursprung S., Aviles-Rivero A.I., Etmann C., McCague C., Beer L. 2020. Machine learning for COVID-19 detection and prognostication using chest radiographs and CT scans: a systematic methodological review. arXiv Preprint arXiv:2008.06388. [Google Scholar]
  • 49.Zou J., Huss M., Abid A., Mohammadi P., Torkamani A., Telenti A. A primer on deep learning in genomics. Nature Genet. 2019;51(1):12–18. doi: 10.1038/s41588-018-0295-5. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 50.Waldstein S.M., Seeböck P., Donner R., Sadeghipour A., Bogunović H., Osborne A., Schmidt-Erfurth U. Unbiased identification of novel subclinical imaging biomarkers using unsupervised deep learning. Sci. Rep. 2020;10(1):1–9. doi: 10.1038/s41598-020-69814-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 51.Li L., Wu F., Yang G., Xu L., Wong T., Mohiaddin R., Firmin D., Keegan J., Zhuang X. Atrial scar quantification via multi-scale CNN in the graph-cuts framework. Med. Image Anal. 2020;60 doi: 10.1016/j.media.2019.101595. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 52.N.D. Lane, S. Bhattacharya, P. Georgiev, C. Forlivesi, F. Kawsar, An early resource characterization of deep learning on wearables, smartphones and internet-of-things devices, in: Proceedings of the 2015 International Workshop on Internet of Things Towards Applications, 2015, pp. 7–12.
  • 53.Chen A.I., Balter M.L., Maguire T.J., Yarmush M.L. Deep learning robotic guidance for autonomous vascular access. Nat. Mach. Intell. 2020;2(2):104–115. [Google Scholar]
  • 54.Esteva A., Robicquet A., Ramsundar B., Kuleshov V., DePristo M., Chou K., Cui C., Corrado G., Thrun S., Dean J. A guide to deep learning in healthcare. Nature Med. 2019;25(1):24–29. doi: 10.1038/s41591-018-0316-z. [DOI] [PubMed] [Google Scholar]
  • 55.Obermeyer Z., Emanuel E.J. Predicting the future—big data, machine learning, and clinical medicine. N. Engl. J. Med. 2016;375(13):1216. doi: 10.1056/NEJMp1606181. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 56.Krause J., Gulshan V., Rahimy E., Karth P., Widner K., Corrado G.S., Peng L., Webster D.R. Grader variability and the importance of reference standards for evaluating machine learning models for diabetic retinopathy. Ophthalmology. 2018;125(8):1264–1272. doi: 10.1016/j.ophtha.2018.01.034. [DOI] [PubMed] [Google Scholar]
  • 57.D. Rebholz-Schuhmann, A.J. Jimeno-Yepes, E. van Mulligen, N. Kang, J. Kors, D. Milward, P. Corbett, E. Buyko, K. Tomanek, E. Beisswanger, et al. The calbc silver standard corpus for biomedical named entities-a study in harmonizing the contributions from four independent named entity taggers, in: Proceedings of the Seventh International Conference on Language Resources and Evaluation, 2010.
  • 58.Kirby J.C., Speltz P., Rasmussen L.V., Basford M., Gottesman O., Peissig P.L., Pacheco J.A., Tromp G., Pathak J., Carrell D.S. PheKB: a Catalog and workflow for creating electronic phenotype algorithms for transportability. J. Am. Med. Inform. Assoc. 2016;23(6):1046–1052. doi: 10.1093/jamia/ocv202. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 59.Zhao Z.-Q., Zheng P., Xu S.-t., Wu X. Object detection with deep learning: A review. IEEE Trans. Neural Netw. Learn. Syst. 2019;30(11):3212–3232. doi: 10.1109/TNNLS.2018.2876865. [DOI] [PubMed] [Google Scholar]
  • 60.Singh S.P., Kumar A., Darbari H., Singh L., Rastogi A., Jain S. 2017 International Conference on Computer, Communications and Electronics (Comptelix) IEEE; 2017. Machine translation using deep learning: An overview; pp. 162–167. [Google Scholar]
  • 61.D. Amodei, S. Ananthanarayanan, R. Anubhai, J. Bai, E. Battenberg, C. Case, J. Casper, B. Catanzaro, Q. Cheng, G. Chen, et al. Deep speech 2: End-to-end speech recognition in English and Mandarin, in: International Conference on Machine Learning, 2016, pp. 173–182.
  • 62.Nassif A.B., Shahin I., Attili I., Azzeh M., Shaalan K. Speech recognition using deep neural networks: A systematic review. IEEE Access. 2019;7:19143–19165. [Google Scholar]
  • 63.Purwins H., Li B., Virtanen T., Schlüter J., Chang S.-Y., Sainath T. Deep learning for audio signal processing. IEEE J. Sel. Top. Sign. Proces. 2019;13(2):206–219. [Google Scholar]
  • 64.Elsken T., Metzen J.H., Hutter F. Neural architecture search: A survey. J. Mach. Learn. Res. 2019;20(55):1–21. [Google Scholar]
  • 65.Holzinger A., Malle B., Saranti A., Pfeifer B. Towards multi-modal causability with graph neural networks enabling information fusion for explainable AI. Inf. Fusion. 2021;71:28–37. [Google Scholar]
  • 66.Zhang W., Liu F., Luo L., Zhang J. Predicting drug side effects by multi-label learning and ensemble learning. BMC Bioinformatics. 2015;16(1):365. doi: 10.1186/s12859-015-0774-y. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 67.Yang G., Raschke F., Barrick T.R., Howe F.A. Manifold learning in MR spectroscopy using nonlinear dimensionality reduction and unsupervised clustering. Magn. Reson. Med. 2015;74(3):868–878. doi: 10.1002/mrm.25447. [DOI] [PubMed] [Google Scholar]
  • 68.Zhao L.P., Bolouri H. Object-oriented regression for building predictive models with high dimensional omics data from translational studies. J. Biomed. Inform. 2016;60:431–445. doi: 10.1016/j.jbi.2016.03.001. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 69.Kim S.G., Theera-Ampornpunt N., Fang C.-H., Harwani M., Grama A., Chaterji S. Opening up the blackbox: an interpretable deep neural network-based classifier for cell-type specific enhancer predictions. BMC Syst. Biol. 2016;10(2):243–258. doi: 10.1186/s12918-016-0302-3. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 70.Hao J., Kim Y., Kim T.-K., Kang M. PASNet: pathway-associated sparse deep neural network for prognosis prediction from high-throughput data. BMC Bioinformatics. 2018;19(1):1–13. doi: 10.1186/s12859-018-2500-z. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 71.Bernardini M., Romeo L., Misericordia P., Frontoni E. Discovering the type 2 diabetes in electronic health records using the sparse balanced support vector machine. IEEE J. Biomed. Health Inf. 2019;24(1):235–246. doi: 10.1109/JBHI.2019.2899218. [DOI] [PubMed] [Google Scholar]
  • 72.Eck A., Zintgraf L.M., de Groot E., de Meij T.G., Cohen T.S., Savelkoul P., Welling M., Budding A. Interpretation of microbiota-based diagnostics by explaining individual classifier decisions. BMC Bioinformatics. 2017;18(1):441. doi: 10.1186/s12859-017-1843-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 73.Ge W., Huh J.-W., Park Y.R., Lee J.-H., Kim Y.-H., Turchin A. AMIA Annual Symposium Proceedings. Vol. 2018. American Medical Informatics Association; 2018. An interpretable ICU mortality prediction model based on logistic regression and recurrent neural networks with LSTM units; p. 460. [PMC free article] [PubMed] [Google Scholar]
  • 74.Zuallaert J., Godin F., Kim M., Soete A., Saeys Y., De Neve W. Splicerover: interpretable convolutional neural networks for improved splice site prediction. Bioinformatics. 2018;34(24):4180–4188. doi: 10.1093/bioinformatics/bty497. [DOI] [PubMed] [Google Scholar]
  • 75.Suh J., Yoo S., Park J., Cho S.Y., Cho M.C., Son H., Jeong H. Development and validation of explainable AI-based decision-supporting tool for prostate biopsy. BJU Int. 2020 doi: 10.1111/bju.15122. [DOI] [PubMed] [Google Scholar]
  • 76.Singh A., Mohammed A.R., Zelek J., Lakshminarayanan V. Applications of Machine Learning 2020, vol. 11511. International Society for Optics and Photonics; 2020. Interpretation of deep learning using attributions: application to ophthalmic diagnosis; p. 115110A. [Google Scholar]
  • 77.Kwon B.C., Choi M.-J., Kim J.T., Choi E., Kim Y.B., Kwon S., Sun J., Choo J. Retainvis: Visual analytics with interpretable and interactive recurrent neural networks on electronic medical records. IEEE Trans. Vis. Comput. Graphics. 2018;25(1):299–309. doi: 10.1109/TVCG.2018.2865027. [DOI] [PubMed] [Google Scholar]
  • 78.Zhang J., Kowsari K., Harrison J.H., Lobo J.M., Barnes L.E. Patient2vec: A personalized interpretable deep representation of the longitudinal electronic health record. IEEE Access. 2018;6:65333–65346. [Google Scholar]
  • 79.Choi E., Bahadori M.T., Sun J., Kulas J., Schuetz A., Stewart W. Retain: An interpretable predictive model for healthcare using reverse time attention mechanism. Adv. Neural Inf. Process. Syst. 2016;29:3504–3512. [Google Scholar]
  • 80.Kaji D.A., Zech J.R., Kim J.S., Cho S.K., Dangayach N.S., Costa A.B., Oermann E.K. An attention based deep learning model of clinical events in the intensive care unit. PLoS One. 2019;14(2) doi: 10.1371/journal.pone.0211057. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 81.Shickel B., Loftus T.J., Adhikari L., Ozrazgat-Baslanti T., Bihorac A., Rashidi P. DeepSOFA: a continuous acuity score for critically ill patients using clinically interpretable deep learning. Sci. Rep. 2019;9(1):1–12. doi: 10.1038/s41598-019-38491-0. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 82.Hu H., Xiao A., Zhang S., Li Y., Shi X., Jiang T., Zhang L., Zhang L., Zeng J. DeepHINT: understanding HIV-1 integration via deep learning with attention. Bioinformatics. 2019;35(10):1660–1667. doi: 10.1093/bioinformatics/bty842. [DOI] [PubMed] [Google Scholar]
  • 83.Izadyyazdanabadi M., Belykh E., Cavallo C., Zhao X., Gandhi S., Moreira L.B., Eschbacher J., Nakaji P., Preul M.C., Yang Y. International Conference on Medical Image Computing and Computer-Assisted Intervention. Springer; 2018. Weakly-supervised learning-based feature localization for confocal laser endomicroscopy glioma images; pp. 300–308. [Google Scholar]
  • 84.Zhao G., Zhou B., Wang K., Jiang R., Xu M. International Conference on Medical Image Computing and Computer-Assisted Intervention. Springer; 2018. Respond-CAM: Analyzing deep models for 3D imaging data by visualizations; pp. 485–492. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 85.Couture H.D., Marron J.S., Perou C.M., Troester M.A., Niethammer M. International Conference on Medical Image Computing and Computer-Assisted Intervention. Springer; 2018. Multiple instance learning for heterogeneous images: Training a CNN for histopathology; pp. 254–262. [Google Scholar]
  • 86.Lee H., Yune S., Mansouri M., Kim M., Tajmir S.H., Guerrier C.E., Ebert S.A., Pomerantz S.R., Romero J.M., Kamalian S. An explainable deep-learning algorithm for the detection of acute intracranial haemorrhage from small datasets. Nat. Biomed. Eng. 2019;3(3):173. doi: 10.1038/s41551-018-0324-9. [DOI] [PubMed] [Google Scholar]
  • 87.Kim J., Kim H.J., Kim C., Kim W.H. Artificial intelligence in breast ultrasonography. Ultrasonography (Seoul, Korea) 2020 doi: 10.14366/usg.20117. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 88.Rajpurkar P., Park A., Irvin J., Chute C., Bereket M., Mastrodicasa D., Langlotz C.P., Lungren M.P., Ng A.Y., Patel B.N. AppendiXNet: Deep learning for diagnosis of appendicitis from a small dataset of CT exams using video pretraining. Sci. Rep. 2020;10(1):1–7. doi: 10.1038/s41598-020-61055-6. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 89.Porumb M., Stranges S., Pescapè A., Pecchia L. Precision medicine and artificial intelligence: A pilot study on deep learning for hypoglycemic events detection based on ECG. Sci. Rep. 2020;10(1):1–16. doi: 10.1038/s41598-019-56927-5. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 90.R. Caruana, Y. Lou, J. Gehrke, P. Koch, M. Sturm, N. Elhadad, Intelligible models for healthcare: Predicting pneumonia risk and hospital 30-day readmission, in: Proceedings of the 21th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, 2015, pp. 1721–1730.
  • 91.Letham B., Rudin C., McCormick T.H., Madigan D. Interpretable classifiers using rules and Bayesian analysis: Building a better stroke prediction model. Ann. Appl. Stat. 2015;9(3):1350–1371. [Google Scholar]
  • 92.Che Z., Purushotham S., Khemani R., Liu Y. AMIA Annual Symposium Proceedings, vol. 2016. American Medical Informatics Association; 2016. Interpretable deep models for ICU outcome prediction; p. 371. [PMC free article] [PubMed] [Google Scholar]
  • 93.Ming Y., Qu H., Bertini E. Rulematrix: Visualizing and understanding classifiers with rules. IEEE Trans. Vis. Comput. Graphics. 2018;25(1):342–352. doi: 10.1109/TVCG.2018.2864812. [DOI] [PubMed] [Google Scholar]
  • 94.Xiao C., Ma T., Dieng A.B., Blei D.M., Wang F. Readmission prediction via deep contextual embedding of clinical concepts. PLoS One. 2018;13(4) doi: 10.1371/journal.pone.0195024. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 95.Davoodi R., Moradi M.H. Mortality prediction in intensive care units (ICUs) using a deep rule-based fuzzy classifier. J. Biomed. Inform. 2018;79:48–59. doi: 10.1016/j.jbi.2018.02.008. [DOI] [PubMed] [Google Scholar]
  • 96.Lee H., Kim S.T., Ro Y.M. Interpretability of Machine Intelligence in Medical Image Computing and Multimodal Learning for Clinical Decision Support. Springer; 2019. Generation of multimodal justification using visual word constraint model for explainable computer-aided diagnosis; pp. 21–29. [Google Scholar]
  • 97.Prentzas N., Nicolaides A., Kyriacou E., Kakas A., Pattichis C. 2019 IEEE 19th International Conference on Bioinformatics and Bioengineering (BIBE) IEEE; 2019. Integrating machine learning with symbolic reasoning to build an explainable AI model for stroke prediction; pp. 817–821. [Google Scholar]
  • 98.Pan L., Liu G., Mao X., Li H., Zhang J., Liang H., Li X. Development of prediction models using machine learning algorithms for girls with suspected central precocious puberty: retrospective study. JMIR Med. Inform. 2019;7(1) doi: 10.2196/11728. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 99.Ghafouri-Fard S., Taheri M., Omrani M.D., Daaee A., Mohammad-Rahimi H., Kazazi H. Application of single-nucleotide polymorphisms in the diagnosis of autism spectrum disorders: a preliminary study with artificial neural networks. J. Mol. Neurosci. 2019;68(4):515–521. doi: 10.1007/s12031-019-01311-1. [DOI] [PubMed] [Google Scholar]
  • 100.Kovalev M.S., Utkin L.V., Kasimov E.M. Survlime: A method for explaining machine learning survival models. Knowl.-Based Syst. 2020;203 doi: 10.1016/j.knosys.2020.106164. http://www.sciencedirect.com/science/article/pii/S0950705120304044. [DOI] [Google Scholar]
  • 101.Meldo A., Utkin L., Kovalev M., Kasimov E. The natural language explanation algorithms for the lung cancer computer-aided diagnosis system. Artif. Intell. Med. 2020;108 doi: 10.1016/j.artmed.2020.101952. [DOI] [PubMed] [Google Scholar]
  • 102.Panigutti C., Perotti A., Pedreschi D. Proceedings of the 2020 Conference on Fairness, Accountability, and Transparency. Association for Computing Machinery; New York, NY, USA: 2020. Doctor XAI: An ontology-based approach to black-box sequential data classification explanations; pp. 629–639. (FAT* ’20). [DOI] [Google Scholar]
  • 103.Lauritsen S.M., Kristensen M., Olsen M.V., Larsen M.S., Lauritsen K.M., Jø rgensen M.J., Lange J., Thiesson B. Explainable artificial intelligence model to predict acute critical illness from electronic health records. Nature Commun. 2020;11(1):1–11. doi: 10.1038/s41467-020-17431-x. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 104.Lee L.M., Gostin L.O. Ethical collection, storage, and use of public health data: a proposal for a national privacy protection. JAMA. 2009;302(1):82–84. doi: 10.1001/jama.2009.958. [DOI] [PubMed] [Google Scholar]
  • 105.S. Narayan, M. Gagné, R. Safavi-Naini, Privacy preserving EHR system using attribute-based infrastructure, in: Proceedings of the 2010 ACM Workshop on Cloud Computing Security Workshop, 2010, pp. 47–52.
  • 106.Dolin R.H., Alschuler L., Boyer S., Beebe C., Behlen F.M., Biron P.V., Shabo A. HL7 clinical document architecture, release 2. J. Am. Med. Inform. Assoc. 2006;13(1):30–39. doi: 10.1197/jamia.M1888. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 107.Mandl K.D., Kohane I.S. Escaping the EHR trap—the future of health IT. N. Engl. J. Med. 2012;366(24):2240–2242. doi: 10.1056/NEJMp1203102. [DOI] [PubMed] [Google Scholar]
  • 108.Mandel J.C., Kreda D.A., Mandl K.D., Kohane I.S., Ramoni R.B. SMART On FHIR: a standards-based, interoperable apps platform for electronic health records. J. Am. Med. Inform. Assoc. 2016;23(5):899–908. doi: 10.1093/jamia/ocv189. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 109.Topol E.J. High-performance medicine: the convergence of human and artificial intelligence. Nature Med. 2019;25(1):44–56. doi: 10.1038/s41591-018-0300-7. [DOI] [PubMed] [Google Scholar]
  • 110.Yu K.-H., Kohane I.S. Framing the challenges of artificial intelligence in medicine. BMJ Qual. Saf. 2019;28(3):238–241. doi: 10.1136/bmjqs-2018-008551. [DOI] [PubMed] [Google Scholar]
  • 111.Miller D.D. The medical AI insurgency: what physicians must know about data to practice with intelligent machines. NPJ Digit. Med. 2019;2(1):1–5. doi: 10.1038/s41746-019-0138-5. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 112.Challen R., Denny J., Pitt M., Gompels L., Edwards T., Tsaneva-Atanasova K. Artificial intelligence, bias and clinical safety. BMJ Qual. Saf. 2019;28(3):231–237. doi: 10.1136/bmjqs-2018-008370. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 113.DeCamp M., Lindvall C. Latent bias and the implementation of artificial intelligence in medicine. J. Am. Med. Inform. Assoc. 2020;27(12):2020–2023. doi: 10.1093/jamia/ocaa094. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 114.Esmaeilzadeh P. Use of AI-based tools for healthcare purposes: a survey study from consumers’ perspectives. BMC Med. Inform. Decis. Mak. 2020;20(1):1–19. doi: 10.1186/s12911-020-01191-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 115.England J.R., Cheng P.M. Artificial intelligence for medical image analysis: a guide for authors and reviewers. Am. J. Roentgenol. 2019;212(3):513–519. doi: 10.2214/AJR.18.20490. [DOI] [PubMed] [Google Scholar]
  • 116.Gomolin A., Netchiporouk E., Gniadecki R., Litvinov I.V. Artificial intelligence applications in dermatology: Where do we stand? Front. Med. 2020;7 doi: 10.3389/fmed.2020.00100. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 117.Zeiler M.D., Fergus R. European Conference on Computer Vision. Springer; 2014. Visualizing and understanding convolutional networks; pp. 818–833. [Google Scholar]
  • 118.K. Simonyan, A. Vedaldi, A. Zisserman, Deep inside convolutional networks: Visualising image classification models and saliency maps, in: Workshop At International Conference on Learning Representations, 2014.
  • 119.R.R. Selvaraju, M. Cogswell, A. Das, R. Vedantam, D. Parikh, D. Batra, Grad-cam: Visual explanations from deep networks via gradient-based localization, in: Proceedings of the IEEE International Conference on Computer Vision, 2017, pp. 618–626.
  • 120.Z. Zhang, Y. Xie, F. Xing, M. McGough, L. Yang, Mdnet: A semantically and visually interpretable medical image diagnosis network, in: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2017, pp. 6428–6436.
  • 121.Croskerry P., Cosby K., Graber M.L., Singh H. CRC Press; 2017. Diagnosis: Interpreting the Shadows. [Google Scholar]
  • 122.Quinn T.P., Jacobs S., Senadeera M., Le V., Coghlan S. 2020. The three ghosts of medical AI: Can the black-box present deliver? arXiv preprint arXiv:2012.06000. [DOI] [PubMed] [Google Scholar]
  • 123.Holzinger A., Biemann C., Pattichis C.S., Kell D.B. 2017. What do we need to build explainable AI systems for the medical domain? arXiv preprint arXiv:1712.09923. [Google Scholar]
  • 124.Tonekaboni S., Joshi S., McCradden M.D., Goldenberg A. In: Proceedings of the 4th Machine Learning for Healthcare Conference. Doshi-Velez F., Fackler J., Jung K., Kale D., Ranganath R., Wallace B., Wiens J., editors. vol. 106. PMLR; Ann Arbor, Michigan: 2019. What clinicians want: Contextualizing explainable machine learning for clinical end use; pp. 359–380. (Proceedings of Machine Learning Research). http://proceedings.mlr.press/v106/tonekaboni19a.html. [Google Scholar]
  • 125.Holzinger A., Langs G., Denk H., Zatloukal K., Müller H. Causability and explainability of artificial intelligence in medicine. Wiley Interdiscip. Rev. Data Min. Knowl. Discov. 2019;9(4) doi: 10.1002/widm.1312. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 126.Hossain M.S., Muhammad G., Guizani N. Explainable AI and mass surveillance system-based healthcare framework to combat COVID-I9 like pandemics. IEEE Netw. 2020;34(4):126–132. [Google Scholar]
  • 127.Khodabandehloo E., Riboni D., Alimohammadi A. HealthXAI: Collaborative and explainable AI for supporting early diagnosis of cognitive decline. Future Gener. Comput. Syst. 2021;116:168–189. [Google Scholar]
  • 128.Kallianos K., Mongan J., Antani S., Henry T., Taylor A., Abuya J., Kohli M. How far have we come? Artificial intelligence for chest radiograph interpretation. Clin. Radiol. 2019;74(5):338–345. doi: 10.1016/j.crad.2018.12.015. [DOI] [PubMed] [Google Scholar]
  • 129.Zucco C., Liang H., Di Fatta G., Cannataro M. 2018 IEEE International Conference on Bioinformatics and Biomedicine (BIBM) IEEE; 2018. Explainable sentiment analysis with applications in medicine; pp. 1740–1747. [Google Scholar]
  • 130.Langlotz C.P., Allen B., Erickson B.J., Kalpathy-Cramer J., Bigelow K., Cook T.S., Flanders A.E., Lungren M.P., Mendelson D.S., Rudie J.D. A roadmap for foundational research on artificial intelligence in medical imaging: from the 2018 NIH/RSNA/ACR/The academy workshop. Radiology. 2019;291(3):781–791. doi: 10.1148/radiol.2019190613. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 131.London A.J. Artificial intelligence and black-box medical decisions: accuracy versus explainability. Hastings Cent. Rep. 2019;49(1):15–21. doi: 10.1002/hast.973. [DOI] [PubMed] [Google Scholar]
  • 132.Stiglic G., Kocbek S., Pernek I., Kokol P. Comprehensive decision tree models in bioinformatics. PLoS One. 2012;7(3) doi: 10.1371/journal.pone.0033812. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 133.Valdes G., Luna J.M., Eaton E., Simone II C.B., Ungar L.H., Solberg T.D. Mediboost: a patient stratification tool for interpretable decision making in the era of precision medicine. Sci. Rep. 2016;6:37854. doi: 10.1038/srep37854. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 134.Tjoa E., Guan C. A survey on explainable artificial intelligence (XAI): Toward medical XAI. IEEE Trans. Neural Netw. Learn. Syst. 2020 doi: 10.1109/TNNLS.2020.3027314. [DOI] [PubMed] [Google Scholar]
  • 135.Payrovnaziri S.N., Chen Z., Rengifo-Moreno P., Miller T., Bian J., Chen J.H., Liu X., He Z. Explainable artificial intelligence models using real-world electronic health record data: a systematic scoping review. J. Am. Med. Inform. Assoc. 2020 doi: 10.1093/jamia/ocaa053. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 136.Pocevičiūtė M., Eilertsen G., Lundström C. Artificial Intelligence and Machine Learning for Digital Pathology. Springer; 2020. Survey of XAI in digital pathology; pp. 56–88. [Google Scholar]
  • 137.Tosun A.B., Pullara F., Becich M.J., Taylor D.L., Chennubhotla S.C., Fine J.L. Artificial Intelligence and Machine Learning for Digital Pathology. Springer; 2020. Histomapr™: An explainable AI (xAI) platform for computational pathology solutions; pp. 204–227. [Google Scholar]
  • 138.Wold S., Esbensen K., Geladi P. Principal component analysis. Chemometr. Intell. Lab. Syst. 1987;2(1–3):37–52. [Google Scholar]
  • 139.Comon P. Independent component analysis, a new concept? Signal Process. 1994;36(3):287–314. [Google Scholar]
  • 140.Belkin M., Niyogi P. Laplacian eigenmaps and spectral techniques for embedding and clustering. Adv. Neural Inf. Process. Syst. 2001;14:585–591. [Google Scholar]
  • 141.Carvalho D.V., Pereira E.M., Cardoso J.S. Machine learning interpretability: A survey on methods and metrics. Electronics. 2019;8(8):832. [Google Scholar]
  • 142.Linardatos P., Papastefanopoulos V., Kotsiantis S. Explainable AI: A review of machine learning interpretability methods. Entropy. 2021;23(1):18. doi: 10.3390/e23010018. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 143.Lundberg S.M., Erion G., Chen H., DeGrave A., Prutkin J.M., Nair B., Katz R., Himmelfarb J., Bansal N., Lee S.-I. From local explanations to global understanding with explainable AI for trees. Nat. Mach. Intell. 2020;2(1):2522–5839. doi: 10.1038/s42256-019-0138-9. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 144.Montavon G., Samek W., Müller K.-R. Methods for interpreting and understanding deep neural networks. Digit. Signal Process. 2018;73:1–15. [Google Scholar]
  • 145.Shrikumar A., Greenside P., Kundaje A. In: Proceedings of the 34th International Conference on Machine Learning. Precup D., Teh Y.W., editors. vol. 70. 2017. Learning important features through propagating activation differences; pp. 3145–3153. (Proceedings of Machine Learning Research). [Google Scholar]
  • 146.Springenberg J., Dosovitskiy A., Brox T., Riedmiller M. ICLR (Workshop Track) 2015. Striving for simplicity: the all convolutional net. http://lmb.informatik.uni-freiburg.de/Publications/2015/DB15a. [Google Scholar]
  • 147.Bach S., Binder A., Montavon G., Klauschen F., Müller K.-R., Samek W. On pixel-wise explanations for non-linear classifier decisions by layer-wise relevance propagation. PLoS One. 2015;10(7) doi: 10.1371/journal.pone.0130140. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 148.Chen H., Lundberg S., Lee S.-I. Explainable AI in Healthcare and Medicine: Building a Culture of Transparency and Accountability. Springer International Publishing; Cham: 2021. Explaining models by propagating Shapley values of local components; pp. 261–270. [Google Scholar]
  • 149.Bahdanau D., Cho K., Bengio Y. In: 3rd International Conference on Learning Representations, ICLR 2015, San Diego, CA, USA, May 7-9, 2015, Conference Track Proceedings. Bengio Y., LeCun Y., editors. 2015. Neural machine translation by jointly learning to align and translate. http://arxiv.org/abs/1409.0473. [Google Scholar]
  • 150.Putelli L., Gerevini A.E., Lavelli A., Serina I. International Conference of the Italian Association for Artificial Intelligence. Springer; 2019. Applying self-interaction attention for extracting drug-drug interactions; pp. 445–460. [Google Scholar]
  • 151.D. Mascharka, P. Tran, R. Soklaski, A. Majumdar, Transparency by design: Closing the gap between performance and interpretability in visual reasoning, in: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2018, pp. 4942–4950.
  • 152.B. Zhou, A. Khosla, A. Lapedriza, A. Oliva, A. Torralba, Learning deep features for discriminative localization, in: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2016, pp. 2921–2929.
  • 153.Polino A., Pascanu R., Alistarh D. International Conference on Learning Representations. 2018. Model compression via distillation and quantization. https://openreview.net/forum?id=S1XolQbRW. [Google Scholar]
  • 154.M. Wu, M. Hughes, S. Parbhoo, M. Zazzi, V. Roth, F. Doshi-Velez, Beyond sparsity: Tree regularization of deep models for interpretability, in: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 32, 2018.
  • 155.Hinton G., Vinyals O., Dean J. NIPS Deep Learning and Representation Learning Workshop. 2015. Distilling the knowledge in a neural network. http://arxiv.org/abs/1503.02531. [Google Scholar]
  • 156.Hinton G., Frosst N. 2017. Distilling a neural network into a soft decision tree. https://arxiv.org/pdf/1711.09784.pdf. [Google Scholar]
  • 157.Yang X., Zhang H., Cai J. Auto-encoding and distilling scene graphs for image captioning. IEEE Trans. Pattern Anal. Mach. Intell. 2020 doi: 10.1109/TPAMI.2020.3042192. [DOI] [PubMed] [Google Scholar]
  • 158.Li X.-H., Cao C.C., Shi Y., Bai W., Gao H., Qiu L., Wang C., Gao Y., Zhang S., Xue X. A survey of data-driven and knowledge-aware explainable AI. IEEE Trans. Knowl. Data Eng. 2020 [Google Scholar]
  • 159.Lage I., Chen E., He J., Narayanan M., Kim B., Gershman S., Doshi-Velez F. 2019. An evaluation of the human-interpretability of explanation. arXiv preprint arXiv:1902.00006. [PMC free article] [PubMed] [Google Scholar]
  • 160.H. Lakkaraju, E. Kamar, R. Caruana, J. Leskovec, Faithful and customizable explanations of black box models, in: Proceedings of the 2019 AAAI/ACM Conference on AI, Ethics, and Society, 2019, pp. 131–138.
  • 161.Das D., Ito J., Kadowaki T., Tsuda K. An interpretable machine learning model for diagnosis of Alzheimer’s disease. PeerJ. 2019;7 doi: 10.7717/peerj.6543. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 162.Liang Y., Li S., Yan C., Li M., Jiang C. Explaining the black-box model: A survey of local interpretation methods for deep neural networks. Neurocomputing. 2021;419:168–182. [Google Scholar]
  • 163.Montavon G., Binder A., Lapuschkin S., Samek W., Müller K.-R. Explainable AI: Interpreting, Explaining and Visualizing Deep Learning. Springer; 2019. Layer-wise relevance propagation: an overview; pp. 193–209. [Google Scholar]
  • 164.Lin Y., Dong S., Yeh Y., Wu Y., Lan G., Liu C., Chu T.C. Emergency management and infection control in a radiology department during an outbreak of severe acute respiratory syndrome. Br. J. Radiol. 2005;78(931):606–611. doi: 10.1259/bjr/17161223. [DOI] [PubMed] [Google Scholar]
  • 165.K. He, X. Zhang, S. Ren, J. Sun, Deep residual learning for image recognition, in: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2016, pp. 770–778.
  • 166.Zhou Z.-H. Department of Computer Science & Technology, Nanjing University; 2004. Multi-instance learning: A survey. [Google Scholar]
  • 167.Bekker A.J., Goldberger J. 2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) IEEE; 2016. Training deep neural-networks based on unreliable labels; pp. 2682–2686. [Google Scholar]
  • 168.Qian X., Lin Y., Zhao Y., Yue X., Lu B., Wang J. Objective ventricle segmentation in brain CT with ischemic stroke based on anatomical knowledge. BioMed Res. Int. 2017;2017 doi: 10.1155/2017/8690892. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 169.Cherukuri V., Ssenyonga P., Warf B.C., Kulkarni A.V., Monga V., Schiff S.J. Learning based segmentation of CT brain images: application to postoperative hydrocephalic scans. IEEE Trans. Biomed. Eng. 2017;65(8):1871–1884. doi: 10.1109/TBME.2017.2783305. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 170.Ronneberger O., Fischer P., Brox T. International Conference on Medical Image Computing and Computer-Assisted Intervention. Springer; 2015. U-net: Convolutional networks for biomedical image segmentation; pp. 234–241. [Google Scholar]
  • 171.Deng J., Dong W., Socher R., Li L.-J., Li K., Fei-Fei L. 2009 IEEE Conference on Computer Vision and Pattern Recognition. IEEE; 2009. Imagenet: A large-scale hierarchical image database; pp. 248–255. [Google Scholar]
  • 172.Dong C., Loy C.C., He K., Tang X. Image super-resolution using deep convolutional networks. IEEE Trans. Pattern Anal. Mach. Intell. 2015;38(2):295–307. doi: 10.1109/TPAMI.2015.2439281. [DOI] [PubMed] [Google Scholar]
  • 173.Grandvalet Y., Bengio Y. Advances in Neural Information Processing Systems. 2005. Semi-supervised learning by entropy minimization; pp. 529–536. [Google Scholar]
  • 174.T.-H. Vu, H. Jain, M. Bucher, M. Cord, P. Pérez, Advent: Adversarial entropy minimization for domain adaptation in semantic segmentation, in: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2019, pp. 2517–2526.
  • 175.Kingma D.P., Ba J. 2014. Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980. [Google Scholar]
  • 176.Zhang K., Liu X., Shen J., Li Z., Sang Y., Wu X., Zha Y., Liang W., Wang C., Wang K. Clinically applicable AI system for accurate diagnosis, quantitative measurements, and prognosis of COVID-19 pneumonia using computed tomography. Cell. 2020 doi: 10.1016/j.cell.2020.08.029. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 177.Wang L., Wong A. 2020. COVID-Net: A tailored deep convolutional neural network design for detection of COVID-19 cases from chest radiography images. arXiv preprint arXiv:2003.09871. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 178.Li L., Qin L., Xu Z., Yin Y., Wang X., Kong B., Bai J., Lu Y., Fang Z., Song Q. Artificial intelligence distinguishes COVID-19 from community acquired pneumonia on chest CT. Radiology. 2020 doi: 10.1148/radiol.2020200905. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 179.Ouyang X., Huo J., Xia L., Shan F., Liu J., Mo Z., Yan F., Ding Z., Yang Q., Song B. 2020. Dual-sampling attention network for diagnosis of COVID-19 from community acquired pneumonia. arXiv preprint arXiv:2005.02690. [DOI] [PubMed] [Google Scholar]
  • 180.Geirhos R., Rubisch P., Michaelis C., Bethge M., Wichmann F.A., Brendel W. 2018. Imagenet-trained CNNs are biased towards texture; increasing shape bias improves accuracy and robustness. arXiv preprint arXiv:1811.12231. [Google Scholar]
  • 181.Carrington A.M., Manuel D.G., Fieguth P.W., Ramsay T., Osmani V., Wernly B., Bennett C., Hawken S., McInnes M., Magwood O. 2021. Deep ROC analysis and AUC as balanced average accuracy to improve model selection, understanding and interpretation. arXiv preprint arXiv:2103.11357. [DOI] [PubMed] [Google Scholar]
  • 182.Maaten L.v.d., Hinton G. Visualizing data using t-SNE. J. Mach. Learn. Res. 2008;9(Nov):2579–2605. [Google Scholar]
  • 183.Lundberg S.M., Lee S.-I. Advances in Neural Information Processing Systems. 2017. A unified approach to interpreting model predictions; pp. 4765–4774. [Google Scholar]
  • 184.Zhou Z., Siddiquee M.M.R., Tajbakhsh N., Liang J. Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support. Springer; 2018. UNet++: A nested U-Net architecture for medical image segmentation; pp. 3–11. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 185.Long C., Xu H., Shen Q., Zhang X., Fan B., Wang C., Zeng B., Li Z., Li X., Li H. Diagnosis of the coronavirus disease (COVID-19): rRT-PCR or CT? Eur. J. Radiol. 2020;126 doi: 10.1016/j.ejrad.2020.108961. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 186.Roberts M., Driggs D., Thorpe M., Gilbey J., Yeung M., Ursprung S., Aviles-Rivero A.I., Etmann C., McCague C., Beer L. Common pitfalls and recommendations for using machine learning to detect and prognosticate for COVID-19 using chest radiographs and CT scans. Nat. Mach. Intell. 2021;3(3):199–217. [Google Scholar]
  • 187.Driggs D., Selby I., Roberts M., Gkrania-Klotsas E., Rudd J.H., Yang G., Babar J., Sala E., Schönlieb C.-B., collaboration A.-C. Radiological Society of North America; 2021. Machine Learning for COVID-19 Diagnosis and Prognostication: Lessons for Amplifying the Signal While Reducing the Noise. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 188.Du M., Liu N., Hu X. Techniques for interpretable machine learning. Commun. ACM. 2019;63(1):68–77. [Google Scholar]

RESOURCES