Skip to main content
Frontiers in Robotics and AI logoLink to Frontiers in Robotics and AI
. 2024 Nov 28;11:1444763. doi: 10.3389/frobt.2024.1444763

Enhancing interpretability and accuracy of AI models in healthcare: a comprehensive review on challenges and future directions

Mohammad Ennab 1,*,, Hamid Mcheick 1,
PMCID: PMC11638409  PMID: 39677978

Abstract

Artificial Intelligence (AI) has demonstrated exceptional performance in automating critical healthcare tasks, such as diagnostic imaging analysis and predictive modeling, often surpassing human capabilities. The integration of AI in healthcare promises substantial improvements in patient outcomes, including faster diagnosis and personalized treatment plans. However, AI models frequently lack interpretability, leading to significant challenges concerning their performance and generalizability across diverse patient populations. These opaque AI technologies raise serious patient safety concerns, as non-interpretable models can result in improper treatment decisions due to misinterpretations by healthcare providers. Our systematic review explores various AI applications in healthcare, focusing on the critical assessment of model interpretability and accuracy. We identify and elucidate the most significant limitations of current AI systems, such as the black-box nature of deep learning models and the variability in performance across different clinical settings. By addressing these challenges, our objective is to provide healthcare providers with well-informed strategies to develop innovative and safe AI solutions. This review aims to ensure that future AI implementations in healthcare not only enhance performance but also maintain transparency and patient safety.

Keywords: artificial intelligence, machine learning, deep learning, healthcare, interpretability, explainability, accuracy

1 Introduction

In recent years, AI technology has enabled the widespread adoption of machine learning (ML) and deep learning (DL) methods in healthcare. These techniques have automated various processes like screening, diagnostics, and treatment (Zhongqi and Jia, 2022). However, DL algorithms are often considered “black boxes” due to their intricate training and numerous parameters, making it hard to decipher the connection between inputs and outputs (Zhang and Zhang, 2023). Although advanced DL methods can significantly improve speed and accuracy without human intervention, establishing an interpretable AI framework remains imperative. Understanding algorithms’ inner workings and prediction bases is essential, not just competitively but also for stakeholder trust (Stinson and Vlaad, 2024). Discussing trustworthy AI involves the terms accuracy, interpretability, and explainability, which are occasionally used interchangeably. In data mining and machine learning, interpretability is the interface between people and decision models that accurately represents the model and is intelligible to humans (Savindu Herath Pathirannehelage et al., 2024).

2 Research gaps

While current literature highlights the use of AI in healthcare, the research gaps in terms of model generalizability, lack of real-time interpretability, and inclusion of diverse clinical data remain relatively unexplored. Additionally, the integration of uncertainty quantification with interpretability models is an underexamined area that could improve AI adoption in healthcare settings. Gaps related to real-time feedback for clinicians and the role of user-centered design in AI development also need further exploration (Helman et al., 2022).

This paper characterizes the challenges of AI applications in healthcare focused on interpretability and accuracy to ensure accountability and regulatory compliance. Our survey helps healthcare providers develop appropriate strategies to rapidly implement innovative solutions safely. The goal is consolidating existing knowledge on AI systems and the interpretability models to help researchers swiftly grasp the state-of-the-art and determine areas needing more research (Saeed and Omlin, 2023). We also distinguish our review from earlier studies by focusing on both the technical and practical limitations of AI in healthcare, which are often overlooked in previous reviews, such as those focusing solely on specific AI techniques or healthcare applications.

The primary contributions of this work include:

  • • Summarizing the current state of AI systems in healthcare.

  • • Identifying and detailing the key limitations encountered when developing AI systems, with a focus on both interpretability and accuracy.

  • • Discussing the research gaps and future challenges in AI application in healthcare.

  • • Offering a detailed comparison of how this review differs from previous literature, including works like (Sadeghi et al., 2024).

Additionally, we consider the control strategy presented by Stefanelli et al. (2023) in their work on sensorimotor control in prosthetics, which integrates both force and temperature information to replicate human reflexive behavior during manipulation. This approach provides valuable insights into designing more human-like AI systems in healthcare by considering multimodal sensory inputs.

3 Methodology

To achieve our objectives, we systematically reviewed 61 AI systems for classifying and treating various diseases in healthcare. The selected systems were analyzed by extracting how they address the challenges of assessing AI systems in terms of interpretability and accuracy. Our work defines the black box of healthcare AI applications, emphasizing model interpretability and accuracy.

3.1 Search strategy

We conducted our search using multiple databases, including Google Scholar, PubMed, and IEEE Xplore, to ensure comprehensive coverage of the literature. The search was conducted using the following keywords, combined with Boolean operators:

  • • “Artificial Intelligence” AND “Healthcare” AND “Accuracy”

  • • “Deep Learning” AND “Interpretability” AND “Healthcare”

  • • “Machine Learning” AND “Explainability” AND “Healthcare”

3.2 Inclusion criteria

  • • Peer-reviewed articles focused on healthcare AI applications.

  • • Studies published between 2010 and 2024.

  • • Articles emphasizing interpretability and/or accuracy (both ML and DL).

3.3 Exclusion criteria

  • • Articles not focused on AI techniques (i.e., non technical).

  • • Non-healthcare AI applications.

  • • Legislative or legal discussions that do not contribute to the technical understanding of interpretability.

3.4 Data extraction

For each selected study, we extracted information on:

  • • The AI model used.

  • • The metrics employed to assess interpretability and accuracy.

  • • The context in which the AI model was applied.

  • • A detailed discussion on the results achieved.

The results of the literature analysis are summarized in Table 1.

TABLE 1.

Summary of selected models that applied AI models in various healthcare contexts.

Study AI model Accuracy metric Interpretability Context Results
Zhongqi and Jia (2022) Deep Learning 95% Black-box Diagnostic Imaging High accuracy but limited interpretability
Jing et al. (2023) Neural Networks 92% Explainable AI (XAI) Predictive Modeling Improved performance with post hoc explanation methods
Stinson and Vlaad (2024) Deep Neural Networks 97% None Screening and Diagnostics Excellent accuracy but no real-time interpretability
Helman et al. (2022) Random Forests 89% Global Interpretability Treatment Decision Support High interpretability but moderately lower accuracy
Mella et al. (2023) Support Vector Machines 91% Local Interpretability Diagnostics High accuracy with interpretable decision boundaries
Hamm et al. (2023) Deep Learning 95% Available Prostate cancer High accuracy
Subramani et al. (2023) Deep Learning 96% Black-box Cardiovascular diseases High accuracy
Nicolae et al. (2020) Machine Learning Treatment Dose Planning Black-box Rectal cancer Effective planning
Nicolae et al. (2020) Machine Learning Effectively reduce planning time Black-box Prostate cancer Efficient planning
Afrash et al. (2023) Deep Learning 83% Black-box Gastric cancer High accuracy
Subramani et al. (2023) Deep Learning 96% Black-box Cardiovascular diseases High accuracy

The PRISMA diagram shown in Figure 1 illustrates the article selection process:

FIGURE 1.

FIGURE 1

PRISMA diagram illustrates the article selection process.

3.5 Study limitations

Our review primarily focuses on AI models applied in diagnostic imaging tasks. While we have considered various healthcare applications (e.g., screening, and treatment decision-making), our scope does not extend to other domains like genomics or personalized medicine, which could benefit from AI innovations. Additionally, the lack of real-time applicability and the high computational cost of certain models have not been deeply explored in this review.

4 Summary of selected literature

We reviewed 61 papers that applied AI models in various healthcare contexts. Table 1 summarizes the key information extracted from each paper, including the AI model used, the metrics for accuracy and interpretability, and the results achieved.

5 Key metrics in AI model development

Despite the numerous benefits, AI models in healthcare are not without limitations. These include:

5.1 The accuracy

Accuracy is paramount in healthcare applications, especially for tasks like diagnostic imaging and predictive modeling. AI systems can often surpass human experts in detecting abnormalities (Esteva et al., 2017). However, the performance of AI models is typically evaluated using historical data, split into training and testing sets. This method, though widely used, does not always reflect real-world performance (Baeza-Yates, 2022). The trade-off between model complexity and accuracy is an ongoing challenge.

5.2 The interpretability

Interpretability in AI refers to the degree to which a human can understand and trust the decisions made by AI systems. In the context of healthcare, interpretability is paramount because medical professionals rely on these systems to make critical decisions about patient care. Without a clear understanding of how AI systems arrive at their conclusions, healthcare providers may be reluctant to adopt these technologies or may misinterpret their recommendations, potentially leading to adverse patient outcomes (Ennab and Mcheick, 2022b). Techniques like activation visualization enable us to infer which features in an image are represented by each feature map within every layer of the trained convolutional neural network. When activation visualization is needed, it involves inputting sample medical images into the convolutional neural network for analysis and observing the patterns in the results (Kim, 2023).

5.3 The interpretability-accuracy trade-off

As highlighted (Ennab and Mcheick, 2022a), there is often a trade-off between interpretability and accuracy. While simpler models like decision trees are more interpretable, they may not achieve the same level of accuracy as more complex models, such as deep neural networks (Escalante et al., 2018).

6 AI models in healthcare

The field of medicine has employed AI technology to automate various stages of clinical research, offering valuable assistance for clinical decision-making (Wu Yazhou and Chen, 2022). Utilizing AI techniques across diverse medical domains brings advantages such as enhancing diagnostic accuracy and reducing both time and labor requirements. Leveraging the latest AI advancements, typical applications experiencing revolutionary changes include intelligent screening, precise diagnosis, risk prediction, and supportive therapy (Khalifa et al., 2024) as shown in Figure 2.

FIGURE 2.

FIGURE 2

A diagram illustrating AI models in healthcare.

6.1 Conversational AI

Conversational AI is a technology that leverages conversation-oriented techniques, facilitating dynamic interactions and widespread engagements across various users and platforms (Kaushik, 2023). It refers to a set of technologies that communicate with humans using “natural language” through hardware, software, etc., by combining natural language processing (NLP), ML, DL, and conversation processing technologies into a single platform (Han et al., 2023). Within the healthcare sector, conversational AI is facilitating several significant use cases that have an impact on both healthcare providers and patients. In healthcare, five distinct use cases have been recognized:

  • (1) Disseminating Information: AI assistants can engage in conversational interactions to provide answers to frequently asked questions (FAQs) concerning specific illnesses, health conditions, or pandemics (Haptik, 2023). It has the potential to increase awareness of a specific health-related issue or disaster by providing quick access to accurate, trustworthy, and timely information, e.g., Dr. LalPathLabs (Gottlieb et al., 2023).

  • (2) Scheduling appointments: Conversational AI enables patients to easily arrange and reschedule medical appointments (Haptik, 2023).

  • (3) Patients care: Healthcare organizations have effectively used AI Assistants to automate the answer to frequently asked questions and the resolution of regular, repetitive chores and diagnostics chains (Haptik, 2023).

  • (4) Managing and tracking patients’ health: conversational AI aids individuals in monitoring their physical wellbeing by detecting symptoms at an early stage and facilitating online consultations with healthcare professionals when necessary (Haptik, 2023). Additionally, patients are equipped with the resources and information required to manage their own health effectively (Han et al., 2023).

  • (5) Enhancing Patient Assistance: conversational AI fosters a data-driven approach within healthcare, empowering patients and caregivers with the necessary information for informed decision-making (Haptik, 2023).

6.2 Intelligent screening

AI technology has been applied in the screening of various malignant tumors, and it can automatically screen the beginning and malignant areas of suspected cancer. However, it should be noted that the accuracy of the model has a major impact on the clinical decision-making of physicians (Chen et al., 2024). When the model’s predictions are inaccurate, the effectiveness of its supplementary screening is frequently significantly diminished. Moreover, for diseases with low occurrence rates and limited sample sizes, the presence of false positives cannot be disregarded, making it advisable to conduct manual reviews for verification. Consequently, considerable challenges persist in the application of AI models in clinical settings, and it is essential to factor in the potential adverse repercussions of model-assisted screening during the design of AI tools. Furthermore, it is noteworthy that all AI systems operated as black boxes without real-time interpretability techniques (Wu Yazhou and Chen, 2022).

6.3 Screening for digestive tumors

Kiani et al., (2020) developed a deep learning system for liver pathology image analysis; hence, the system can be self-employed in the screening of both hepatocellular carcinoma and cholangiocarcinoma. When tested against the validation set, this system has a performance accuracy of 88.5% and boasts the same accuracy of 88.5% for the independent test set. A similar system was developed by Sinha et al., (2024), which carried out the analysis of colonoscopy images with the help of machine learning. Precisely, the system detects and differentiates the adenomas, which need to be extracted, from nonneoplastic polyps, while having a predictive accuracy of 98.1%. A similar artificial intelligence (AI)-based deep learning system was developed by Wang et al., (2020) for colonoscopy picture analysis; the study showed that the Adenoma Detection Rate (ADR) increased tremendously after AI application compared with that in the traditional group. Therefore, it can be said that with this system, the polyp and tumor detection rate of colonoscopy have been well increased. On the other hand, Chen et al. (2020) and Wu L (2019) have developed a deep learning-based image analysis system for esophagogastroduodenoscopy (EGD), which can classify the stage of duodenal diseases. This system leads to an excellent reduction of the blind spot’s missed diagnosis rate, which is 5.9% and 3.4% missed rate.

6.4 Screening for other tumors

In a study by McGenity et al. (2024), an innovative deep learning-driven system for analyzing slice pathological images was introduced. This system demonstrated the capability to automatically diagnose and classify breast cancer. Impressively, the overall accuracy rate achieved a notable 83.1% when compared to pathological results, which are considered the gold standard. Whereas, Hossain et al. (2024) presented a chest CT processing system based on deep learning, achieving a sensitivity rate of 91% for metastasis detection, thereby enabling the automated screening of metastatic breast cancer. Mohamed et al. (2022) introduced a fully automated breast cancer detection system. Initially, the U-Net network is employed to automatically extract and segregate the breast area from the surrounding body, which is considered noise in the breast cancer detection model. This yielded an accuracy of 99.33%, sensitivity of 100%, and specificity of 98.67%. Lotter et al. (2021) introduced an annotation-efficient DL approach that attains state-of-the-art performance in fields like mammogram classification, showcasing average sensitivity enhancements of up to 14% for AI methods compared to mammography experts. Yoo et al. (2018) presented a deep learning-based ultrasound image analysis system, which elevated the screening sensitivity for thyroid cancer from 84% to 92.0%, and successfully achieved automated screening for both benign and malignant thyroid nodules. Masood et al. (2018) Constructed a lung CT image processing system using the Internet of Medical Things (IoMT) and DL, which successfully predicted the malignant transformation stage of pulmonary nodules, achieving a classification accuracy rate of 84.6%. Wu Yazhou and Chen (2022) and Zadeh Shirazi et al. (2021) employed a Deep Convolutional Neural Network (DCNN) as a semantic segmentation model for segmenting seven distinct tumor regions, achieving a segmentation accuracy of 70%.

6.5 Screening for eye diseases

Hassan et al. (2024) utilized DL techniques to process retinal images, enabling the automated screening and severity grading of diabetic retinopathy. The AI method exhibited a sensitivity and specificity of 100% and 88.4%, respectively, for diagnosing severe lesions. With the ability to obtain the sensitivity of 85.2% and specificity of 92% for lesions. Archana and Jeevaraj (2024) developed a DL-based system for processing the images of the eye, and it provided an outstanding AUC-ROC from 99.3% to 99.7% for cataract classification. This aids in conducting automated screening and collective management of cataract-affected persons. Lombardi et al. (2021) proposed a new deep learning system to predict brain age through native resting-state scalp EEG raw data and reached an accuracy of 89.7%. Wu et al. (2018) found in a study on the diagnosis of fungal keratitis, that automatic hyphae detection technology is sensitive at 89.3% and specific at 95.7%. The AUC value of the ROC curve is 94.6%, which is timely, accurate, objective, and quantitated to provide evaluation criteria for fungal keratitis.

6.6 Intelligent diagnosis

AI systems within healthcare provide patients with reliable and tailored services that go beyond traditional testing. They effectively assist patients in overseeing their individual wellbeing, tracking medical conditions, delivering treatments, offering psychological guidance, and handling dietary considerations. This holistic support contributes to extending patient longevity and enhancing their overall quality of life (Origimid, 2022). However, the effectiveness of the AI model is limited by the size of the training dataset. A model trained on one dataset may not perform optimally when applied to a different dataset with different characteristics. Therefore, it is important to incorporate an external test dataset during model training to evaluate its ability to generalize. Additionally, while many AI-driven diagnostic methods focus on analyzing medical imaging data, it is essential to recognize that clinically meaningful research outcomes require a holistic assessment of various patient indicators by healthcare professionals. Therefore, future research efforts should prioritize the comprehensive use of diverse clinical data to improve the effectiveness and wide-ranging applicability of AI models (Data sharing in the age of deep learning, 2023).

6.7 Diagnosis of infectious diseases

The outbreak of novel coronavirus disease (COVID-19) provides a suitable application opportunity for AI technology. AI technology has made a significant progress in the diagnosis, typing, risk prediction, and adjuvant treatment of COVID-19 (Ennab and Mcheick, 2022b). Shorfuzzaman et al. (2021) introduced a deep learning fusion framework that incorporates the principles of transfer learning, achieving intelligent diagnosis of COVID-19 patients with a classification accuracy of 95.5%. Quiroz et al. (2021) confirmed that the ML approach can be employed to automatically assess the severity of COVID-19, aiding in the classification and diagnosis of COVID-19 patients with an accuracy rate of 96.0%, a sensitivity of 84.5%, and a specificity of 92.9%. This enables the prioritization of follow-up diagnosis and treatment. Salem et al. (2023) developed a machine and deep learning system to identify metabolites and clinical features linked to gallstone disease. Pang et al. (2019) developed an innovative YOLOv3-arch model specifically designed to identify cholelithiasis and classify gallstones in CT images. This model significantly enhanced accuracy, achieving a 92.7% accuracy rate for identifying granular gallstones and an average accuracy of 80.3% for identifying muddy gallstones. However, it is important to note a major limitation of these models: healthcare workers cannot predict COVID-19 in patients with other underlying conditions such as COPD, hypertension, asthma, and more (Ennab and Mcheick, 2022b).

6.8 Diagnosis of medical diseases

As medical imaging technology advances and clinical diagnosis accuracy improves, clinical diagnostic approaches employing DL technology have experienced significant growth (Ennab and Mcheick, 2022b). Arbabshirani et al. (2018) and Kim et al. (2024), each constructed a DL-based brain CT image processing system, with accuracy values of (73%) and (84.6%), realizing automatic detection of acute neurological events such as stroke. Lo et al. (2021) suggested an automated ischemic stroke diagnosis method using deep DL, achieving a sensitivity of 98.1%, specificity of 96.9%, and an accuracy rating of 99.3%. This method effectively offers clinicians valuable diagnostic recommendations for acute ischemic stroke cases. Bibi et al. (2020) created a system that combines DL and the IoMT to achieve swift and secure identification and classification of leukemia, boasting an impressive average accuracy rate of 99.6%. This system facilitates real-time communication between healthcare professionals and patients regarding leukemia detection, diagnosis, and treatment, effectively conserving clinicians’ time and efforts. Jing et al. (2023) developed a DL system aimed at predicting the likelihood of recurrence and disability outcomes in patients who have experienced a transient ischemic attack or ischemic stroke. Eckardt et al. (2022) created a deep learning model specifically tailored for detecting NPM1 mutation and forecasting the presence of acute myeloid leukemia, attaining an impressive accuracy level of 95%. Bukhari et al. (2022) developed a deep learning framework for identifying leukemia cancer in microscopic blood samples by incorporating squeeze and excitation learning techniques, achieving an accuracy rate of 89.72%. Hamedan et al. (2020) employed Expert Systems (ES) to assess chronic kidney disease, integrating AI technology with artificial expert judgments. The results indicated that the ES significantly outperformed all other models in performance for predicting chronic kidney disease when the accuracy of the ES was 92.1%, sensitivity was 95.4%, and specificity was 88.9%.

6.9 Diagnosis of surgical diseases

Deep learning image knee analytical system was designed for knee injury diagnosing, e.g., injured knee estimation, such as anterior cruciate ligament and meniscus tears by Dratsch et al. (2024). Deep learning models in Siouras et al. (2022) related to MRI-associated knee injury estimation with accuracies ranging from 72.5% to 100%. The processing system based on deep learning for X-ray images was developed by Lindsey et al. (2018) to detect and localize fractures. After they integrated AI technology, the sensitivity in fracture detection of the clinicians improved from 80.8% to 91.5%, and the specificity increased from 87.5% to 93.9%, the CT image analysis system, designed by Fu et al. (2019), intended for automatic diagnosis and visualization of interfemoral fractures by identifying the most probable area of the fracture.

6.10 Adjuvant therapy

The goal of is to aid individuals with early invasive cancer in accurately predicting the personalized survival benefits, allowing them to make informed treatment decisions, without supplanting conventional treatments (Zheng et al., 2024). Adjuvant therapy relies on machine learning models and has the potential to enhance patient survival (Howard et al., 2020). Machine learning techniques can accurately predict patient outcomes under different treatment regimens by modeling the complex interactions between risk factors in a data-driven manner (Alaa et al., 2020).

6.11 Treatment decision support

Radiation therapy plays a pivotal role in the treatment of various tumors (Frisch et al., 2024). Within the treatment process, it is imperative to precisely outline the organ at risk (OAR) to provide guidance for radiotherapy and predict the patient’s prognosis (Hall et al., 2022). Bird et al. (2021) constructed ML model using a multi-center dataset, with the goal of establishing precise and broadly applicable radiotherapy protocols for rectal cancer. Through rigorous validation studies and regulatory approvals, this treatment decision support system mitigates prescription errors and augments existing error warning mechanisms (Yang et al., 2023). Rawson et al. (2021) constructed machine learning-driven systems for antimicrobial prescription decision-making, offering clinical decision support in antibiotic management. The AI-generated prescription recommendations have achieved a level of accuracy similar to that of clinicians. Bozyel et al. (2024) introduced an ML-based prescription recognition system to enable automated early detection and correction of prescription errors in heart disease patients, achieving a clinical effectiveness rate of 85%. Hooshmand et al. (2021) utilized DL techniques to explore potential COVID-19 drug candidates that exhibit minimal side effects and promising efficacy. Ke et al. (2020) employed DL methodologies to pinpoint existing drugs with the potential to combat COVID-19, ultimately discovering over 80 promising candidates for coronavirus treatment.

6.12 AI systems for medical surgical robots

Surgical robot technology stands as the foundational element of Computer Integrated Surgery (CIS). It brings quantification to the complete medical process, spanning from diagnosis and procedural planning to surgery, recovery, and observation (Lee et al., 2024). By integrating robotics and IT technology into the medical domain, it aims to enhance surgical procedures, making them safer and more efficient through rigorous objectification (Reddy et al., 2023). Surgical robots typically fall into two primary categories. This is identified to be the first category where robot systems are developed, which are designed to operate on body cavities such as the abdomen, chest, and lumbar regions. An illustrative example is the da Vinci robot system, which conducts surgery using a master-slave approach relying on 3D imaging (Biswas et al., 2023). The second category pertains to surgical robot systems designed for addressing lesions within the brain, bones, or muscles. These systems create a 3D model of the patient’s lesion using pre-existing medical images. The surgery planning, therefore, becomes a pointer of other factors like the lesion location, lesion condition, image information regarding the lesion, and the position of the patient in relation to the operating table, along with the 3D patient coordinate system realized before surgery (Chiou et al., 2022). Employing surgical robots yields several advantages, including reduced surgery duration, lowered risk of bleeding and infection for patients, and decreased fatigue and hand tremors for medical staff during procedures. Besides, it has benefits to hospitals by the patients having short stay lengths in hospital, therefore, increasing bed turnover, and positive patient outcomes through minimally invasive surgical technologies (Lee et al., 2024). At the moment, IR is used in broad clinical fields, such as orthopedics, biliary tract, throat, and liver surgeries, to mention but a few (Gumbs et al., 2021; Zhang and Zhang, 2023) used IR in spinal surgery in whose application brings out the accuracy in screw positioning and uses a few intraoperative fluoroscopies and has a low incidence in postoperative complications. Xie et al. (2021) used da Vinci surgical system in managing biliary cysts in infants below 1 year and brought out that the use of IR is safe in this clinical application. Arora et al. (2024) performed robotic surgery, in this case, transoral, for the extraction of masses in the throat, which was well exposed visually and that one encountered no major adverse reactions. Serednicki et al. (2023) used IR in liver surgery, and one enjoys the benefits of low blood loss and few adhesions, therefore short stay days in hospital and the duration for postoperative recovery is short.

7 The interpretability models

With the widespread adoption of AI solutions in healthcare, directly understanding models’ working mechanisms and opening the black box has become increasingly important (Pavlidis, 2024). Building confidence in machine learning models is now necessary for fully adopting AI systems. Thus, model interpretability is highly sought-after, especially in high-risk domains requiring dependability like healthcare (Ennab and Mcheick, 2022a). AI models in healthcare may provide contradictory results to medical institutions, potentially having disastrous consequences across cultures and situations. Interpretable approaches can clarify how a model choice is made, allowing tracking each output result and controlling model outputs (Longo et al., 2024). Interpretable AI models enable users to query, comprehend, correct, and even advance the learning system. Overall, interpretable learning models allow end-users like doctors to assess the model before taking any further action. By justifying predictions, interpretable machine and deep learning models give users the option to reject or accept the predictions and recommendations (Salih, 2022).

7.1 Types of the interpretability models

7.1.1 Ante-hoc(Intrinsic)

The fastest way to easily interpret a model is designing it with an interpretable structure initially. A simple model like a decision tree is easy for humans to interpret by looking at its structure. A model with a simple structure was termed Intrinsic (originally equipped) because it already secured interpretability itself, also expressed as having transparency (Cao et al., 2024). The advantage of Intrinsic is explaining “how the model works.” However, due to the trade-off, intrinsic models have low accuracy (Carvalho et al., 2019).

7.1.2 Post-hoc

If the model itself lacks explanatory power, its prediction results must be interpreted post hoc. Most interpretability techniques in machine and deep learning are post hoc. Ideally, the model would have high accuracy and explanatory power, but this is rare in reality (Cao et al., 2024). Complex high-performing models are commonly used, with post hoc analysis done. Post-hoc examples are explained with the viewpoints introduced later (Moradi and Samwald, 2021).

7.1.3 Global

The global technique describes all model predictions based on understanding its logic, or at the module level, scopes describing how one module affects predictions (Carvalho et al., 2019).

7.1.4 Local

Local techniques describe specific decisions or predictions. A range describing a prediction group by bundling predictions is also a local technique. Compared to global methods, local techniques have a smaller scope to explain, making them relatively feasible and inexpensive. Additionally, even if the overall prediction trend is unexplained, one or a few predictions can be described nearly perfectly (Moradi and Samwald, 2021).

7.2 The interpretability models

The categorization of the interpretability models may vary slightly depending on specific implementations and use cases as shown in Table 2.

TABLE 2.

Categorization of the interpretability models.

Interpretability model Public Local Post-hoc (model-agnostic) Ante-hoc (model-intrinsic)
The interpretability based
Bayesian Nonparametric
GAM
MAPLE
Anchors
Perturbation-Based Methods
Attention Based
SHAP
Grad-CAM
Textual Justification
LIME
Testing Concept Activation Vectors
Similar Images

7.2.1 The interpretability-based model using the relative weight of the image features

Ennab and Mcheick (2022a) presented an interpretability-based model using statistics and probability principles to train datasets by determining the relative weights of variables indicating their respective significance in predicting and estimating disease likelihood. The variables are either patient symptoms or characteristics of injured organ regions in medical images. Dividing each variable’s weight by the sum of all weights gives the relative weights. Training the dataset determines infection likelihood. Data were collected as previously described in Mohammad Ennab and Hamid Mcheick (22 October 2022).

7.2.2 LIME (local interpretable model-agnostic explanation)

LIME is a model-agnostic technique explaining which features are most important in a feature space region (Shi et al., 2020). LIME’s core idea is computing a Local Surrogate Model in a region of interest, which is an easily interpretable model like linear or decision tree trained to mimic a complex model’s behavior (Garreau and von Luxburg, 2020). For an explanation, LIME creates new similar data points with slightly altered values. Feeding the perturbed points into the complex model reveals relationships between perturbed features and predictions, captured by the surrogate model (Garreau and von Luxburg, 2020).

7.2.3 SHAP

SHAP uses game theory to measure each attribute’s impact on the prediction process. The Shapley value evenly divides advantages among contributing parties (features) when contributions are unequal (Shapley, 2016). In other words, Shapley values are based on features interacting to influence predictions toward a value. It attempts to evenly distribute contributions across all feature subgroups. Specifically, the Shapley value uniformly distributes the difference between the prediction and average prediction among the instance’s feature values needing explanation (Strumbelj and Kononenko, 2011). SHAP values provide a unique additive feature importance measure satisfying attribution features (local accuracy, missingness, consistency). These features represent intuitive rules for determining the final prediction, translatable to the machine learning problem. However, direct Shapley value computation requires efficient computation, needing to check every permutation combination (Mitchell et al., 2022).

7.2.4 Grad-CAM

Class Activation Mapping (CAM) is a technique widely used in computer vision to create visual representations displaying contributions of different image regions to neural network predictions (Selvaraju et al., 2017). This produces a heatmap-resembling image with each pixel signifying the activation level for a class. Pixel values range from 0 to 1, typically shown as a 0 to 255 grayscale image. Higher scores indicate regions in the original image having a stronger influence on the network’s response or prediction. Overlaying CAM onto the original image enhances visual appeal and informativeness. Unlike CAM requiring model structural changes and retraining, Grad-CAM utilizes pretrained weights to backpropagate gradients to the desired parameter layer (like convolutional) when predicting an image (Schöttl, 2020). This yields a gradient matrix with identical dimensions to the parameter layer’s output feature map. By globally average pooling the gradient matrix across spatial dimensions, a vector with equal length to the feature map’s channel count emerges. This vector contributes to weighing the diverse feature map channels, ultimately creating a heatmap visualization. As Grad-CAM avoids model architectural adjustments and retraining, it offers a more flexible, efficient CAM alternative (Liu et al., 2023). Conventionally, a neural network’s classification module uses a fully connected model processing extracted features, converting them into class probability scores via a softmax layer. The highest scoring class then dictates the ultimate prediction. Grad-CAM takes a different approach by not just discriminating between classes but also pinpointing relevant image regions. This is achieved by exploiting gradients (derivatives) from the final convolutional layer’s feature map (Vinogradova et al., 2020). These gradients serve as a tool to emphasize important areas significantly impacting the eventual prediction.

7.2.5 Bayesian nonparametric model

Guo et al. (2018) created a Bayesian nonparametric approach to build a parameter space with infinite dimensions. In other words, model size can fluctuate in response to data increases or decreases, determined by the number of data parameters used. It requires few assumptions to learn data and perform clustering. Growing data can also be continuously aggregated into proper classifications, this model also makes predictions concurrently. A spatial data model comprises all issue-related properties manageable based on the unique learning problem (Ribeiro et al., 2018). The core of the non-parametric Bayesian statistical model is setting the data probability distribution function to an arbitrary, flexible distribution rather than a specific parametric one and placing a prior distribution on this to perform posterior inference. In Bayesian statistical models, non-parametric modelling allows flexible modelling of data distribution, random effect distribution, or parameter of interest prior distribution, presenting a wider probability model class. Thus, the prior distribution for random distribution is key in nonparametric Bayesian statistical modelling, with the probability most frequently used (Noh et al., 2014).

7.2.6 GAM

Kraus et al. (2024) introduced GAM, a generalized additive global variable weight technique considering neural network swarm forecast patterns. GAM’s global interpretation describes the neural network’s non-linear representation. GAM also enables modifying subpopulation granularity and tracking global interpretations for particular samples. In statistics, a generalized additive model (GAM) is a linear model where predicted variable values are the aggregation of several unknown smooth functions defined for the predictors. The purpose is inferring a predictor smooth function whose aggregate composition approximates it. This structure is easily interpretable, allowing the user to see each variable’s importance, i.e., its effect on the predicted output via its function (Linardatos et al., 2021).

7.2.7 MAPLE

The key difference between using MAPLE as a black-box model explanation versus a predictive model is fitting MAPLE to the black-box model’s prediction in the first case and the response variable in the second. Since MAPLE is a very accurate predictive model providing correct predictions, it avoids trading off performance and interpretability. It finds global trends using local examples and explanations. MAPLE differs from other frameworks by its training distributions (El Shawi et al., 2019). Data were collected as previously described in Mohammad Ennab and Hamid Mcheick (22 October 2022). Maple alters tree ensembles to provide local explanations that detect global trends and example-based explanations. It uses the ensemble to determine the most important training points for a new prediction, building a linear model from those points for prediction and local explanation (Plumb et al., 2018).

7.2.8 Anchors

Anchors is a model-independent, rule-based local explainer approach (Ribeiro et al., 2018). Anchors ensure that projections from the same anchor are roughly equal. Anchors, in other words, identify the features that are sufficient to correct the forecast while adjusting the others that do not influence the prediction. The bottom-up strategy, in which anchors are constructed sequentially, is one type of anchor building (Ribeiro et al., 2018). Data were collected as previously described in Mohammad Ennab and Hamid Mcheick (22 October 2022). Anchor describes individual predictions for black-box classification models by finding decision rules that sufficiently “anchor” the predictions. A rule freezes a prediction if changes in other feature values do not affect the prediction. Anchor leverages reinforcement learning techniques with graph search algorithms to reduce the number of model calls (required running time) to a minimum while still being able to recover from the local optimization (T, 2020).

7.2.9 Perturbation-based methods

Perturbation is the most basic technique for assessing input property modifications on output. This involves eliminating, masking, or altering specific inputs, then forward passing output and comparing to the original output. This is similar to the sensitivity analysis performed in parametric control system models. The input features that have the greatest impact on the outcome are prioritized. Because a forward pass must be performed after perturbing each set of features in the input, it is computationally intensive. In the case of picture data, the perturbation is accomplished by covering areas of the image with a grey patch, thereby obscuring them from the system’s view (Singh et al., 2020). In images, disturbance is done by covering areas with a grey patch, concealing them from the system’s view (Ennab and Mcheick, 2022a). The perturbation-based methods are broadly divided into input sampling-based methods and input optimization methods. RISE (Randomized Input Sample for Explanation) is the most representative algorithm among input sampling-based methods that put a random mask on the input image and saves the AI output. After around 8,000 repetitions, a linear combination of random masks and AI output values is performed. In this way, random masks made with high output values show consistency in feature exposure, allowing for a proper explanation (LG AI Researcher, 2021). However, the problem with this approach is that it requires about 8,000 outputs to illustrate. Another limitation is that the results are different each time it is performed on the same input image because a random mask is used (LG AI Researcher, 2021). The most representative algorithm in the latter input optimization method is an algorithm called extremal perturbation, which finds feature parts showing high output values by the optimization method. The problem with this approach, however, is that it relies only on the optimization method. When an accurate solution is obtained, a very interpretable explanation can be obtained, but when an exact solution is not obtained, features that seem completely unrelated may be displayed. In addition, since it uses a numerical optimization method, it has a computation time problem (LG AI Researcher, 2021).

7.2.10 Attention based

The fundamental concept of attention is motivated by how people pay attention to various areas of an image or other data sources in order to interpret them. The technique employed attention mechanisms, which included an image model and a language model, to show the detection process (Li et al., 2021). The interpretation in attention-based involves combining mechanisms for the selective traits of dominant features with attention towards the reported trait. One part of this will be selecting some of the hidden states over the time steps and on top of that, adding an attention layer to the present deep learning model. An attention score is computed at each important feature or time sequence to denote its importance. The attention mechanism found in the language model was used to learn the mapping between diagnostic reports and sights (Singh et al., 2020).

7.2.11 Testing concept activation vectors

In Kim et al. (2017), Concept Activation Vectors (TCAV) is another innovative approach used to explain the acquired characteristics of successive layers in terms of human-understandable concepts to domain experts without deep learning comprehension. It uses the network’s directional derivative in concept space in a similar way that saliency maps use input feature space. The directional derivative of the network in concept space is treated as a saliency map in TCAV. In this way, TCAV is perfectly suitable as the approach for detection of microaneurysms and aneurysms in the retina when saliency maps are used for the classification of the Diabetic Retinopathy (DR) level prediction and provides understandable reasoning to the physician for the level of DR. This provides justification whether a conceptual or physical structure is present in the image (Singh et al., 2020). However, many clinical concepts in medicine, such as structural texture or tissue shape, cannot be fully described by TCAV directly to prove their existence or non-existence (Kim et al., 2018). Many clinical concepts, such as form, texture, or shape, cannot be effectively described in terms of presence or absence and require a continuous scale of assessment (Ennab and Mcheick, 2022a).

7.2.12 Similar images

Stano et al. (2020) proposed research assessing layers of a 3D-CNN using a Gaussian mixture model (GMM) and binary encoding of training and test pictures based on their GMM components to offer explanations for comparable 3D images. As an explanation for its result, the software returned activation-wise similar training pictures utilizing the atlas. The model includes a perceptual code in binary vector space that defines the CNN’s processing of the input sample (Kim et al., 2018). A collection of the most perceptually similar and dissimilar samples may be retrieved from an existing atlas of labelled samples in order to support and further explain the choice made by the CNN model by calculating distances between pairs of samples in this perceptual encoding space. Applications of this model include Computer-Aided Diagnosis software using Computed Tomography (CT) data from medical imaging tests (Stano et al., 2020). The same imaging model was carried on the 3D MNIST datasets and Magnetic Resonance Imaging (MRI) datasets, and the findings were congenial atrophy conditions. There is an indication in some of the cases that the breakthrough of similarity with activation is invariant to the picture spatial orientation, and impact may be there in choices related to pictures returning (Singh et al., 2020).

7.2.13 Textual justification

The justification for the choice was given in the form of words or phrases and, on general, can communicate directly to both expert and non-expert users (Singh et al., 2020). A diagnostic phase and visual heatmaps for the breast-mass classification was developed using an explanation model that took input from the visual features of a classifier as well as the prediction embeddings (Lee et al., 2018). Justification generator was trained to generate justifications in the existence of a small number of medical reports using a visual word constraint loss (Singh et al., 2020). Data were taken as described earlier in Mohammad Ennab and Hamid Mcheick (22 October 2022). A diagnostic network and a justification generator are the two parts of the overall design of the model. Any general CADx network (classifier of malignant mass and benign mass) can be used as the diagnostic network. A visual characteristic and a diagnosis made by the diagnosis network are used by the justification generator. A visual word constraint loss is developed in the training stage to efficiently train the justification generator by preventing the training set’s sentences from being duplicated (Lee et al., 2019).

7.3 Characteristics of the interpretability models

It is important that when choosing the interpretation models to implement, many-paged features should be carefully and perhaps realized that no one fits all solution may exist:

  • 1. Model complexity vs. model interpretability: Some interpretability method simplifies complex models to be interpretable. Still, in that simplification, the majority of the time, a tradeoff between model accuracy and interpretability are made (Johansson et al., 2011).

  • 2. Suitability to different model types: not all interpretability method has a similar level of suitability for different kinds of machine learning models. Some work well with large deep neural networks, while others work better for decision tress or linear (Escalante et al., 2018).

  • 3. Local explanations: Most of the interpretation methods aim to present local explanations: for each prediction, the user can get an idea of which features that have played into account for that particular output. That is to say that these methods do not aim to have a global perception of the model’s behavior (Linardatos et al., 2020).

  • 4. Consistency: The interpretability methods can explain why the explanations are consistent depend on the way the input samples are chosen and hence might produce different explanations for marginally different sampled data points (Hu et al., 2019).

  • 5. Discrete explanation: Some of them offer differences in terms of interpretation where they provide interpretations in a discrete or rule-based way; therefore, the fine-grained nature of the original model might miss this kind of methods (Gilpin et al., 2018).

  • 6. Interpreting complex models: On the other hand, the interpretation of complex models like deep neural networks might be quite troublesome, and yes, the explanation might be done (Carvalho et al., 2019).

Table 3 summarizes the limitations associated with various interpretability models. Each model offers distinct advantages and challenges, with trade-offs between local and global explanations, computational overhead, and applicability across different architectures. While some models (like LIME and SHAP) provide local insights at the cost of consistency, others (like Grad–CAM and Bayesian Nonparametric) are tailored for specific use cases, such as convolutional neural networks. The table highlights how different interpretability methods suit various needs but may have constraints related to stability, complexity, and applicability beyond specific models or domains.

TABLE 3.

Limits of the interpretability models.

Model Limits
The interpretability based Offers detailed local explanations but may overlook the overall model behavior. The explanations can change based on the selected samples during analysis
LIME Relies on data samples tailored extensively to estimate the local model. Explanations can differ with sample selection, leading to inconsistent results. High computational overhead
SHAP Calculation of Shapley values is intensive, especially with high dimensionality. Based on cooperative game theory assumptions, which might not always hold
Grad-CAM Focused on Convolutional Neural Networks, might not be effective for other models. Provides class activation maps but lacks deep feature-level explanations
Bayesian Nonparametric Designed for CNNs, not useful for other model architectures. Provides class activation maps but lacks detailed feature-level explanations
Generalized Additive Models (GAM) Cannot capture complex interactions between features, less expressive power. Assumes linear and additive relationships, which might not hold in reality
MAPLE Provides local-level explanations at the expense of the global view. Explanations can vary depending on the samples used during analysis
Anchors Powerful for black-box models, less effective for transparent models. Provides discrete explanations, which might not capture model subtleties
Perturbation-Based Methods Computationally expensive for complex models. Explanations may vary due to random sampling, affecting stability
Attention-Based Methods Difficult to implement and complex. Provides importance to input data parts but lacks clear feature explanations
Testing Concept Activation Vectors Designed for neural networks, limited applicability to other models. Understanding results requires deep learning expertise
Similar Images Effectiveness limited by dataset quality. Provides feature-level information but explanations are image-level
Textual Justification Effective only if people can understand and believe them. Can potentially bias or leave ambiguous text explanations

7.4 Assessment of the interpretability models

The interpretability models can mainly be classified under the qualitative and quantitative models. The models that are applied to the quantitative evaluation are as follows:

  • 1. KAR method (keep and retrain): the modularity of the analysis on how the removal of the least significant N % pixel features in the saliency map is done in respect of the change in retrained model accuracy (Kim et al., 2019).

  • 2. ROAR method (remove and retrain): the removability analysis of how the accuracy of the retrained model is affected when excluding the most important N % pixel features in the saliency map (Kim et al., 2019).

The following are used as measures of qualitative evaluation:

  • 1. Coherence: This means that the input pattern that is closely related to the prediction given by the interpretability method needs to have an attribute that is somehow (Gilpin et al., 2018).

  • 2. Selectivity: in the case that input image, the exclusion of pixels that are rated important in the saliency map by the method, this, in turn leads to reduction of related probability that is related to the model prediction category that corresponds (Fuhrman et al., 2022).

  • 3. Implementation invariance: In the case of two models, for explanation, being similar, that is to say, two models that provide the same input produces the same output, then this, in turn means that interpretability method should give the same for two models (Carvalho et al., 2019).

  • 4. Class sensitivity: The explanation that is generated by interpretability method should be sensitive to the category (Nielsen et al., 2022).

  • 5. Explanation continuity interpretability methods should give similar explanations to similar input (Longo et al., 2024).

Figure 3 explains the categorization of the interpretability models into quantitative and qualitative methods. The quantitative models (KAR and ROAR) focus on performance changes with model retraining based on feature importance. In contrast, qualitative models (such as Coherence, Selectivity, Implementation Invariance, Class Sensitivity, and Explanation Continuity) assess interpretability based on the behavior and consistency of explanations across different inputs and predictions.

FIGURE 3.

FIGURE 3

Assessment of interpretability models categorized into quantitative and qualitative evaluation methods.

7.5 Recapitulation of main points

The application of AI in healthcare has demonstrated immense potential in automating diagnosis, treatment decisions, and patient monitoring. However, several key challenges persist:

  • • Advantages: AI systems have shown remarkable accuracy in tasks like diagnostic imaging, risk prediction, and disease screening. The ability of deep learning models to process large datasets quickly and accurately has outperformed traditional methods in several areas (e.g., cancer detection and diagnostic imaging).

  • • Limitations: The black-box nature of many deep learning models makes it difficult for healthcare providers to trust and interpret their decisions. The lack of transparency and interpretability remains a critical barrier to widespread adoption. Additionally, AI models often struggle to generalize across diverse patient populations, leading to reduced performance in real-world clinical settings.

  • • Challenges: Another significant challenge is the computational cost and real-time applicability of certain models. Many AI systems require significant computational resources, which limits their deployment in resource-constrained healthcare facilities. Furthermore, the integration of uncertainty quantification with interpretability models remains an underexplored area that could provide more reliable and explainable AI systems.

8 Lessons learned and future directions

Through our systematic review, several trends and common issues have emerged:

  • • Lack of real-time interpretability: Most high-accuracy AI models operate as black boxes, which limits their clinical applicability.

  • • Trade-offs between accuracy and interpretability: Models that prioritize accuracy often do so at the expense of interpretability, and vice versa.

  • • Generalizability issues: The majority of studies focus on specific datasets, making it difficult to generalize their findings to broader clinical populations.

8.1 Future directions

To address these challenges, future research should focus on developing hybrid models that balance accuracy and interpretability. Additionally, incorporating uncertainty quantification methods can improve the reliability of AI models in healthcare. Additionally, as AI continues to evolve, several key areas require further investigation:

  • • Improved Interpretability: Future research should focus on developing models that balance both accuracy and interpretability. Techniques such as Local Interpretable Model-Agnostic Explanations (LIME) and SHAP (Shapley Additive Explanations) could be further explored to enhance the transparency of deep learning models. Additionally, more effort is needed to integrate uncertainty quantification techniques to improve model reliability in real-time clinical settings.

  • • Incorporation of Multimodal Data: AI systems that can process diverse types of clinical data, such as medical imaging, patient history, and genetic information, are likely to provide more comprehensive and accurate predictions. Research should explore how multimodal AI systems can improve patient outcomes by integrating various data sources.

  • • User-Centered Design: A key challenge in AI healthcare applications is the need for user-centered design. Future research should explore how AI models can be developed in collaboration with healthcare providers to ensure that they are user-friendly, interpretable, and aligned with clinical workflows.

  • • AI in Personalized Medicine: While this review focused primarily on diagnostic imaging, future research should explore how AI can be applied to other domains, such as genomics and personalized medicine. AI models that can predict patient-specific outcomes and suggest personalized treatment plans based on genetic data have the potential to revolutionize healthcare.

  • • Learning-Based Manipulation: In the context of learning-based manipulation, as discussed in Stefanelli et al. (2023), future research should investigate how AI systems can integrate multiple sensory inputs (e.g., force, temperature) to mimic human-like responses in medical robotics and prosthetics. Testing and refining these methodologies in clinical scenarios could lead to more adaptive and responsive AI-driven prosthetic systems.

9 Conclusion

This review systematically examined AI models in healthcare, focusing on the trade-offs between accuracy and interpretability. We highlighted the most commonly used models and metrics, and identified key challenges, such as the black-box nature of deep learning models and generalizability issues. Our findings suggest that future AI research should prioritize transparency and safety, particularly in high-risk healthcare applications. By addressing these issues, healthcare providers can develop AI solutions that not only enhance performance but also build trust among stakeholders, ultimately leading to safer and more effective patient care. Additionally, incorporating strategies like those presented by Stefanelli et al. (2023), which integrate force and temperature information in prosthetic control, could offer new avenues for enhancing AI systems’ human-like capabilities in healthcare.

Funding Statement

The author(s) declare that no financial support was received for the research, authorship, and/or publication of this article.

Author contributions

ME: Data curation, Formal Analysis, Resources, Validation, Writing–original draft, Writing–review and editing. HM: Conceptualization, Supervision, Writing–original draft.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

  1. Afrash M. R., Shafiee M., Kazemi-Arpanahi H. (2023). Establishing machine learning models to predict the early risk of gastric cancer based on lifestyle factors. BMC Gastroenterol. 23 (1), 6. 10.1186/s12876-022-02626-x [DOI] [PMC free article] [PubMed] [Google Scholar]
  2. Alaa A., Gurdasani D., Harris A., Rashbass J., Schaar M. v. d. (2020) Machine learning to guide the use of adjuvant therapies for breast cancer. 10.21203/rs.3.rs-53594/v1 [DOI] [Google Scholar]
  3. Arbabshirani M. R., Fornwalt B. K., Mongelluzzo G. J., Suever J. D., Geise B. D., Patel A. A., et al. (2018). Advanced machine learning in action: identification of intracranial hemorrhage on computed tomography scans of the head with clinical workflow integration. NPJ Digit. Med. 1, 9. 10.1038/s41746-017-0015-z [DOI] [PMC free article] [PubMed] [Google Scholar]
  4. Archana R., Jeevaraj P. S. E. (2024). Deep learning models for digital image processing: a review. Artif. Intell. Rev. 57 (1), 11. 10.1007/s10462-023-10631-z [DOI] [Google Scholar]
  5. Arora A., Faulkner J., Paleri V., Kapoor K., Al-Lami A., Olaleye O., et al. (2024). New robotic platform for transoral robotic surgery: an IDEAL stage 0 study. BMJ Surg. interventions, & health Technol. 6 (1), e000181. 10.1136/bmjsit-2022-000181 [DOI] [PMC free article] [PubMed] [Google Scholar]
  6. Baeza-Yates R. (2022). “Ethical challenges in AI,” in WSDM ’22: The Fifteenth ACM International Conference on Web Search and Data Mining, Virtual Event, Tempe, AZ, USA, February 21–25, 2022 (ACM; ), 1–2. 10.1145/3488560.3498370 [DOI] [Google Scholar]
  7. Bibi N., Ud Din I., Sikandar M., Almogren A., Ali S. (2020). IoMT-based automated detection and classification of leukemia using deep learning. J. Healthc. Eng. 6648574, 6648574. 10.1155/2020/6648574 [DOI] [PMC free article] [PubMed] [Google Scholar]
  8. Bird D., McCallum H., Nix M. G., Teo M., Gilbert A., Casanova N., et al. (2021). Multicentre, deep learning, synthetic-CT generation for ano-rectal MR-only radiotherapy treatment planning. Radiother. Oncol. 156 (1), 23–28. 10.1016/j.radonc.2020.11.027 [DOI] [PMC free article] [PubMed] [Google Scholar]
  9. Biswas P., Sikander S., Kulkarni P. (2023). Recent advances in robot-assisted surgical systems. Biomed. Eng. Adv. 6, 100109. 10.1016/j.bea.2023.100109 [DOI] [Google Scholar]
  10. Bozyel S., Şimşek E., Koçyiğit Burunkaya D., Güler A., Korkmaz Y., Şeker M., et al. (2024). Artificial intelligence-based clinical decision support systems in cardiovascular diseases. Anatol. J. Cardiol. 28 (2), 74–86. 10.14744/AnatolJCardiol.2023.3685 [DOI] [PMC free article] [PubMed] [Google Scholar]
  11. Bukhari M., et al. (2022). “A deep learning framework for leukemia cancer detection in microscopic blood samples using squeeze and excitation learning,” in Mathematical problems in engineering. Editor Ali N. 10.1155/2022/2801227 [DOI] [Google Scholar]
  12. Cao S., et al. (2024). A systematic literature review on explainability for machine/deep learning-based software engineering research. ArXiv, abs/2401.14617. [Google Scholar]
  13. Carvalho D. V., Pereira E. M., Cardoso J. S. (2019). Machine learning interpretability: a survey on methods and metrics. Electronics 8 (8), 832. 10.3390/electronics8080832 [DOI] [Google Scholar]
  14. Chen D., Wu L., Li Y., Zhang J., Liu J., Huang L., et al. (2020). Comparing blind spots of unsedated ultrafine, sedated, and unsedated conventional gastroscopy with and without artificial intelligence: a prospective, single-blind, 3-parallel-group, randomized, single-center trial. Gastrointest. Endosc. 91 (2), 332–339.e3. 10.1016/j.gie.2019.09.016 [DOI] [PubMed] [Google Scholar]
  15. Chen K., Shao A., Burapacheep J., Li Y. (2024). Conversational AI and equity through assessing GPT-3’s communication with diverse social groups on contentious topics. Sci. Rep. 14 (1), 1561. 10.1038/s41598-024-51969-w [DOI] [PMC free article] [PubMed] [Google Scholar]
  16. Chiou S.-Y., Zhang Z.-Y., Liu H.-L., Yan J.-L., Wei K.-C., Chen P.-Y. (2022). Augmented reality surgical navigation system for external ventricular drain. Basel, Switzerland: Healthcare; 10 (10). 10.3390/healthcare10101815 [DOI] [PMC free article] [PubMed] [Google Scholar]
  17. Data sharing in the age of deep learning (2023). Nature Biotechnology 41 (4), 433. 10.1038/s41587-023-01770-3 [DOI] [PubMed] [Google Scholar]
  18. Dratsch T., Zäske C., Siedek F., Rauen P., Hokamp N. G., Sonnabend K., et al. (2024). Reconstruction of 3D knee MRI using deep learning and compressed sensing: a validation study on healthy volunteers. Eur. Radiol. Exp. 8 (1), 47. 10.1186/s41747-024-00446-0 [DOI] [PMC free article] [PubMed] [Google Scholar]
  19. Eckardt J.-N., Middeke J. M., Riechert S., Schmittmann T., Sulaiman A. S., Kramer M., et al. (2022). Deep learning detects acute myeloid leukemia and predicts NPM1 mutation status from bone marrow smears. Leukemia 36 (1), 111–118. 10.1038/s41375-021-01408-w [DOI] [PMC free article] [PubMed] [Google Scholar]
  20. El Shawi R., et al. (2019). “Interpretability in HealthCare A comparative study of local machine learning interpretability techniques,” in 2019 IEEE 32nd international symposium on computer-based medical systems (CBMS), 275–280. 10.1109/CBMS.2019.00065 [DOI] [Google Scholar]
  21. Ennab M., Mcheick H. (2022a). Designing an interpretability-based model to explain the artificial intelligence algorithms in healthcare. Diagn. (Basel, Switzerland) 12 (7), 1557. 10.3390/diagnostics12071557 [DOI] [PMC free article] [PubMed] [Google Scholar]
  22. Ennab M., Mcheick H. (2022b). Survey of COVID-19 prediction models and their limitations. Inf. Syst. 11 (2), 14–21. 10.11648/j.ijiis.20221102.11 [DOI] [Google Scholar]
  23. Escalante H. J., Escalera S., Guyon I., Baró X., Güçlütürk Y., Güçlü U. (2018). “Explainable and interpretable models in computer vision and machine learning,” in Springer Verlag (The Springer Series on Challenges in Machine Learning; ). Available at: https://inria.hal.science/hal-01991623 . [Google Scholar]
  24. Esteva A, Kuprel B., Novoa R. A., Ko J., Swetter S. M., Blau H. M., et al. (2017). Erratum: corrigendum: Dermatologist-level classification of skin cancer with deep neural networks. Nature 546 (7660), 686. 10.1038/nature22985 [DOI] [PubMed] [Google Scholar]
  25. Frisch S., Libuschewski H., Peters S., Gerß J., von Hoff K., Kortmann R. D., et al. (2024). Radiation therapy plays an important role in the treatment of atypical teratoid/rhabdoid tumors: analysis of the EU-RHAB cohorts and their precursors. Int. J. Radiat. Oncol. Biol. Phys. 119 (4), 1147–1157. 10.1016/j.ijrobp.2024.01.200 [DOI] [PubMed] [Google Scholar]
  26. Fu Y., Liu R., Liu Y., Lu J. (2019). Intertrochanteric fracture visualization and analysis using a map projection technique. Med. Biol. Eng. Comput. 57 (3), 633–642. 10.1007/s11517-018-1905-1 [DOI] [PubMed] [Google Scholar]
  27. Fuhrman J. D., Gorre N., Hu Q., Li H., El Naqa I., Giger M. L. (2022). A review of explainable and interpretable AI with applications in COVID‐19 imaging. Med. Phys. 49 (1), 1–14. 10.1002/mp.15359 [DOI] [PMC free article] [PubMed] [Google Scholar]
  28. Garreau D., von Luxburg U. (2020). “Explaining the explainer: a first theoretical analysis of LIME,” in Proceedings of the twenty third international Conference on artificial Intelligence and statistics. PMLR (proceedings of machine learning research). Editors Chiappa S., Calandra R., 1287–1296. Available at: https://proceedings.mlr.press/v108/garreau20a.html. [Google Scholar]
  29. Gilpin L. H., et al. (2018). “Explaining explanations: an overview of interpretability of machine learning,” in 2018 IEEE 5th International Conference on data science and advanced analytics (DSAA) (IEEE; ), 80–89. [Google Scholar]
  30. Gottlieb M., Kline J. A., Schneider A. J., Coates W. C. (2023). ChatGPT and conversational artificial intelligence: friend, foe, or future of research? Am. J. Emerg. Med. 70, 81–83. 10.1016/j.ajem.2023.05.018 [DOI] [PubMed] [Google Scholar]
  31. Gumbs A. A., Frigerio I., Spolverato G., Croner R., Illanes A., Chouillard E., et al. (2021). Artificial intelligence surgery: how do we get to autonomous actions in surgery? Sensors (Basel, Switzerland) 21 (16), 5526. 10.3390/s21165526 [DOI] [PMC free article] [PubMed] [Google Scholar]
  32. Guo W., et al. (2018). Explaining deep learning models--A bayesian non-parametric approach. Adv. neural Inf. Process. Syst. 31. [Google Scholar]
  33. Hall W. A., Paulson E., Li X. A., Erickson B., Schultz C., Tree A., et al. (2022). Magnetic resonance linear accelerator technology and adaptive radiation therapy: an overview for clinicians. CA a cancer J. Clin. 72 (1), 34–56. 10.3322/caac.21707 [DOI] [PMC free article] [PubMed] [Google Scholar]
  34. Hamedan F., Orooji A., Sanadgol H., Sheikhtaheri A. (2020). Clinical decision support system to predict chronic kidney disease: a fuzzy expert system approach. Int. J. Med. Inf. 138 (104134), 104134. 10.1016/j.ijmedinf.2020.104134 [DOI] [PubMed] [Google Scholar]
  35. Hamm C. A., Baumgärtner G. L., Biessmann F., Beetz N. L., Hartenstein A., Savic L. J., et al. (2023). Interactive explainable deep learning model informs prostate cancer diagnosis at MRI. Radiology 307, 222276. 10.1148/radiol.222276 [DOI] [PubMed] [Google Scholar]
  36. Han T., et al. (2023) MedAlpaca – an open-source collection of medical conversational AI models and training data. 10.48550/arXiv.2304.08247 [DOI] [Google Scholar]
  37. Haptik T. (2023). “haptik”, conversational AI in healthcare. Spring. Available at: https://www.haptik.ai/blog/conversational-ai-healthcare-use-cases/. [Google Scholar]
  38. Hassan S. A., Akbar S., Khan H. U. (2024). Detection of central serous retinopathy using deep learning through retinal images. Multimedia Tools Appl. 83 (7), 21369–21396. 10.1007/s11042-023-16206-y [DOI] [Google Scholar]
  39. Helman S., Terry M. A., Pellathy T., Williams A., Dubrawski A., Clermont G., et al. (2022). Engaging clinicians early during the development of a graphical user display of an intelligent alerting system at the bedside. Int. J. Med. Inform. 159, 104643. 10.1016/j.ijmedinf.2021.104643 [DOI] [PMC free article] [PubMed] [Google Scholar]
  40. Hooshmand S. A., Zarei Ghobadi M., Hooshmand S. E., Azimzadeh Jamalkandi S., Alavi S. M., Masoudi-Nejad A. (2021). A multimodal deep learning-based drug repurposing approach for treatment of COVID-19. Mol. Divers 25 (3), 1717–1730. 10.1007/s11030-020-10144-9 [DOI] [PMC free article] [PubMed] [Google Scholar]
  41. Hossain M. M., Walid M. A. A., Galib S. S., Azad M. M., Rahman W., Shafi A., et al. (2024). COVID-19 detection from chest CT images using optimized deep features and ensemble classification. Syst. Soft Comput. 6, 200077. 10.1016/j.sasc.2024.200077 [DOI] [Google Scholar]
  42. Howard F. M., Kochanny S., Koshy M., Spiotto M., Pearson A. T. (2020). Machine learning-guided adjuvant treatment of head and neck cancer. JAMA Netw. open 3 (11), e2025881. 10.1001/jamanetworkopen.2020.25881 [DOI] [PMC free article] [PubMed] [Google Scholar]
  43. Hu Y., et al. (2019). “Multi-modal probabilistic prediction of interactive behavior via an interpretable model,” in 2019 IEEE intelligent vehicles symposium (IV) (IEEE; ), 557–563. [Google Scholar]
  44. Jing J., Liu Z., Guan H., Zhu W., Zhang Z., Meng X., et al. (2023). A deep learning system to predict recurrence and disability outcomes in patients with transient ischemic attack or ischemic stroke. Adv. Intell. Syst. 5 (4), 2200240. 10.1002/aisy.202200240 [DOI] [Google Scholar]
  45. Johansson U., Sönströd C., Norinder U., Boström H. (2011). Trade-off between accuracy and interpretability for predictive in silico modeling. Future Med. Chem. 3 (6), 647–663. 10.4155/fmc.11.23 [DOI] [PubMed] [Google Scholar]
  46. Kaushik P. (2023). Artificial intelligence accelerated transformation in the healthcare industry. Amity J. Prof. Pract. 3. 10.55054/ajpp.v3i01.630 [DOI] [Google Scholar]
  47. Ke Y.-Y., Peng T.-T., Yeh T. K., Huang W. Z., Chang S. E., Wu S. H., et al. (2020). Artificial intelligence approach fighting COVID-19 with repurposing drugs. Biomed. J. 43 (4), 355–362. 10.1016/j.bj.2020.05.001 [DOI] [PMC free article] [PubMed] [Google Scholar]
  48. Khalifa M., Albadawy M., Iqbal U. (2024). Advancing clinical decision support: the role of artificial intelligence across six domains. Comput. Methods Programs Biomed. Update 5, 100142. 10.1016/j.cmpbup.2024.100142 [DOI] [Google Scholar]
  49. Kiani A., Uyumazturk B., Rajpurkar P., Wang A., Gao R., Jones E., et al. (2020). Impact of a deep learning assistant on the histopathologic classification of liver cancer. NPJ Digit. Med. 3, 23. 10.1038/s41746-020-0232-8 [DOI] [PMC free article] [PubMed] [Google Scholar]
  50. Kim B., et al. (2017) Interpretability beyond feature attribution: quantitative testing with concept activation vectors (TCAV). 10.48550/ARXIV.1711.11279 [DOI] [Google Scholar]
  51. Kim B., et al. (2019). “Why are saliency maps noisy? Cause of and solution to noisy saliency maps,” in 2019 IEEE/CVF international conference on computer vision workshop (ICCVW), Seoul, Korea (South), October 27–28, 2019 (IEEE; ), 4149–4157. [Google Scholar]
  52. Kim H., Yoo S. K., Kim J. S., Kim Y. T., Lee J. W., Kim C., et al. (2024). Clinical feasibility of deep learning-based synthetic CT images from T2-weighted MR images for cervical cancer patients compared to MRCAT. Sci. Rep. 14 (1), 8504. 10.1038/s41598-024-59014-6 [DOI] [PMC free article] [PubMed] [Google Scholar]
  53. Kim K. (2023). “COGNEX”, Overview of interpretable machine learning: (2) main interpretation methods of deep learning models in image recognition problems, 21 February. Available at: https://www.cognex.com/ko-kr/blogs/deep-learning/research/overview-interpretable-machine-learning-2-interpreting-deep-learning-models-image-recognition (Accessed February 21, 2023).
  54. Lee A., Baker T. S., Bederson J. B., Rapoport B. I. (2024). Levels of autonomy in FDA-cleared surgical robots: a systematic review. npj Digit. Med. 7 (1), 103. 10.1038/s41746-024-01102-y [DOI] [PMC free article] [PubMed] [Google Scholar]
  55. Lee G., Jeong J., Seo S., Kim C., Kang P. (2018). Sentiment classification with word localization based on weakly supervised learning with a convolutional neural network. Knowledge-Based Syst. 152 (15), 70–82. 10.1016/j.knosys.2018.04.006 [DOI] [Google Scholar]
  56. Lee H., Kim S. T., Ro Y. M. (2019). Generation of multimodal justification using visual word constraint model for explainable computer-aided diagnosis, abs/1906.03922. CoRR, 21–29. 10.1007/978-3-030-33850-3_3 [DOI] [Google Scholar]
  57. LG AI Researcher (2021). New algorithm SISE of “explainable artificial intelligence” (XAI), Naver. Available at: https://m.post.naver.com/viewer/postView.naver?volumeNo=30627518&memberNo=52249799. (Accessed May 20, 2024)
  58. Li A., Xiao F., Zhang C., Fan C. (2021). Attention-based interpretable neural network for building cooling load prediction. Appl. Energy 299, 117238. 10.1016/j.apenergy.2021.117238 [DOI] [Google Scholar]
  59. Linardatos P., Papastefanopoulos V., Kotsiantis S. (2020). Explainable ai: a review of machine learning interpretability methods. Entropy 23 (1), 18. 10.3390/e23010018 [DOI] [PMC free article] [PubMed] [Google Scholar]
  60. Linardatos P., Papastefanopoulos V., Kotsiantis S. (2021) Entropy [preprint]. [DOI] [PMC free article] [PubMed] [Google Scholar]
  61. Lindsey R., Daluiski A., Chopra S., Lachapelle A., Mozer M., Sicular S., et al. (2018). Deep neural network improves fracture detection by clinicians. Proc. Natl. Acad. Sci. U. S. A. 115 (45), 11591–11596. 10.1073/pnas.1806905115 [DOI] [PMC free article] [PubMed] [Google Scholar]
  62. Liu B., Guo J., Chen C. L. P., Wu X., Zhang T. (2023). Fine-grained interpretability for EEG emotion recognition: concat-aided grad-CAM and systematic brain functional network. IEEE Trans. Affect. Comput. 15, 671–684. 10.1109/taffc.2023.3288885 [DOI] [Google Scholar]
  63. Lo C.-M., Hung P.-H., Lin D.-T. (2021). Rapid assessment of acute ischemic stroke by computed tomography using deep convolutional neural networks. J. Digit. Imaging 34 (3), 637–646. 10.1007/s10278-021-00457-y [DOI] [PMC free article] [PubMed] [Google Scholar]
  64. Lombardi A., Monaco A., Donvito G., Amoroso N., Bellotti R., Tangaro S. (2021). Brain age prediction with morphological features using deep neural networks: Results from predictive analytic competition 2019. Front. Psychiatry. 11, 619629. [DOI] [PMC free article] [PubMed] [Google Scholar]
  65. Longo L., Brcic M., Cabitza F., Choi J., Confalonieri R., Ser J. D., et al. (2024). Explainable Artificial Intelligence (XAI) 2.0: a manifesto of open challenges and interdisciplinary research directions. Inf. Fusion 106, 102301. 10.1016/j.inffus.2024.102301 [DOI] [Google Scholar]
  66. Lotter W., Haslam B., Diab A. R., Kim J. G., Grisot G., Wu E., et al. (2021). Robust breast cancer detection in mammography and digital breast tomosynthesis using an annotation-efficient deep learning approach. Nat. Med. 27 (2), 244–249. 10.1038/s41591-020-01174-9 [DOI] [PMC free article] [PubMed] [Google Scholar]
  67. Masood A., Sheng B., Li P., Hou X., Wei X., Qin J., et al. (2018). Computer-assisted decision support system in pulmonary cancer detection and stage classification on CT images. J. Biomed. Inf. 79, 117–128. 10.1016/j.jbi.2018.01.005 [DOI] [PubMed] [Google Scholar]
  68. McGenity C., Clarke E. L., Jennings C., Matthews G., Cartlidge C., Freduah-Agyemang H., et al. (2024). Artificial intelligence in digital pathology: a systematic review and meta-analysis of diagnostic test accuracy. npj Digit. Med. 7 (1), 114. 10.1038/s41746-024-01106-8 [DOI] [PMC free article] [PubMed] [Google Scholar]
  69. Mella M. H., Chabrillac E., Dupret-Bories A., Mirallie M., Vergez S. (2023). Transoral robotic surgery for head and neck cancer: advances and residual knowledge gaps. J. Clin. Med. 12 (6), 2303. 10.3390/jcm12062303 [DOI] [PMC free article] [PubMed] [Google Scholar]
  70. Mitchell R., Cooper J., Frank E., Holmes G. (2022). Sampling permutations for shapley value estimation. J. Mach. Learn. Res. 23 (43), 1–46. [Google Scholar]
  71. Mohamed E. A., Rashed E. A., Gaber T., Karam O. (2022). Deep learning model for fully automated breast cancer detection system from thermograms. PLOS ONE 17 (1), 1–20. 10.1371/journal.pone.0262349 [DOI] [PMC free article] [PubMed] [Google Scholar]
  72. Moradi M., Samwald M. (2021). Post-hoc explanation of black-box classifiers using confident itemsets. Expert Syst. Appl. 165, 113941. 10.1016/j.eswa.2020.113941 [DOI] [Google Scholar]
  73. Nicolae A., Semple M., Lu L., Smith M., Chung H., Loblaw A., et al. (2020). Conventional vs machine learning-based treatment planning in prostate brachytherapy: results of a Phase I randomized controlled trial. Brachytherapy 19 (4), 470–476. 10.1016/j.brachy.2020.03.004 [DOI] [PubMed] [Google Scholar]
  74. Nielsen I. E., Dera D., Rasool G., Ramachandran R. P., Bouaynaya N. C. (2022). Robust explainability: a tutorial on gradient-based attribution methods for deep neural networks. IEEE Signal Process. Mag. 39 (4), 73–84. 10.1109/msp.2022.3142719 [DOI] [Google Scholar]
  75. Noh HS, Park J., Sim G., Yu J. E., Chung Y. (2014). Nonparametric bayesian statistical models in biomedical research. Korean J. Appl. Statistics 27 (6), 867–889. 10.5351/kjas.2014.27.6.867 [DOI] [Google Scholar]
  76. Origimid (2022). Artificial intelligence assisted diagnosis. Available at: https://www.origimed.com/Product/ai (Accessed May 20, 2024).
  77. Pang S., Ding T., Qiao S., Meng F., Wang S., Li P., et al. (2019). A novel YOLOv3-arch model for identifying cholelithiasis and classifying gallstones on CT images. PLoS one 14 (6), e0217647. 10.1371/journal.pone.0217647 [DOI] [PMC free article] [PubMed] [Google Scholar]
  78. Pavlidis G. (2024). Unlocking the black box: analysing the EU artificial intelligence act’s framework for explainability in AI. Law, Innovation Technol. 16, 293–308. 10.1080/17579961.2024.2313795 [DOI] [Google Scholar]
  79. Plumb G., Molitor D., Talwalkar A. S. (2018). “Model agnostic supervised local explanations,” in Advances in neural information processing systems. Editor Bengio S., et al. (Curran Associates, Inc.). Available at: https://proceedings.neurips.cc/paper/2018/file/b495ce63ede0f4efc9eec62cb947c162-Paper.pdf (Accessed January 6, 2024). [Google Scholar]
  80. Quiroz J. C., Feng Y. Z., Cheng Z. Y., Rezazadegan D., Chen P. K., Lin Q. T., et al. (2021). Development and validation of a machine learning approach for automated severity assessment of COVID-19 based on clinical and imaging data: retrospective study. JMIR Med. Inf. 9 (2), e24572. 10.2196/24572 [DOI] [PMC free article] [PubMed] [Google Scholar]
  81. Rawson T. M., Hernandez B., Moore L. S. P., Herrero P., Charani E., Ming D., et al. (2021). A real-world evaluation of a case-based reasoning algorithm to support antimicrobial prescribing decisions in acute care. Clin. Infect. Dis. 72 (12), 2103–2111. 10.1093/cid/ciaa383 [DOI] [PubMed] [Google Scholar]
  82. Reddy K., Gharde P., Tayade H., Patil M., Reddy L. S., Surya D. (2023). Advancements in robotic surgery: a comprehensive overview of current utilizations and upcoming frontiers. Cureus. 15 (12), e50415. 10.7759/cureus.50415 [DOI] [PMC free article] [PubMed] [Google Scholar]
  83. Ribeiro M. T., Singh S., Guestrin C. (2018). Anchors: high-precision model-agnostic explanations. Proc. AAAI Conf. Artif. Intell. 32 (1). 10.1609/aaai.v32i1.11491 [DOI] [Google Scholar]
  84. Sadeghi Z., Alizadehsani R., Cifci M. A., Kausar S., Rehman R., Mahanta P., et al. (2024). A review of explainable artificial intelligence in healthcare. Comput. Electr. Eng. 118, 109370. 10.1016/j.compeleceng.2024.109370 [DOI] [Google Scholar]
  85. Saeed W., Omlin C. (2023). Explainable AI (XAI): a systematic meta-survey of current challenges and future opportunities. Knowledge-Based Syst. 263, 110273. 10.1016/j.knosys.2023.110273 [DOI] [Google Scholar]
  86. Salem N. M., Jack K. M., Gu H., Kumar A., Garcia M., Yang P., et al. (2023). Machine and deep learning identified metabolites and clinical features associated with gallstone disease. Comput. Methods Programs Biomed. Update 3, 100106. 10.1016/j.cmpbup.2023.100106 [DOI] [Google Scholar]
  87. Salih S. (2022). Understanding machine learning interpretability. Available at: https://towardsdatascience.com/understanding-machine-learning-interpretability-168fd7562a1a (Accessed February 6, 2024).
  88. Savindu Herath Pathirannehelage Y. R. S., Krogh G. von, von Krogh G. (2024). Design principles for artificial intelligence-augmented decision making: an action design research study. Eur. J. Inf. Syst. 0 (0), 1–23. 10.1080/0960085X.2024.2330402 [DOI] [Google Scholar]
  89. Schöttl A. (2020). “A light-weight method to foster the (Grad) CAM interpretability and explainability of classification networks,” in 2020 10th international conference on advanced computer information technologies (ACIT), Deggendorf, Germany, September 16–18, 2020 (IEEE; ), 348–351. [Google Scholar]
  90. Selvaraju R. R., et al. (2017). “Grad-cam: visual explanations from deep networks via gradient-based localization,” in Proceedings of the IEEE international conference on computer vision, Venice, Italy, October 22–29, 2017 (IEEE; ), 618–626. [Google Scholar]
  91. Serednicki W. A., Hołówko W., Major P., Małczak P., Pędziwiatr M. (2023). Minimizing blood loss and transfusion rate in laparoscopic liver surgery: a review. Wideochir. Inne. Tech. Maloinwazyjne 18 (2), 213–223. 10.5114/wiitm.2022.124088 [DOI] [PMC free article] [PubMed] [Google Scholar]
  92. Shapley L. S. (2016) “17. A value for n-person games,” in Contributions to the theory of games (AM-28), eds Kuhn H. W., Tucker A. W. Vol. II. (Princeton University Press; ), pp. 307–318. 10.1515/9781400881970-018 [DOI] [Google Scholar]
  93. Shi S., Du Y., Fan W. (2020) An extension of LIME with improvement of interpretability and fidelity. [Google Scholar]
  94. Shorfuzzaman M., Masud M., Alhumyani H., Anand D., Singh A. (2021). Artificial neural network-based deep learning model for COVID-19 patient detection using X-ray chest images. J. Healthc. Eng. 2021, 1–16. 10.1155/2021/5513679 [DOI] [PMC free article] [PubMed] [Google Scholar]
  95. Singh A., Sengupta S., Lakshminarayanan V. (2020). Explainable deep learning models in medical image analysis. arXiv 6, 52. 10.3390/jimaging6060052 [DOI] [PMC free article] [PubMed] [Google Scholar]
  96. Sinha S., Vegesna R., Mukherjee S., Kammula A. V., Dhruba S. R., Wu W., et al. (2024). PERCEPTION predicts patient response and resistance to treatment using single-cell transcriptomics of their tumors. Nat. Cancer 5, 938–952. 10.1038/s43018-024-00756-7 [DOI] [PubMed] [Google Scholar]
  97. Siouras A., Moustakidis S., Giannakidis A., Chalatsis G., Liampas I., Vlychou M., et al. (2022). Knee injury detection using deep learning on MRI studies: a systematic review. Diagn. (Basel, Switzerland) 12 (2), 537. 10.3390/diagnostics12020537 [DOI] [PMC free article] [PubMed] [Google Scholar]
  98. Stano M., Benesova W., Martak L. S. (2020). “Explainable 3D convolutional neural network using GMM encoding,” in Twelfth international conference on machine vision (ICMV 2019). Editors Osten W., Nikolaev D. P. (SPIE; ), 507–513. 10.1117/12.2557314 [DOI] [Google Scholar]
  99. Stefanelli E., Cordella F., Gentile C., Zollo L. (2023). Hand prosthesis sensorimotor control inspired by the human somatosensory system. Robotics 12 (5), 136. 10.3390/robotics12050136 [DOI] [Google Scholar]
  100. Stinson C., Vlaad S. (2024). A feeling for the algorithm: diversity, expertise, and artificial intelligence. Big Data & Soc. 11. 10.1177/20539517231224247 [DOI] [Google Scholar]
  101. Strumbelj E., Kononenko I. (2011) ‘A general method for visualizing and explaining black-box regression models, , 21–30. 10.1007/978-3-642-20267-4_3 [DOI] [Google Scholar]
  102. Subramani S., Varshney N., Anand M. V., Soudagar M. E. M., Al-Keridis L. A., Upadhyay T. K., et al. (2023). Cardiovascular diseases prediction by machine learning incorporation with deep learning. Front. Med. 10, 1150933. 10.3389/fmed.2023.1150933 [DOI] [PMC free article] [PubMed] [Google Scholar]
  103. T J. (2020). [Interpretable machine learning (5-8)] scope rule (anchor). Tistory. Available at: https://eair.tistory.com/27 (Accessed May 12, 2024). [Google Scholar]
  104. Vinogradova K., Dibrov A., Myers G. (2020). “Towards interpretable semantic segmentation via gradient-weighted class activation mapping (student abstract),” in Proceedings of the AAAI conference on artificial intelligence, New York, United States, February 2020, 13943–13944. [Google Scholar]
  105. Wang P., Liu X., Berzin T. M., Glissen Brown J. R., Liu P., Zhou C., et al. (2020). Effect of a deep-learning computer-aided detection system on adenoma detection during colonoscopy (CADe-DB trial): a double-blind randomised study. Lancet Gastroenterol. Hepatol 5 (4), 343–351. 10.1016/S2468-1253(19)30411-X [DOI] [PubMed] [Google Scholar]
  106. Wu X., Tao Y., Qiu Q., Wu X. (2018). Application of image recognition-based automatic hyphae detection in fungal keratitis. Australas. Phys. Eng. Sci. Med. 41 (1), 95–103. 10.1007/s13246-017-0613-8 [DOI] [PubMed] [Google Scholar]
  107. Wu Yazhou Y. D., Chen X. (2022). Research progress and prospect of artificial intelligence in the clinical field. J. Army Med. Univ. 44, 89–102. 10.16016/j.1000-5404.202112092 [DOI] [Google Scholar]
  108. Xie X., Li K., Wu Y., Ai C., Wang Q., Wang C., et al. (2021). Preliminary experiences with robot-assisted choledochal cyst excision using the da vinci surgical system in children below the age of one. Front. Pediatr. 23 (9), 741098. 10.3389/fped.2021.741098 [DOI] [PMC free article] [PubMed] [Google Scholar]
  109. Yang W., Wei Y., Wei H., Chen Y., Huang G., Li X., et al. (2023). Survey on explainable AI: from approaches, limitations and applications aspects. Human-Centric Intell. Syst. 3 (3), 161–188. 10.1007/s44230-023-00038-y [DOI] [Google Scholar]
  110. Yoo Y. J., Ha E. J., Cho Y. J., Kim H. L., Han M., Kang S. Y. (2018). Computer-aided diagnosis of thyroid nodules via ultrasonography: initial clinical experience. Korean J. Radiol. 19 (4), 665–672. 10.3348/kjr.2018.19.4.665 [DOI] [PMC free article] [PubMed] [Google Scholar]
  111. Zadeh Shirazi A., McDonnell M. D., Fornaciari E., Bagherian N. S., Scheer K. G., Samuel M. S., et al. (2021). A deep convolutional neural network for segmentation of whole-slide pathology images identifies novel tumour cell-perivascular niche interactions that are associated with poor survival in glioblastoma. Br. J. Cancer 125 (3), 337–350. 10.1038/s41416-021-01394-x [DOI] [PMC free article] [PubMed] [Google Scholar]
  112. Zhang J., Zhang Z. (2023). Ethics and governance of trustworthy medical artificial intelligence. BMC Med. Inf. Decis. Mak. 23 (1), 7. 10.1186/s12911-023-02103-9 [DOI] [PMC free article] [PubMed] [Google Scholar]
  113. Zheng S., Rammohan N., Sita T., Teo P. T., Wu Y., Lesniak M., et al. (2024). GlioPredictor: a deep learning model for identification of high-risk adult IDH-mutant glioma towards adjuvant treatment planning. Sci. Rep. 14 (1), 2126. 10.1038/s41598-024-51765-6 [DOI] [PMC free article] [PubMed] [Google Scholar]
  114. Zhongqi H., Jia Z. C. (2022). Development and application of software testing under artificial intelligence, 21ks. Available at: https://www.21ks.net/lunwen/rgznlw/180797.html (Accessed May 23, 2024).

Articles from Frontiers in Robotics and AI are provided here courtesy of Frontiers Media SA

RESOURCES