Skip to main content
. 2023 Jan 5;23(2):634. doi: 10.3390/s23020634

Table 2.

The following table is a collection of papers that have used interpretable methods in Section 4 to improve the algorithm.

Paper Organ XAI Modality Contribution
[42] bone CAM X-ray The model aims to predict the degree of knee damage and pain value through X-ray image.
[43] lung CAM Ultrasound, X-ray It uses three kinds of lung ultrasound images as datasets, and two networks, VGG-16 and VGG-CAM, to classify three kinds of pneumonia.
[44] breast CAM X-ray It proposes a globally-aware multiple instance classifier (GMIC) that uses CAM to identify the most informative regions with local and global information.
[45] lung CAM X-ray, CT The study improves two models, one of them based on MobileNet to classify COVID-19 CXR images, the other one is ResNet for CT image classification.
[46] lung CAM CT It selects healthy and COVID-19 patient’s data for training DRE-Net model.
[47] lung Grad-CAM CT It proposes a method of deep feature fusion. It achieves better performance than the single use of CNN.
[48] chest Grad-CAM ultrasound The paper proposes a semi-supervised model based on attention mechanism and disentangled. It then uses Grad-CAM to improve model’s explainable.
[49] lung Grad-CAM X-ray It provides a computer-aided detection, which is composed of the Discrimination-DL and the Localization-DL, and uses Grad-CAM to locate abnormal areas in the image.
[50] colon Grad-CAM colonoscopy The study proposes DenseNet121 to predict if the patient has ulcerative colitis (UC).
[51] colon Grad-CAM whole-slide images It investigates the potential of a deep learning-based system for automated MSI prediction.
[52] lung Grad-CAM CT It shows a classifier based on the Res2Net network. The study uses Activation Mapping to increase the interpretability of the overall Joint Classification and Segmentation system.
[53] chest Grad-CAM CT It proposes a neighboring aware graph neural network (NAGNN) for COVID-19 detection based on chest CT images.
[54] lung Grad-CAM, LIME X-ray This work provides a COVID-19 X-ray dataset, and proposes a COVID-CXNet based on CheXNet using transfer learning.
[55] lung Grad-CAM, LIME X-ray, CT It compares five DL models and uses the visualization method to explain NASNetLarge.
[56] breast Attention X-ray It provides the triple-attention learning A3 Net model to diagnose 14 chest diseases.
[57] bone Attention CT The study introduces a multimodal spatial attention module (MSAM). It uses an attention mechanism to focus on the area of interest.
[58] colon Attention colonoscopy The proposed Focus U-Net achieves an average DSC and IoU of 87.8% and 80.9%, respectively.
[59] lung, skin Saliency CT, X-ray The work presents quantitative assessment metrics for saliency XAI. Three different saliency algorithms were evaluated.
[60] lung SHAP EHR The study introduces a predictive length of stay framework to deal with imbalanced EHR datasets.
[61] - SHAP EHR The study presents an explainable clinical decision support system (CDSS) to help clinicians identify women at risk for Gestational Diabetes Mellitus (GDM).
[62] - SHAP radiomics The study proposes a pipeline for interactive medical image analysis via radiomics.
[63] lung SHAP CT This paper provides a model to predict mutation in patients with non-small cell lung cancer.
[64] chest SHAP EHR In this paper, it compares the performance of different ML methods (RSFs, SSVMs, and XGB and CPH regression) and uses SHAP value to interpret the models.
[65] chest LIME, SHAP X-ray The study proposes a unified pipeline to improve explainability for CNN using multiple XAI methods.
[66] lung SHAP, LIME, Scoped Rules EHR The study provides a comparison among three feature-based XAI techniques on EHR dataset. The results show that the use of these techniques can not replace human experts.
[67] chest Image caption CT It proposes Medical-VLBERT for COVID-19 CT report generation.

CAM: class activation map, AUC: area under the ROC curve, ROC: receiver operating characteristic curve, LIME: local interpretable model-agnostic explanation, EHR: electronic health record, SHAP: Shapley additive explanations.