Self-eXplainable AI for Medical Image Analysis
Self-eXplainable AI for Medical Image Analysis
Post-hoc XAI
making areas such as medical image analysis, has led to
arXiv:2410.02331v1 [cs.CV] 3 Oct 2024
S-XAI
training process of deep learning models. This approach
allows models to generate inherent explanations that are
Explainable Input Explanation
closely aligned with their internal decision-making pro- Explainable Model
cesses. Such enhanced transparency significantly sup-
ports the trustworthiness, robustness, and accountabil- Fig. 1. Illustration of post-hoc XAI versus Self-eXplainable AI (S-XAI).
ity of AI systems in real-world medical applications. To
facilitate the development of S-XAI methods for medical
image analysis, this survey presents an comprehensive
answering (VQA). Deep neural networks (DNNs) automat-
review across various image modalities and clinical appli- ically learn features from input data and produce optimal
cations. It covers more than 200 papers from three key outputs. However, the inherent complexity nature of DNNs
perspectives: 1) input explainability through the integration hinder our understanding of the decision-making processes
of explainable feature engineering and knowledge graph, 2) behind these models. Consequently, DNNs are often consid-
model explainability via attention-based learning, concept-
based learning, and prototype-based learning, and 3) out-
ered as black-box models, which has raised concerns about
put explainability by providing counterfactual explanation their transparency, interpretability, and accountability for their
and textual explanation. Additionally, this paper outlines successful deployment in real-world clinical applications [1].
the desired characteristics of explainability and existing To tackle the challenge of developing more trustworthy AI
evaluation methods for assessing explanation quality. Fi- systems, research efforts are increasingly focusing on various
nally, it discusses the major challenges and future research
directions in developing S-XAI for medical image analysis. eXplainable AI (XAI) methods, enhancing transparency [2],
fairness [3], and robustness [4]. However, most XAI methods
Index Terms— Self-eXplainable Artificial Intelligence (S- aim to generate explanations for the outputs of black-box
XAI), Medical Image Analysis, Input Explainability, Model
Explainability, Output Explainability, S-XAI Evaluation
AI models after they have been trained, a category known
as post-hoc XAI, as illustrated in Fig. 1 top. These methods
utilize additional explanation models or algorithms to provide
I. I NTRODUCTION insights into the decision-making process of the primary AI
Artificial intelligence (AI), particularly deep learning, has model. In the field of medical image analysis, commonly used
driven significant advancements in medical image analysis, post-hoc XAI techniques include feature attribution methods,
including applications in disease diagnosis, lesion segmenta- such as gradient-based approaches (e.g., LRP [5], CAM [6])
tion, medical report generation (MRG), and visual question and perturbation-based approaches (e.g., LIME [7], Kernel
SHAP [8]). Additionally, some methods explored concept
This work was supported by the Hong Kong Innovation and Technol- attributions, learning human-defined concepts from the internal
ogy Fund (Project No. MHP/002/22), HKUST (Project No. FS111) and
Research Grants Council of the Hong Kong (No. R6003-22 and T45- activations of DNNs (e.g., TCAV [9], CAR [10]). Post-hoc
401/22-N). XAI techniques are often model-agnostic, indicating that they
J. Hou, Y. Bie, H. Wang, and A. Tan are with the Department of can be flexibly applied to a variety of already-trained black-
Computer Science and Engineering, Hong Kong University of Science
and Technology, Hong Kong, China (email: [email protected]) box AI models.
S. Liu is with the Department of Engineering, Shenzhen MSU-BIT Since post-hoc explanations are generated separately from
University, Shenzhen, China (email: [email protected]) the primary AI model, several valid concerns have been
L. Luo is with the Department of Biomedical Informatics, Harvard
University, Cambridge, USA (email: luyang [email protected]) raised: 1) these explanations may not always be faithful to
H. Chen is with the Department of Computer Science and Engineer- the actual decision-making process of black-box models [11],
ing, Department of Chemical and Biological Engineering and Division of [12]; 2) they may lack sufficient detail to fully elucidate
Life Science, Hong Kong University of Science and Technology, Hong
Kong, China; HKUST Shenzhen-Hong Kong Collaborative Innovation the model’s functioning [13]. These limitations of post-hoc
Research Institute, Futian, Shenzhen, China. (email: [email protected]) XAI approaches are particularly problematic in high-stakes
2 GENERIC COLORIZED JOURNAL, VOL. XX, NO. XX, XXXX 2017
Input Explainability (Sec III) Model Explainability (Sec IV) Output Explainability (Sec V)
Fig. 3. Overview of Self-eXplainable AI (S-XAI) frameworks including input explainability, model explainability, and output explainability.
This survey concentrates on S-XAI methods for medical A. Explainable Feature Engineering
image analysis that allow models to inherently explain their
Feature engineering focuses on transforming raw images
own decision-making. As depicted in Fig. 3, we introduce a
into a more useful set of human-interpretable features. This
new taxonomy of S-XAI based on the three key components
process is crucial for traditional machine learning methods to
of DNNs.
achieve accurate predictions, but it can be time-consuming and
1) Input Explainability (Sec. III): Input explainability
demands significant domain expertise. In contrast, deep learn-
focuses on integrating additional explainable inputs with deep
ing models automatically extract features from raw images,
features of medical images obtained from various anatomical
simplifying the manual crafting process but often resulting
locations and modalities to produce final predictions. By
in reduced interpretability. A promising approach to enhance
incorporating external knowledge and context-specific infor-
input explainability is to incorporate explainable feature engi-
mation, the accuracy and reliability of these predictions can
neering into deep learning, which injects domain knowledge
be significantly improved.
2) Model Explainability (Sec. IV): Model explainability into the model, as shown in Fig. 4(a). This integration en-
aims to design inherently intepretable model architectures of hances the model’s interpretability by ensuring that the learned
DNNs. Instead of explaining a black-box model, transforming features are relevant and meaningful for clinical applications.
the model into an interpretable format enhances understanding Ultimately, this method improves model performance and
of how it processes information. offers valuable insights into the decision-making process.
3) Output Explainability (Sec. V): Output explainability A common strategy in explainable feature engineering is
refers to the model’s ability to generate not just predictions to combine both handcrafted and deep features from an input
for various medical image tasks but also accompanying expla- image to make final predictions [24], [25]. For example, Kapse
nations through an explanation generator. This capability aids et al. [24] introduce a self-interpretable multiple instance
in understanding the rationale behind the model’s predictions, learning (SI-MIL) framework that simultaneously learns from
facilitating informed medical decision-making. deep image features and handcrafted morphometric and spatial
The following sections summarize and categorize the most descriptors. They assess the local and global interpretability
relevant works on S-XAI methods applied to medical image of SI-MIL through statistical analysis, a user study, and key
analysis. Comprehensive lists of the reviewed S-XAI meth- interpretability criteria. Another line of approach involves
ods are provided, detailing the employed S-XAI techniques, incorporating interpretable clinical variables as additional in-
publication year, anatomical location, image modality, medical puts alongside the images, often utilizing multimodal learning
application, and the datasets used. techniques [26], [27]. For instance, Xiang et al. [26] introduce
OvcaFinder, an interpretable model that combines deep learn-
III. I NPUT E XPLAINABILITY ing predictions from ultrasound images with Ovarian–Adnexal
In this section, we will explore input explainability by Reporting and Data System scores provided by radiologists, as
integrating external domain knowledge, focusing on two key well as routine clinical variables for diagnosing ovarian cancer.
approaches, i.e., a) explainable feature engineering (Sec. III- This approach enhances diagnostic accuracy and explains the
A) and b) knowledge graph (Sec. III-B). As shown in Fig. 4, impact of key features and regions on the prediction outcomes.
these explainable inputs will interact with the deep features of Discussion: Although explainable feature engineering can
image inputs and be combined to support final predictions. be time-consuming, it brings valuable prior knowledge and en-
4 GENERIC COLORIZED JOURNAL, VOL. XX, NO. XX, XXXX 2017
(a) Explainable Feature Engineering foundation for medical decision-making, clinical research, and
healthcare analytics [31]–[33]. By harnessing the medical prior
Feature Engineering Clinical Variables E Encoder
knowledge encoded in the graph, AI models can gain valuable
Shape Local Body Temperature
P
insights, identify patterns, predict patient outcomes, assist in
Predictor
Color Textual Complete Blood Count diagnosis, recommend personalized treatments, and ultimately
… …
Fusion improve patient care and outcomes [34]–[38]. For example,
Handcrafted features
Liu et al. [36] and Huang et al. [37] develop KGs based on the
Image
professional perspective related to medical images to enhance
image understanding. Another way to utilize prior knowledge
E P “Pneumonia” is by collecting a large number of relationship triples to
create a domain-knowledge-enhanced medical VQA dataset.
For instance, Liu et al. [38] extract a set of 52.6K triplets
1) Prior Knowledge Graph 2) Data Knowledge Graph in the format < head, relation, tail > containing medi-
Ultrasound cal knowledge from OwnThink (https://www.ownthink.com).
Normal Other Finding CT X-ray
Text
They then use this external information to create SLAKE, a
Heart Spine Pleural …
large-scale, semantically annotated, and knowledge-enhanced
Cardiomegaly Scoliosis … …
Inter-data Intra-data bilingual dataset for training and testing Med-VQA systems.
Prior KGs enhance S-XAI models by integrating expert-
3) Hybrid Knowledge Graph
derived knowledge and medical facts, enabling these models
(b) Knowledge Graph to better understand key medical concepts and make more in-
formed predictions. However, the creation of these KGs largely
Fig. 4. Input explainability that incorporates (a) explainable feature depends on specialized expertise, making the process labor-
engineering (b) knowledge graph as additional inputs.
intensive. Furthermore, these KGs often lack the adaptability
required for analyzing dynamic clinical datasets.
hances the interpretability of deep learning models concerning 2) Data Knowledge Graph: A data KG differs from a prior
input features. Despite the increasing research in this area, KG in its construction methodology. While a prior KG relies
most studies prioritize accuracy improvements, with limited on expert insights and established medical facts, a data KG is
analysis given to the explainability. Additionally, effective built directly from the dataset itself. This means that instead of
information fusion and interaction poses a key challenge. relying solely on pre-existing knowledge, the data knowledge
graph leverages the inherent information contained within the
B. Knowledge Graph dataset. This approach allows the data KG to provide a unique
A knowledge graph (KG) is a structured representation of perspective and the potential to discover previously unknown
factual knowledge that captures relationships between entities relationships and correlations within the data [64]–[67]. There
in a specific area. It provides a way to organize and represent are two primary approaches to leveraging data knowledge
knowledge in a semantically rich and interconnected manner for enhancing the explainability of AI models: 1) extracting
and plays a crucial role in enhancing the interpretability of S- knowledge directly from the dataset [43], [45], [46], [49],
XAI models. Recently, integrating structured domain knowl- [51], [57], [61]. Liu et al. [49] employ a bipartite graph
edge into downstream tasks has attracted significant attention convolutional network to model the intrinsic geometric and
of both industry and academia [28]–[30]. This growing interest semantic relation of ipsilateral views, and an inception graph
stems from the recognition that leveraging domain knowledge convolutional network to model the structural similarities of
can greatly improve the performance and effectiveness of bilateral views. Huang et al. [61] develop a medical KG based
various applications. As shown in Fig 4(b), regarding medical on the types of diseases and questions concerned by patients
imaging analysis, the utilization of KG can be broadly cate- during their treatment process. 2) Transferring knowledge from
gorized into three categories: 1) prior KG, which serves as a pre-trained models. For example, Qi et al. [52] use a pre-
foundational resource that gathers existing domain expertise trained U-Net to segment lung lobes and then model both
and established medical knowledge; 2) data KG, which is the intra-image and inter-image relationships of these lobes
derived from the analysis of large-scale medical imaging and in-batch images through their respective graphs. Elbatel
datasets; and 3) hybrid KG, which combines the strengths of et al. [53] distill knowledge from pre-training models to small
both prior and data KGs for medical image analysis. models for disease classification.
1) Prior Knowledge Graph: A prior KG in the medical Overall, constructing a data KG involves leveraging the
domain is a specialized KG that captures and organizes facts inherent characteristics of the dataset itself to build a graph
information about medical concepts and their relationships. structure to assist S-XAI models. However, it is important to
It can be constructed from a multi-sources, including med- note that these methods often harbor inherent biases that can
ical literature, electronic health records, medical ontologies, vary significantly across different datasets.
clinical guidelines, and expert opinions. This graph serves as 3) Hybrid Knowledge Graph: A hybrid KG integrates both
a comprehensive repository of medical knowledge, encom- the prior KG and the data KG, representing an interactive
passing details about diseases, symptoms, treatments, medi- approach. The prior KG provides a static foundation of estab-
cations, anatomical structures, and more. It provides a vital lished medical facts, while the data KG utilizes dataset charac-
AUTHOR et al.: PREPARATION OF PAPERS FOR IEEE TRANSACTIONS AND JOURNALS (FEBRUARY 2017) 5
TABLE I
I NPUT EXPLAINABILITY METHODS BASED ON KNOWLEDGE GRAPH (KG). T HE ABBREVIATIONS HERE ARE CLS: CLASSIFICATION , DET:
DETECTION , MRG: MEDICAL REPORT GENERATION , VQA: VISUAL QUESTION ANSWERING .
teristics to dynamically update and enhance this foundational into a graph format is labor-intensive and costly, requiring
knowledge. By incorporating data-specific insights discovered constant updates and refinements to incorporate the latest
from the dataset, the hybrid KG allows for the integration of research findings, clinical guidelines, and emerging medical
new information and the refinement of existing knowledge. data to maintain up-to-date prior medical knowledge. Another
This dynamic updating process ensures that the KG remains challenge lies in the heterogeneity of medical image data. With
up-to-date and relevant. Consequently, the hybrid KG com- the continuous growth of medical image data, the variety of
bines the strengths of both the prior and data KG, offering a image modalities expands, complicating their effective inte-
more comprehensive and adaptable knowledge representation gration within KGs. Developing robust algorithms to extract
for S-XAI models in the medical field [35], [47], [48], [56], meaningful features from medical images and link them with
[58], [59], [62], [63]. For instance, Wu et al. [48] implement a relevant medical KGs remains an ongoing research endeavor.
triplet extraction module to extract medical information from
reports, combining entity descriptions with visual signals at the IV. M ODEL E XPLAINABILITY
image patch level for medical diagnosis. For the medical report
generation tasks, Li et al. [56] decompose medical report In this section, we present model explainability by design-
generation into explicit medical abnormality graph learning ing interpretable model architectures, such as attention-based
and subsequent natural language modeling. Each node in the learning (Sec. IV-A), concept-based learning (Sec. IV-B), and
abnormality graph represents a possible clinical abnormality prototype-based learning (Sec. IV-C).
based on prior medical knowledge, with the correlations
among these nodes encoded as edge weights to inform clinical A. Attention-based Learning
diagnostic decisions. Hu et al. [63] utilize large language
Attention-based learning aims to capture specific areas in an
models to extract labels and build a large-scale medical VQA
image that are relevant to the prediction task while suppressing
dataset, Medical-CXR-VQA. They then leverage graph neural
irrelevant regions based on feature maps. Therefore, it can
networks to learn logical reasoning paths based on this dataset
be naturally combined with S-XAI methods to provide visual
for medical visual question answering task.
explanations that enhance model decision-making [2], [19],
In summary, the construction of a hybrid KG relies on [68]. We categorize attention-based S-XAI models into 1)
prior knowledge and involves automatically adjusting the structure-guided attention models and 2) loss-guided attention
nodes or edges based on the data characteristics. This process models. As illustrated in Fig. 5, the former specifically designs
ensures that the KG remains aligned with the specific domain the attention structure and obtains model predictions directly
knowledge and captures the most relevant, data-specific in- from the attention map, while the latter constrains the attention
formation. It provides a comprehensive representation of both map using a loss function to ensure attention map align with
data-specific knowledge and prior knowledge, enhancing the an ideal interpretable distribution.
interpretability of S-XAI models. 1) Structure-guided Attention Model: As shown in the top
Discussion: The utilization of medical KGs in medical branch of Fig. 5, structure-guided attention models are char-
image analysis poses both challenges and promising oppor- acterized by the association between the attention structure
tunities. First, integrating diverse prior medical knowledge in the model and the components directly influencing the
6 GENERIC COLORIZED JOURNAL, VOL. XX, NO. XX, XXXX 2017
1) Structure-Guided Attention Models three directions, the 3D attention map can be visualized to
explain the model’s decision-making process. For the fast MRI
Visual Attention Explanations
reconstruction task, Huang et al. [77] propose a shifted win-
Attention Weighted dows deformable attention mechanism which uses reference
Image Combination points to impose spatial constraints on attention and directly
A=en>on Es>ma>on
combines the outputs from the attention modules of different
E P “Pneumonia” windows to produce the model’s reconstruction results.
Although structure-guided attention maps can provide ex-
Attention Mechanism planations for model predictions, they are still difficult to align
with clear human-understandable decision-making basis.
Generated Attention Reference
2) Loss-guided Attention Model: As shown in the bottom
Attention Loss Attention branch of Fig. 5, loss-guided attention models use interpretable
labels (i.e., reference attention maps) to construct loss func-
2) Loss-Guided A.en0on Models
tions that directly constrain the generated attention maps. This
Fig. 5. Attention-based learning, including 1) structure-guided and 2) method encourages the model to focus on areas that are un-
loss-guided attention models. X-ray images borrowed from [69]. derstandable and beneficial for making predictions. Benefiting
from lesion area annotations and professional analyses by
model’s predictions. This allows the generated attention map to doctors, which provide clear references for model decisions,
effectively explain the model’s predictions. Jetley et al. [70] is loss-guided attention learning techniques are commonly used
the first to introduce attention learning for XAI. They propose in medical image analysis.
an attention estimator which calculates feature compatibility Using ground-truth masks of regions of interest (RoIs) to
scores to weight feature maps as feature activation scores, guide the generation of attention maps is a widely adopted
which are then directly used as input for a linear classifier. approach in medical image classification [80], [83], [87]. For
This approach guides the model’s attention toward areas that instance, Yang et al. [80] directly optimize the attention maps
are more relevant to its decision-making while suppressing by a Dice loss, which encourages the model to focus on
irrelevant regions. Fukui et al. [71] present an Attention target areas that are highly relevant to the classification of
Branch Network (ABN), which replaces the fully connec- breast cancer microscopy images. To alleviate the challenge
tion layer of Class Activation Mapping (CAM) [6] using a of obtaining pixel-level annotations, Yin et al. [87] pre-train a
convolution layer to output class probabilities. There is also histological feature extractor to identify significant clinically
a perception branch to apply a classifier to the combination relevant feature masks, which are then used to guide and
of attention maps from the original features. Furthermore, Li regularize the attention maps. By considering the varying
et al. [72] propose a slot attention-based method in which contributions of histological features for classification, the
the attention output of each slot are directly processed and model can selectively focus on different features based on
summed up by a main block named SCOUTER to support the distribution of nuclei in each instance. In medical image
a specific category, eliminating the need for a linear classifier segmentation, labels corresponding to edges and shapes of
and further improving the model’s transparency. They also use specific regions are often reused to guide attention in learning
the output from the slot attention mechanism to represent the semantic information [102], [105], [109]. Sun et al. [102]
model’s final confidence for each category. Notably, positive combine spatial attention with the attention estimator in U-
and negative interpretations can be controlled through the Net decoders, enabling the model to interpret learned features
parameters in the loss function. This method demonstrates at each resolution. They also introduce a gated shape stream
improved interpretability in the glaucoma diagnosis task. alongside the texture stream, where the resulting shape atten-
Numerous studies have incorporated multiple attention tion maps are aligned with actual edges through binary cross-
mechanisms for medical image classification and segmentation entropy loss, enhancing the cardiac MRI segmentation.
tasks [73]–[75]. For example, Schempler et al. [73] extend Compared with lesion masks, eye tracking data provides a
the attention estimator by extracting local information from more accurate depiction of expert focus, as it captures the way
coarse-scale feature maps for attention gates, facilitating more doctors visually process information during diagnosis. Bhat-
fine-grained visual interpretation for lesion segmentation or tacharya et al. [69] leverage the captured doctors’ attention to
ultrasound diagnosis. Similarly, Gu et al. [74] develop a guide model training. They employ a teacher-student network
comprehensive attention module that enhances model inter- to replicate the visual cognitive behavior of doctors when
pretability through spatial, channel, and scale attention. Their diagnosing diseases on chest radiographs. The teacher model is
segmentation experiments on skin lesions and fetal organs trained based on the visual search patterns of radiologists, and
demonstrate improved performance and better interpretability the student model utilizes an attention loss to predict attention
of target area positioning and scale. Beyond 2D data, how from the teacher network using eye tracking data.
to use attention to explain more complex 3D medical image Discussion: Attention-based S-XAI methods guide model
diagnosis is more challenging. Lozupone et al. [76] present predictions by focusing on critical areas of images, thereby
an attention module that fuses attention weights from sagittal, providing effective attention explanations. Structure-guided
coronal, and axial slices to diagnose Alzheimer’s disease on attention models typically utilize the attention-weighted output
3D MRI brain scans. By integrating these attention scores from as input for the predictor, reflecting the model’s decision-
AUTHOR et al.: PREPARATION OF PAPERS FOR IEEE TRANSACTIONS AND JOURNALS (FEBRUARY 2017) 7
TABLE II
M ODEL EXPLAINABILITY METHODS BASED ON ATTENTION - BASED LEARNING . T HE ABBREVIATIONS HERE ARE CLS: CLASSIFICATION , SEG:
SEGMENTATION , IRE: IMAGE RECONSTRUCTION , REG: REGRESSION
…
the decisions still require further investigation. (LLM/VLM) NO
Infiltration
2) Unsupervised Automatic
3) Generative Concept Learning
Concept Discovering
B. Concept-based Learning
Fig. 6. Concept-based learning, including 1) supervised concept learn-
Concept-based S-XAI methods provide explanations in ing, 2) unsupervised automatic concept discovering, and 3) generative
terms of high-level, human-interpretable attributes rather than concept learning.
low-level, non-interpretable features. This approach reveals
the inner workings of deep learning models using easily
understandable concepts, enabling users to gain deeper insights explicitly described in natural language. For example, Sun et
into underlying reasoning. It also helps in identifying model al. [122] consider the instances segmented by SAM [123] as
biases and allows for adjustments to enhance performance and the concepts of a given image.
trustworthiness. Most concept-based S-XAI methods focus on 1) Supervised Concept Learning: Supervised concept learn-
making decisions based on a set of concepts while also de- ing methods train deep learning models using annotations of
tailing the contribution of each concept to the final prediction textual concepts, particularly by supervising an intermediate
[113]–[116]. These methods introduce concept learning into layer to represent these concepts. A notable example is Con-
the training pipeline of the models, instead of simply analyzing cept Bottleneck Model (CBM) [113], which is an inherently
explainability after training a black-box model (i.e., post-hoc interpretable deep learning architecture. It first maps latent
XAI methods) [117]–[119]. We propose to categorize concept- image features to a concept bottleneck layer, where the number
based S-XAI methods into three types: 1) supervised concept of neurons corresponds to the number of human-defined
learning, 2) unsupervised automatic concept discovering, and concepts, and then predicts final results based on the concept
3) generative concept learning, as shown in Fig. 6. scores from this layer. By enforcing the neurons in the concept
The term Concept has been defined in different ways, which bottleneck layer to learn concept representations supervised by
commonly represents high-level attributes [117], [120], [121]. concept labels, CBMs can directly show each concept’s contri-
In this paper, we suggest adopting a straightforward and easily bution to the final prediction (i.e., class-concept relation) using
understandable categorization: Textual Concepts and Visual the neuron values of the last layer. Specifically, the authors
Concepts. Textual Concepts refer to textual descriptions of of CBM conduct experiments on the knee X-ray dataset OAI
attributes associated with the classes. For example, in Fig. [124] to explore the importance of concepts such as bone spurs
6, the textual concepts for the classes (i.e., “pneumonia” and and calcification in determining arthritis grading. Additionally,
“normal”) include terms like Opacity, Effusion, Infiltration, CBMs allow model editing. When domain experts find certain
etc. Visual Concepts, on the other hand, consist of semanti- predicted concept importance unreasonable, they can easily
cally meaningful features within the image that may not be adjust the model’s predictions by intervening in the weights
8 GENERIC COLORIZED JOURNAL, VOL. XX, NO. XX, XXXX 2017
TABLE III
M ODEL EXPLAINABILITY METHODS BASED ON CONCEPT- BASED LEARNING . T HE ABBREVIATIONS HERE ARE CLS: CLASSIFICATION
of the concept bottleneck layer (test-time intervention). The concepts may not be associated with human-specified textual
CBM architecture has inspired many researchers to develop concepts. However, these methods can still provide concept-
inherently interpretable methods, resulting in a series of based explanations by visualizing the unsupervised concepts
CBM-like methods. For example, Concept Embedding Models and detailing their contributions to the final predictions. For
(CEMs) [159] utilize a group of neurons (concept embeddings) instance, Ghorbani et al. [162] propose Automatic Concept-
instead of a single neuron to represent a concept, which based Explanations (ACE), which automatically extract visual
effectively improves the performance of the original CBM concepts that are meaningful to humans and important for
while preserving its interpretability. Different from CBMs, the network’s predictions. Self-Explaining Neural Networks
Concept Whitening [139], [160] aims to whiten the latent (SENN) [4] first utilize a concept encoder to extract clusters
space of neural networks and aligns the axes of the latent of image representations corresponding to different visual
space with known concepts of interest. Zhao et al. [145] concepts, and also adopt a relevance parametrizer to calculate
introduce a hybrid neuro-probabilistic reasoning algorithm the relevance scores of concepts. The final prediction is
for verifiable concept-based medical image diagnosis, which determined by the combination of discovered concepts and
combines clinical concepts with a Bayesian network. the corresponding relevance scores. Inspired by SENN, Sarkar
The self-explainable nature of concept-based learning mod- et al. [163] propose an ante-hoc explainable framework that
els has led to its application in medical image analysis. includes both a concept encoder and a concept decoder, which
Chauhan et al. [125] propose Interactive CBMs, which can map images into concept space and use the concepts to
request labels for certain concepts from a human collaborator. reconstruct the original images, respectively. Yeh et al. [164]
This method is evaluated on chest and knee X-ray datasets. argue that the discovered concepts may not be sufficient to
Yan et al. [129] discover and eliminate confounding concepts explain model predictions, so they define a completeness score
within datasets using spectral relevance analysis [161], and to evaluate whether the concepts adequately support model
conduct experiments on skin image datasets. Marcinkevics predictions and propose a framework for complete concept-
et al. [144] adapt CBM for prediction tasks with multiple based explanations.
views of ultrasonography and incomplete concept sets. Kim et Since medical concept annotations are costly and require
al. [132] present a medical concept retriever, which connects experts’ efforts, unsupervised automatic concept discovering
medical images with text and densely scores images on is usually adopted to offer concept-based explanations in
concept presence. This enables important tasks in medical AI medical image analysis without expert-annotated labels. For
development and deployment, such as data auditing, model example, Fang et al. [147] address the practical issue of
auditing, and model interpretation, using a CBM architecture classifying infections by proposing a visual concept mining
to develop an inherently interpretable model. (VCM) method to explain fine-grained infectious keratitis
However, a significant challenge in supervised concept images. Specifically, they first use a saliency map based po-
learning is the scarcity of concept annotations, which require tential concept generator to discover visual concepts, and then
labor-intensive efforts from human experts. Therefore, some propose a visual concept-enhanced framework that combines
researchers prefer unsupervised automatic concept discover- both image-level representations and the discovered concept
ing, as it eliminates the need for extra annotations. features for classification. Moreover, Kong et al. [152] develop
2) Unsupervised Automatic Concept Discovering: Models a novel Attribute-Aware Interpretation Learning (AAIL) model
that perform unsupervised concept discovery modify their to discover clinical concepts, and then adopt a fusion module
internal representations to identify concepts within image fea- to integrate these concepts with global features for thyroid
tures without relying on explicit annotations. These discovered nodule diagnosis from ultrasound images.
AUTHOR et al.: PREPARATION OF PAPERS FOR IEEE TRANSACTIONS AND JOURNALS (FEBRUARY 2017) 9
Image
Although unsupervised automatic concept discovering can
offer concept-based explanations, these explanations are ab-
“Nodule”
stract and usually cannot be directly described in natural
language. To alleviate this issue while also addressing the
Prototypical Images
lack of concept annotations, generative concept learning has Compare Because these areas look
similar to typical Nodule images.
become a promising research direction. E
3) Generative Concept Learning: Leveraging foundation 1) Explicit prototype 2) Implicit prototype
Cardiomegaly
models, such as Large Language Models (LLMs) and Vision- extracted from a
certain typical image.
close to features extracted
from a set of typical images.
…
Language Models (VLMs), can assist in generating and la-
…
+ + + ** +
+ + *
beling textual concepts. A notable generative concept learning + ++ * * *
* ** + + ** *
+ + **
method, namely Language Guided Bottlenecks (LaBo) [165], Nodule + * image features prototype
employs an LLM (GPT-3 [166]) to generate textual concepts
for each image category, which are filtered to form the concept Fig. 7. Prototype-based learning, including 1) explicit prototype and 2)
bottleneck layer. LaBo then uses a pre-trained VLM (CLIP implicit prototype. X-ray images borrowed from [170].
[167]) to calculate the similarity between input images and
the generated concepts to obtain concept scores. The final image and then compare the feature maps with the proto-
prediction is based on the multiplication of a weight matrix types to calculate similarities. Ultimately, these similarities
and these concept scores. Label-free CBM [168] employs are combined for the final decision making. This process
a similar pipeline, but trains an independent network that is considered interpretable because the decision making can
includes a concept bottleneck layer. In the medical domain, be clearly attributed to the contribution of each interpretable
Kim et al. [140] enhance LaBo [165] by incorporating a prototype (e.g., by the similarity scores). According to how
more fine-grained concept filtering mechanism and conducted the prototypes are obtained, we define and categorize them to
explainability analysis on dermoscopic images, achieving per- two types: 1) explicit prototypes and 2) implicit prototypes,
formance improvements compared to the baseline. Similarly, as presented in Fig. 7. Explicit prototypes are specific high-
Liu et al. [153] employ ChatGPT and CLIP for explainable dimensional feature representations extracted from certain
zero-shot disease diagnosis on X-ray and CT. Bie et al. [169] training images, whereas implicit prototypes are latent high-
propose an explainable prompt learning framework that lever- dimensional representations that are close to a set of typical
ages medical knowledge by aligning the semantics of images, images’ representations. All existing prototype-based S-XAI
learnable prompts, and clinical concept-driven prompts at mul- models do not require supervision at the prototype level and
tiple granularities, where the category-wise clinical concepts aim to automatically find meaningful prototypes to facilitate
are obtained by eliciting knowledge from LLMs. interpretable decision making.
Discussion: Methods that provide concept-based explana- 1) Explicit prototype based models: The first model of this
tions hold significant importance in medical research and ap- type is ProtopNet [171], which introduces a three-stage train-
plications, particularly in advancing evidence-based medicine. ing scheme that is widely adopted by subsequent research:
By offering human-understandable explanations, these meth- 1) Feature extractor training: in this step, the final layer is
ods have the capability to help doctors and patients better un- frozen, and only the feature extraction backbone is trained.
derstand AI-assisted diagnosis, hence holding the potential to 2) Prototype replacement: this step replaces the learned rep-
make AI technologies effectively supported and disseminated resentations in the prototype layer with the nearest feature
in healthcare. The lack of fine-grained label annotations and patch from the training set. 3) Final layer fine-tuning: in this
the performance-explainability trade-off are the limitations of stage, the feature extractor remains fixed while the parameters
concept-based methods. Thanks to the development of LLMs, of the final layer are fine-tuned. Later works closely follow
researchers are exploring new ways to alleviate these issues, this training scheme while addressing different limitations of
e.g., generative concept learning [140], [153], [165]. In addi- this initial framework. For example, ProtoShare [191] proposes
tion, as there are more and more medical foundation models to share prototypes across different classes to reduce the
being developed, incorporating the knowledge of the models overall number of prototypes and enhance model efficiency. A
and medical experts to efficiently annotate concept labels for similar idea is explored in ProtoPool [192], where prototypes
datasets will be a promising and meaningful direction. Besides are assigned to various classes in a differentiable manner.
the most popular classification task, other medical applications To address the limitation of prior models that use spatially
of concept-based approaches should be further explored. rigid prototypes, ProtoDeform [193] proposes to additionally
learn an offset to obtain prototypes that are more spatially
flexible. TesNet [194] leverages the Grassman manifold to con-
C. Prototype-based Learning struct a transparent embedding space, achieving competitive
Prototype-based S-XAI models aims to provide a decision- accuracy. Inspired by the theory of support vector machines,
making process where a model reasons through comparisons ST-protopnet [195] aims to further improve the accuracy of
with a set of interpretable example prototypes [171]. This prototype-based models by separating prototypes into support
reasoning aligns with human recognition patterns, as humans and trivial prototypes, where support prototypes are located
often identify objects by comparing them to example compo- near the decision boundary in feature space, while trivial ones
nents [172]. These models first extract features from a given lie far from it. To investigate the hierarchical relationships
10 GENERIC COLORIZED JOURNAL, VOL. XX, NO. XX, XXXX 2017
TABLE IV
M ODEL EXPLAINABILITY METHODS BASED ON PROTOTYPE - BASED LEARNING . T HE ABBREVIATIONS HERE ARE CLS: CLASSIFICATION , REG:
REGRESSION .
between classes, Hase et al. [196] propose hierarchical pro- users to identify which specific property is important in the
totypes to offer explanations according to class taxonomy. As corresponding image patch (e.g., is it the color or texture that
prototype-based models are mostly based on linear classifiers, matters in this prototypical area?). This issue can be partially
ProtoKnn [197] explores the usage of k nearest neighbors mitigated using implicit prototype based models.
as a classifier and offers counterfactual explanations within 2) Implicit prototype based models: This type of model
the prototype-based framework. Recognizing the importance follows a similar training scheme as the models based on
of interpretability methods for debugging models, ProtoDe- explicit prototypes, with the major difference in avoiding the
bug [198] proposes an approach where a human supervisor prototype replacement step, or only projecting the prototype
can provide feedback to the discovered prototypes and learn to the training images’ feature patches for visualizations. This
confounder-free prototypes. scheme is simpler than one that includes prototype replace-
Adopting prototype-based S-XAI models in the medical ment step and has different interpretability benefits. Li et al.
domain presents additional challenges. Unlike natural images [199] propose the earliest work using latent prototypes, which
where the representative prototype occupies an area with a leverages a decoder to visualize the meanings of the learned
relatively stable size, medical image features such as disease prototypes. Protoeval [200] designs a set of loss functions to
regions in chest X-ray images can vary significantly in size. encourage the learned latent prototypes to be more stable and
To address this, XProtoNet [170] proposes to predict an occur- consistent across different images. To address the issue of the
rence map and summing the similarity scores within those ar- same prototype potentially representing different concepts in
eas, rather than relying solely on the maximum similarity score the real world, Nauta et al. [201] introduce PIP-net which
as done in ProtopNet. Similarly, [173] introduces prototypes learns prototypes by encouraging the augmented two views of
with square and rectangular spatial dimensions for COVID- the same image patch to be assigned to the same prototype.
19 detection in chest X-rays. In evaluations of ProtopNet, To help users identify the specific properties in an image that
Mohammadjafari et al. [175] observe a performance drop for contribute to the final classification (e.g., color or texture),
Alzheimer’s disease detection using MRI, whereas Carlon et instead of allowing users to observe only one example image
al. [178] report a high-level of interpretability satisfaction from patch per prototype, Ma et al. [202] propose to illuminate
radiologists in breast mass classification using mammograms. prototypical concepts via multiple visualizations. Due to the
In mammogram based breast cancer diagnosis, Wang et al. interpretability benefits of decision trees, ProtoTree [203]
[180] propose to leverage knowledge distillation to improve explores the incorporation of decision trees into prototype-
model performance. To overcome the confounding issue in based models, using latent prototypes as the nodes throughout
mammogram based mass lesion classification, Barnett et al. the decision-making process. Recently, to address the concern
[177] employ a multi-stage framework that identifies the mass that prototype-based models often underperform their black
margin features for malignancy prediction, skipping image box counterparts, Tan et al. [204] develop an automatic
patches that have already been used in previous prototypes prototype discovery and refinement strategy to decompose
during the prototype projection step to improve prototype the parameters of the trained classification head and thus
diversity. In brain tumor classification, MProtoNet [182] intro- guarantees the performance.
duces a new attention module with soft masking and online- Discussion: In terms of performance, implicit prototype
CAM loss applied in 3D multi-parametric MRI. To predict the based models generally outperform explicit ones, probably
brain age based on MR and ultrasound images, Hesse et al. due to the greater flexibility in prototype learning. Regarding
[188] utilize the weighted mean of prototype labels. Addition- interpretability, both types of models offer unique advantages.
ally, INSightR-Net [184] formulates the diabetic retinopathy For example, explicit prototypes can be intuitively explained
grading as a regression task and apply the prototype based through one-to-one mappings to the input image, while im-
framework, while ProtoAL [186] explores an active learning plicit prototypes can be explained using a diverse set of images
setting for prototype-based models in diabetic retinopathy. with similar activations. However, in medical image analysis,
Although these models offer interpretability in a one-to-one current prototype-based S-XAI models primarily utilize ex-
mapping to the input image, they can also make it difficult for plicit prototypes. Therefore, investigating the use of implicit
AUTHOR et al.: PREPARATION OF PAPERS FOR IEEE TRANSACTIONS AND JOURNALS (FEBRUARY 2017) 11
Image Image
E P “Pneumonia” E P “Mass”
Language Model patchy opacities silhouette, present, right [Cardiac silhouette]: Patchy
heart border / pleural effusions, absent opacities silhouette the
Query: What’s right heart border. [Left
3) Free-structured text lower lung zone]: Minimal Encoder Decoder “Healthy”
the findings of left lower lung atelectasis is
Foundation Model Patchy opacities silhouette the right heart
the given x-ray noted. [Mediastinum]: The mediastinal 𝑧
(LLM/VLM) border. The mediastinal contours are
image? normal. There are no pleural effusions. contours are normal. Generative Model Counterfactual Image
Fig. 8. Output explainability that provides (a) textual explanations, including fully-structured, semi-structured, and free-structured text; and (b)
counterfactual explanations. The difference between the generated counterfactual image and raw image (red box) indicates the explanation. X-ray
images borrowed from [205].
prototypes in the medical domain could be a promising avenue certain types of cell attributes along with a concluding
for future research. statement. Additionally, Wang et al. [214] introduce a
hierarchical framework for medical image explanation, which
V. O UTPUT E XPLAINABILITY first predicts semantically related topics and then incorporates
these topics as constraints for the language generation model.
This section discusses output explainability by generating
In the context of hip fracture detection from pelvic X-rays,
explanations alongside model predictions, including textual
Gale et al. [215] utilize a visual attention mechanism to
(Sec. V-A) and counterfactual (Sec. V-B) explanations.
create terms related to location and characteristics, which
are then used to generate sentences structured as: “There is
A. Textual Explanation a [degree of displacement], [+/- comminuted][+/- impacted]
Textual explanations in S-XAI involve generating human- fracture of the [location] neck of femur [+/- with an avulsed
readable descriptions that accompany model predictions as fragment].” More recently, some studies have focused on
part of outputs, similar to image captioning. These methods generating individual sentences based on anatomical regions
use natural language to clarify model decisions and typically [216]–[219]. For example, Tanida et al. [218] introduce
require textual descriptions for supervision. Some studies a Region-Guided Radiology Report Generation (RGRG)
explore the integration of textual explanations with visual ones. method that identifies unique anatomical regions in the chest
We categorize these methods into three types based on the and generates specific descriptions for the most salient areas,
structure of textual explanations: 1) fully-structured, 2) semi- ensuring each sentence in the report is linked to a particular
structured, and 3) free-structured text, as shown in Fig. 8(a). anatomical region. Overall, semi-structured approaches
1) Fully-structured text generation: To address the chal- effectively balance the rigidity of fully structured reports with
lenges posed by complex unstructured medical reports, early the inconsistency of completely free-text reports.
efforts transformed target texts into fully structured formats, 3) Free-structured text generation: With the advancement of
such as descriptive tags, attributes, or fixed templates, rather language models, reports generated for a given input image
than natural language. For example, Pino et al. [206] propose are no longer limited to structured formats; instead, they now
CNN-TRG, which detects abnormalities through multilabel focus on more open, free-structured text descriptions. These
classification and generates reports based on pre-defined tem- approaches typically involve combining an image encoder to
plates. Some works utilize controlled vocabulary terms (e.g., extract visual features with a language model to produce co-
Medical Subject Headings (MeSH) [207]) to describe image herent sentences [221]. Several research efforts provide com-
content instead of relying on unstructured reports. Both Shin et prehensive explanations that include both textual and visual
al. [208] and Gasimova et al. [209] employ CNN-RNN frame- justifications for diagnostic decisions [222], [224]–[226]. For
works to identify diseases and generate corresponding MeSH instance, Spinks and Moens [222] propose a holistic system
sequences, detailing location, severity, and affected organs in that delivers diagnosis results along with generated textual
chest X-ray images. In addition, Rodin et al. [210] present a captions and a realistic medical image representing the closest
multitask and multimodal model to produce a short textual alternative diagnosis as visual evidence. Additionally, Wang
summary structured as “[pathology], [present/absent], [(op- et al. [226] explore a multi-expert Transformer to generate
tional) location], [(optional) severity]”. However, complete de- reports and attention-mapping visualization of key medical
scriptions in natural language are more human-understandable terms and expert tokens.
than a set of simple tags, leading several studies to focus on In addition to directly generating medical reports, some
generating reports in a semi-structured format. research studies have incorporated the classification of patho-
2) Semi-structured text generation: Generating semi- logical terms or tags in two distinct ways. The first approach
structured text involves a partially structured format with utilizes a “classification-report generation” pipeline, integrat-
predefined topics and constraints in the medical report ing a classifier within the report generation network to enhance
generation process. For instance, pathology report generation feature representations [227], [228]. For example, Yuan et al.
methods [211]–[213] produce reports that focus on describing [227] further employ a sentence-level attention mechanism
12 GENERIC COLORIZED JOURNAL, VOL. XX, NO. XX, XXXX 2017
TABLE V
O UTPUT EXPLAINABILITY METHODS THAT PROVIDE TEXTUAL EXPLANATIONS . T HE ABBREVIATIONS HERE ARE MRG: MEDICAL REPORT
GENERATION , CLS: CLASSIFICATION , LOC: LOCATION , SEG: SEGMENTATION , VQA: VISUAL QUESTION ANSWERING , V IS : V ISUAL EXPLANATION .
alongside a word-level attention model to analyze multi-view analyze and respond to open-ended questions about the input
chest X-rays, using predicted medical concepts to improve the images, thanks to their pretraining on extensive datasets of
accuracy of medical reports. Conversely, the second approach image-report pairs. For instance, XrayGPT [237] demonstrates
employs a “report generation-classification” pipeline, leverag- the alignment of a medical visual encoder (MedClip) with
ing interpretable region-of-interest (ROI) characterization for a fine-tuned LLM (Vicuna) using a linear transformation.
final diagnoses. For instance, Zhang et al. [229] construct Given an input image, this combined model can address open-
a pathologist-level interpretable diagnostic framework that ended questions, such as “What are the main findings and
first detects tumour regions in whole slide images (WSIs), impressions from the given X-ray?”. These models not only
then generates natural language descriptions of microscopic excel in medical image captioning but also demonstrate ex-
findings with feature-aware visual attention, and finally estab- ceptional capability in delivering comprehensive explanations
lishes a diagnostic conclusion. Moreover, integrating region for a wide range of medical inquiries. By leveraging their
localization and lesion segmentation can enhance the quality extensive knowledge and understanding, they contribute to
of textual explanations [231], [232], [234], [235]. For instance, the generation of detailed and informative textual explanations
Wang et al. [231] develop a Text-Image Embedding network within the medical field.
(TieNet) that incorporates multi-level attention to highlight Discussion: Textual explanations have demonstrated signifi-
meaningful text words and X-ray image regions for disease cant effectiveness in providing human-interpretable judgments
detection and reporting. Leveraging fine-grained annotations of through natural language. This type of S-XAI approach has
segmentation masks or bounding boxes for lesions, Tian et al. become especially valuable with the advancement of language
[235] combine a segmentation model with a language model, models, enabling the generation of lengthy reports and the
creating a multimodal framework with a semi-supervised at- ability to answer open-ended questions. However, it is crucial
tention mechanism for CT report generation. to enhance the quality and reliability of these generated textual
explanations. Some recent studies utilize techniques such as
Compared to traditional report generation approaches, the
knowledge decoupling [247] and instruction tuning [249] to
utilization of LLMs offers a more interactive and comprehensi-
address challenges like hallucination, thereby improving the
ble method for generating textual explanations. Recent medical
effectiveness and trustworthiness of textual explanations in
VLMs applied to various medical images, such as chest X-rays
medical applications.
(e.g., XrayGPT [237]), skin images (e.g., SkinGPT [238]), and
general medical images (e.g., Med-flamingo [240], LLaMa-
Med [242], MedDr [244], HuatuoGPT-Vision [245]), can
AUTHOR et al.: PREPARATION OF PAPERS FOR IEEE TRANSACTIONS AND JOURNALS (FEBRUARY 2017) 13
TABLE VI
D ESIRABLE QUALITIES OF EXPLANATION METHODS , INDIVIDUAL EXPLANATIONS , AND HUMAN - FRIENDLY EXPLANATIONS .
Explainability Computational Complexity The computational complexity of explanation algorithms. [258], [259]
Methods Generalizability, Portability To increase the utility because of the diversity of model architectures. [259]
[264]. For example, Sayres et al. [266] investigate the impact fixation is an emerging data modality that can provide key
of a deep learning model on doctors’ performance in predict- diagnostic features by tracking the gaze patterns and visual
ing diabetic retinopathy (DR) severity. Ten ophthalmologists attention of clinicians, which is also utilized as the ground
with varying levels of experience read images under three truth of attention maps [271].
conditions: unassisted, predicted grades only, and predicted Concept-based explanation: To evaluate concept-based ex-
grades with heatmaps. The results indicate that AI assistance planations, researchers mainly focus on metrics such as Con-
improves diagnostic accuracy, subjective confidence, and time cept Error [113], [163], T-CAV score [117], Completeness
spent. However, in most cases, the combination of grades Score [164], and Concept Relevance [4], [120]. Additionally,
and heatmaps is only as effective as using grades alone, and other evaluation methods exist. For example, Zarlenga et al.
actually decreased accuracy for patients without DR. [159] propose Concept Alignment Score (CAS) and Mutual
Overall, human-centered evaluations offer the significant ad- Information to evaluate concept-based explainability. Wang et
vantage of providing direct and compelling evidence of the al. [195] adopt Concept Purity to assess the model’s capability
effectiveness of explanations [263]. However, these evaluations to discover concepts that only cover a single shape.
tend to be costly and time-consuming due to the need to recruit Example-based explanation: In the evaluation of example-
expert participants and obtain necessary approvals, as well as based explanations, Nguyen and Martinez [272] establish two
the additional time required for conducting the experiments. quantitative metrics: 1) non-representativeness, which evalu-
Most importantly, these evaluations are inherently subjective. ates how well the examples represent the explanations, thereby
2) Functionality-grounded evaluation: This category of eval- measuring the fidelity of the explanation; and 2) diversity,
uation, which do not involve human-subject investigations, can which gauges the degree of integration within the explanation.
be employed to assess the fidelity of explanations. The accu- Additionally, Huang et al. [200] developed two metrics: 1)
racy of S-XAI methods in generating genuine explanations a consistency score to determine whether the prototype con-
is referred to as the fidelity of an explainer. In this section, sistently highlights the same semantically meaningful areas
we will present a variety of functionality-grounded evaluation across different images, and 2) a stability score to assess
methods for different types of explanations. whether it reliably identifies the same area after the image
Attention-based explanation: In the absence of references, is exposed to noise.
attention-based explanations can be assessed through a causal Textual explanation: The common assessment of textual
framework. For example, Petsiuk et al. [267] introduce two explanations involves using metrics such as BLEU [273],
causal metrics, i.e., deletion and insertion. Following this, ROUGE-L [274], and CIDEr [275] to compare generated
Hooker et al. [268] propose RemOve And Retrain (ROAR), natural language descriptions against ground truth reference
a method that evaluates how the accuracy of a retrained sentences provided by experts. Patricio et al. [17] conduct a
model decreases when essential features in specific regions benchmark study of interpretable medical imaging approaches,
are removed. With the manually annotated ground truth data, specifically evaluating the quality of textual explanations for
such as object bounding boxes or semantic masks, the accuracy chest X-ray images.
of attention-based explanations can be evaluated by comparing Counterfactual explanation: Singla et al. [276] employ three
with these references. Yan et al. [83] and Hou et al. [269] cal- metrics to evaluate counterfactual explanations for chest X-ray
culate the Jaccard index value and the AUC score to measure classification: 1) Frechet Inception Distance (FID) to assess
the effectiveness of attention maps, respectively. Additionally, visual quality, 2) Counterfactual Validity (CV) to determine if
Barnett et al. [270] introduce the Activation Precision metric, the counterfactual aligns with classifier’s predictions, and 3)
which quantifies the proportion of relevant information from Foreign Object Preservation (FOP), which examines whether
the relevant region used to classify the mass margin based patient-specific information is retained. Additionally, they use
on radiologist annotations. Furthermore, human expert eye clinical metrics, including the cardiothoracic ratio and a score
AUTHOR et al.: PREPARATION OF PAPERS FOR IEEE TRANSACTIONS AND JOURNALS (FEBRUARY 2017) 15
for detecting normal costophrenic recess, to illustrate the enhance the input explainability through explainable prompts
clinical utility of the explanations. [284] and knowledge-enhanced prompts [285], ultimately im-
proving model performance.
VII. C HALLENGES AND F UTURE D IRECTIONS 2) Foundation models advance S-XAI: Foundation models
Despite the rapid advancements in S-XAI for medical image learn generally useful representations from the clinical knowl-
analysis, several significant challenges remain unresolved. In edge embedded in medical corpora [286]. By harnessing the
this section, we will analyze the existing challenges and sophisticated capabilities of foundation models, S-XAI meth-
discuss potential future directions to enhance the effectiveness ods can produce user-friendly explanations [287] and support
and reliability of S-XAI in the medical domain. more flexible generative concept-based learning [288], [289].
Moreover, foundation models can facilitate the evaluation of
A. S-XAI Benchmark Construction S-XAI methods that emulate human cognitive processes [290].
Establishing benchmarks for S-XAI in medical image anal-
ysis. is essential. These benchmarks will standardize evalua- C. S-XAI with Human-in-the-Loop
tions, enable fair comparisons between different methods, and
ultimately enhance the reliability of medical AI applications. Integrating Human-in-the-Loop (HITL) processes is crucial
1) Dataset construction: One of the main challenges in for effectively implementing S-XAI in the medical field. This
collecting medical data is the limited availability of doctors to approach not only enhances the overall performance of AI
annotate large datasets. This challenge is even more significant systems but also fosters trust among medical experts.
in S-XAI, where additional fine-grained annotations, such as 1) Enhancing prediction accuracy through human interven-
concepts and textual descriptions, are necessary. As a result, tion: A HITL framework allows for the identification and
medical datasets that meet interpretability standards often have removal of potential confounding factors, such as artifacts
a limited volume of data, reducing the generalizability and or biases in datasets, during the training phase. For instance,
applicability of S-XAI methods in real-world contexts. clinicians can adjust the outputs of predicted concepts, leading
2) Evaluation metrics: Automated evaluation of explana- to a more accurate concept bottleneck model [129]. This
tions generated by S-XAI methods poses another significant collaborative approach can significantly enhance the model’s
challenge. In the medical field, human-centered evaluations of- accuracy by incorporating expert insights.
ten rely on the expertise of clinicians. However, the variability 2) Improving explainability through human feedback: To en-
in expert opinions can lead to biased and subjective assess- sure continuous improvement, a versioning or feedback evalu-
ments [277]. Meanwhile, existing functionality-grounded eval- ation system should be established, enabling the final system to
uations still depend on manual annotations. Thus, developing build trust during hospital evaluations. Achieving this requires
objective metrics to evaluate the quality of model explanations fostering collaboration between S-XAI researchers and clinical
is likely to become an important research focus. practitioners, ensuring that feedback is systematically gathered
To tackle these challenges, future directions include leverag- and used to refine the models.
ing semi-automated annotation tools to assist clinicians in the However, one challenge in integrating HITL processes is
annotation process, thereby easing their workload. Addition- the variability in clinician expertise and availability, which
ally, developing objective metrics and standardized protocols can affect the consistency and quality of human feedback.
to assess the quality of model explanations will be a critical Ensuring that human knowledge is effectively integrated into
research trend in S-XAI. the AI training process without introducing additional biases
or errors is a complex task.
B. S-XAI in the Era of Foundation Models
Foundation models, including large language models
D. Trade-off between Performance and Interpretability
(LLMs) and vision-language models (VLMs), have trans-
formed the AI landscape, finding applications across diverse It is widely believed that as model complexity increases
fields such as natural language processing, computer vision, to enhance performance, the model’s interpretability tends to
and multimodal understanding. Notably, medical LLMs [278]– decline [291], [292]. Conversely, more interpretable models
[280] and medical VLMs [237], [240], [281] are designed may sacrifice some predictive accuracy. However, it is im-
to encode rich domain-specific knowledge. The intersection portant to note that some researches contend that there is no
of medical foundation models and S-XAI presents significant scientific evidence for a general trade-off between accuracy
opportunities for the future of medical AI systems [262]. and interpretability [293]. In fact, recent advancements in
1) S-XAI benefits foundation models: Foundation models are concept-based models [132], [139], [142] have demonstrated
typically large models with an extremely huge number of performance on par with black-box models in medical image
parameters, trained on vast datasets. The complexity of these applications. This achievement depends on the researcher’s
models make it challenging to explore their decision-making ability to identify patterns in an interpretable manner while
processes, which may result in potential biases and a lack of maintaining the flexibility to accurately fit the data [13]. Future
transparency. Apart from leveraging post-hoc XAI techniques S-XAI methods are expected to aim for an optimization of
(e.g., attribution maps [282], [283]) to interpret the decision- both performance and interpretability, potentially providing
making processes of foundation models, S-XAI methods can theoretical foundations for this balance.
16 GENERIC COLORIZED JOURNAL, VOL. XX, NO. XX, XXXX 2017
[50] M. Heath et al., “Current status of the digital database for screening [74] R. Gu et al., “Ca-net: Comprehensive attention convolutional neural
mammography,” in Digital Mammography: Nijmegen, 1998. Springer, networks for explainable medical image segmentation,” IEEE transac-
1998, pp. 457–460. tions on medical imaging, vol. 40, no. 2, pp. 699–711, 2020.
[51] G. Zhao, “Cross chest graph for disease diagnosis with structural [75] C. Barata, M. E. Celebi, and J. S. Marques, “Explainable skin lesion
relational reasoning,” in Proceedings of the 29th ACM International diagnosis using taxonomies,” Pattern Recognition, vol. 110, p. 107413,
Conference on Multimedia, 2021, pp. 612–620. 2021.
[52] B. Qi et al., “Gren: graph-regularized embedding network for weakly- [76] G. Lozupone et al., “Axial: Attention-based explainability for inter-
supervised disease localization in x-ray images,” IEEE Journal of pretable alzheimer’s localized diagnosis using 2d cnns on 3d mri brain
Biomedical and Health Informatics, vol. 26, no. 10, pp. 5142–5153, scans,” arXiv preprint arXiv:2407.02418, 2024.
2022. [77] J. Huang et al., “Swin deformable attention u-net transformer (sdaut)
[53] M. Elbatel, R. Martı́, and X. Li, “Fopro-kd: fourier prompted effective for explainable fast mri,” in International Conference on Medical Image
knowledge distillation for long-tailed medical image recognition,” IEEE Computing and Computer-Assisted Intervention. Springer, 2022, pp.
Transactions on Medical Imaging, 2023. 538–548.
[54] M. Combalia et al., “Bcn20000: Dermoscopic lesions in the wild,” [78] H. Wang et al., “Breast mass classification via deeply integrating the
arXiv preprint arXiv:1908.02288, 2019. contextual information from multi-view data,” Pattern Recognition,
[55] H. Borgli et al., “Hyperkvasir, a comprehensive multi-class image and vol. 80, pp. 42–52, 2018.
video dataset for gastrointestinal endoscopy,” Scientific data, vol. 7, [79] J. Arevalo et al., “Representation learning for mammography mass
no. 1, p. 283, 2020. lesion classification with convolutional neural networks,” Computer
[56] C. Y. Li et al., “Knowledge-driven encode, retrieve, paraphrase for Methods and Programs in Biomedicine, vol. 127, pp. 248–257, 2016.
medical image report generation,” in Proceedings of the AAAI confer- [Online]. Available: https://www.sciencedirect.com/science/article/pii/
ence on artificial intelligence, vol. 33, no. 01, 2019, pp. 6666–6673. S0169260715300110
[57] F. Liu et al., “Auto-encoding knowledge graph for unsupervised med- [80] H. Yang et al., “Guided soft attention network for classification of
ical report generation,” Advances in Neural Information Processing breast cancer histopathology images,” IEEE transactions on medical
Systems, vol. 34, pp. 16 266–16 279, 2021. imaging, vol. 39, no. 5, pp. 1306–1315, 2019.
[58] M. Li et al., “Dynamic graph enhanced contrastive learning for chest [81] G. Aresta et al., “Bach: Grand challenge on breast cancer histology
x-ray report generation,” in Proceedings of the IEEE/CVF Conference images,” Medical image analysis, vol. 56, pp. 122–139, 2019.
on Computer Vision and Pattern Recognition, 2023, pp. 3334–3343. [82] A. Diaz-Pinto et al., “Cnns for automatic glaucoma assessment us-
[59] K. Kale et al., “Kgvl-bart: Knowledge graph augmented visual lan- ing fundus images: an extensive validation,” Biomedical engineering
guage bart for radiology report generation,” in Proceedings of the online, vol. 18, pp. 1–19, 2019.
17th Conference of the European Chapter of the Association for [83] Y. Yan, J. Kawahara, and G. Hamarneh, “Melanoma recognition via
Computational Linguistics, 2023, pp. 3401–3411. visual attention,” in Information Processing in Medical Imaging: 26th
[60] H. Guo et al., “Medical visual question answering via targeted choice International Conference, IPMI 2019, Hong Kong, China, June 2–7,
contrast and multimodal entity matching,” in International Conference 2019, Proceedings 26. Springer, 2019, pp. 793–804.
on Neural Information Processing. Springer, 2022, pp. 343–354. [84] D. Gutman et al., “Skin lesion analysis toward melanoma detection: A
challenge at the international symposium on biomedical imaging (isbi)
[61] J. Huang et al., “Medical knowledge-based network for patient-oriented
2016, hosted by the international skin imaging collaboration (isic),”
visual question answering,” Information Processing & Management,
arXiv preprint arXiv:1605.01397, 2016.
vol. 60, no. 2, p. 103241, 2023.
[85] N. C. Codella et al., “Skin lesion analysis toward melanoma detection:
[62] X. Hu et al., “Expert knowledge-aware image difference graph rep-
A challenge at the 2017 international symposium on biomedical
resentation learning for difference-aware medical visual question an-
imaging (isbi), hosted by the international skin imaging collaboration
swering,” in Proceedings of the 29th ACM SIGKDD Conference on
(isic),” in 2018 IEEE 15th international symposium on biomedical
Knowledge Discovery and Data Mining, 2023, pp. 4156–4165.
imaging (ISBI 2018). IEEE, 2018, pp. 168–172.
[63] X. Hu et al., “Interpretable medical image visual question answering [86] N. Codella et al., “Skin lesion analysis toward melanoma detection
via multi-modal relationship graph learning,” Medical Image Analysis, 2018: A challenge hosted by the international skin imaging collabora-
vol. 97, p. 103279, 2024. tion (isic),” arXiv preprint arXiv:1902.03368, 2019.
[64] J. Li et al., “Align before fuse: Vision and language representation [87] C. Yin et al., “Focusing on clinically interpretable features: selective
learning with momentum distillation,” Advances in neural information attention regularization for liver biopsy image classification,” in Med-
processing systems, vol. 34, pp. 9694–9705, 2021. ical Image Computing and Computer Assisted Intervention–MICCAI
[65] S. Liu et al., “A hybrid method of recurrent neural network and graph 2021: 24th International Conference, Strasbourg, France, September
neural network for next-period prescription prediction,” International 27–October 1, 2021, Proceedings, Part V 24. Springer, 2021, pp.
Journal of Machine Learning and Cybernetics, vol. 11, pp. 2849–2856, 153–162.
2020. [88] F. Heinemann, G. Birk, and B. Stierstorfer, “Deep learning enables
[66] S. Liu et al., “Multimodal data matters: language model pre-training pathologist-like scoring of nash models,” Scientific reports, vol. 9,
over structured and unstructured electronic health records,” IEEE no. 1, p. 18454, 2019.
Journal of Biomedical and Health Informatics, vol. 27, no. 1, pp. 504– [89] G. Shih et al., “Augmenting the national institutes of health chest
514, 2022. radiograph dataset with expert annotations of possible pneumonia,”
[67] A. Radford et al., “Learning transferable visual models from natural Radiology: Artificial Intelligence, vol. 1, no. 1, p. e180041, 2019.
language supervision,” in International conference on machine learn- [90] D. S. Kermany et al., “Identifying medical diagnoses and treatable
ing. PMLR, 2021, pp. 8748–8763. diseases by image-based deep learning,” cell, vol. 172, no. 5, pp. 1122–
[68] Z. Niu, G. Zhong, and H. Yu, “A review on the attention mechanism 1131, 2018.
of deep learning,” Neurocomputing, vol. 452, pp. 48–62, 2021. [91] P. Lakhani et al., “The 2021 siim-fisabio-rsna machine learning covid-
[69] M. Bhattacharya, S. Jain, and P. Prasanna, “Radiotransformer: a 19 challenge: Annotation and standard exam classification of covid-19
cascaded global-focal transformer for visual attention–guided disease chest radiographs,” Journal of Digital Imaging, vol. 36, no. 1, pp. 365–
classification,” in European Conference on Computer Vision. Springer, 372, 2023.
2022, pp. 679–698. [92] M. E. Chowdhury et al., “Can ai help in screening viral and covid-19
[70] S. Jetley et al., “Learn to pay attention,” arXiv preprint pneumonia?” Ieee Access, vol. 8, pp. 132 665–132 676, 2020.
arXiv:1804.02391, 2018. [93] T. Rahman et al., “Exploring the effect of image enhancement tech-
[71] H. Fukui et al., “Attention branch network: Learning of attention niques on covid-19 detection using chest x-ray images,” Computers in
mechanism for visual explanation,” in Proceedings of the IEEE/CVF biology and medicine, vol. 132, p. 104319, 2021.
conference on computer vision and pattern recognition, 2019, pp. [94] J. Saltz et al., “Stony brook university covid-19 positive cases,” the
10 705–10 714. cancer imaging archive, vol. 4, 2021.
[72] L. Li et al., “Scouter: Slot attention-based classifier for explainable [95] E. Tsai et al., “Data from medical imaging data resource center (midrc)-
image recognition,” in Proceedings of the IEEE/CVF international rsna international covid radiology database (ricord) release 1c-chest x-
conference on computer vision, 2021, pp. 1046–1055. ray, covid+(midrc-ricord-1c),” The Cancer Imaging Archive, vol. 10,
[73] J. Schlemper et al., “Attention gated networks: Learning to leverage 2021.
salient regions in medical images,” Medical image analysis, vol. 53, [96] E. B. Tsai et al., “The rsna international covid-19 open radiology
pp. 197–207, 2019. database (ricord),” Radiology, vol. 299, no. 1, pp. E204–E213, 2021.
22 GENERIC COLORIZED JOURNAL, VOL. XX, NO. XX, XXXX 2017
[97] X. Wang et al., “Hospital-scale chest x-ray database and benchmarks [122] A. Sun et al., “Explain any concept: Segment anything meets concept-
on weakly-supervised classification and localization of common thorax based explanation,” Advances in Neural Information Processing Sys-
diseases,” in IEEE CVPR, vol. 7. sn, 2017, p. 46. tems, vol. 36, 2024.
[98] H. Q. Nguyen et al., “Vindr-cxr: An open dataset of chest x-rays with [123] A. Kirillov et al., “Segment anything,” in Proceedings of the IEEE/CVF
radiologist’s annotations,” Scientific Data, vol. 9, no. 1, p. 429, 2022. International Conference on Computer Vision, 2023, pp. 4015–4026.
[99] B. T. Wyman et al., “Standardization of analysis sets for reporting [124] M. Nevitt, D. Felson, and G. Lester, “The osteoarthritis initiative,”
results from adni mri data,” Alzheimer’s & Dementia, vol. 9, no. 3, pp. Protocol for the cohort study, vol. 1, p. 2, 2006.
332–337, 2013. [125] K. Chauhan et al., “Interactive concept bottleneck models,” in Proceed-
[100] H. R. Roth et al., “Hierarchical 3d fully convolutional networks for ings of the AAAI Conference on Artificial Intelligence, vol. 37, no. 5,
multi-organ segmentation,” arXiv preprint arXiv:1704.06382, 2017. 2023, pp. 5948–5955.
[101] H. R. Roth et al., “Data from pancreas-ct. the cancer imaging archive,” [126] C. Patrı́cio, J. C. Neves, and L. F. Teixeira, “Coherent concept-
IEEE Transactions on Image Processing, vol. 10, p. K9, 2016. based explanations in medical image and its application to skin lesion
[102] J. Sun et al., “Saunet: Shape attentive u-net for interpretable medical diagnosis,” in Proceedings of the IEEE/CVF Conference on Computer
image segmentation,” in Medical Image Computing and Computer Vision and Pattern Recognition, 2023, pp. 3799–3808.
Assisted Intervention–MICCAI 2020: 23rd International Conference, [127] J. Kawahara et al., “Seven-point checklist and skin lesion classification
Lima, Peru, October 4–8, 2020, Proceedings, Part IV 23. Springer, using multitask multimodal neural nets,” IEEE journal of biomedical
2020, pp. 797–806. and health informatics, vol. 23, no. 2, pp. 538–546, 2018.
[103] P. Radau et al., “Evaluation framework for algorithms segmenting short
[128] T. Mendonça et al., “Ph2: A public database for the analysis of
axis cardiac mri.” The MIDAS Journal, 2009.
dermoscopic images,” Dermoscopy image analysis, vol. 2, 2015.
[104] O. Bernard et al., “Deep learning techniques for automatic mri cardiac
multi-structures segmentation and diagnosis: is the problem solved?” [129] S. Yan et al., “Towards trustable skin cancer diagnosis via rewriting
IEEE transactions on medical imaging, vol. 37, no. 11, pp. 2514–2525, model’s decision,” in Proceedings of the IEEE/CVF Conference on
2018. Computer Vision and Pattern Recognition, 2023, pp. 11 568–11 577.
[105] M. Karri, C. S. R. Annavarapu, and U. R. Acharya, “Explainable [130] Y. Bie, L. Luo, and H. Chen, “Mica: Towards explainable skin lesion
multi-module semantic guided attention based network for medical diagnosis via multi-level image-concept alignment,” in Proceedings of
image segmentation,” Computers in Biology and Medicine, vol. 151, the AAAI Conference on Artificial Intelligence, vol. 38, no. 2, 2024,
p. 106231, 2022. pp. 837–845.
[106] P. Tschandl, C. Rosendahl, and H. Kittler, “The ham10000 dataset, [131] R. Daneshjou et al., “Skincon: A skin disease dataset densely annotated
a large collection of multi-source dermatoscopic images of common by domain experts for fine-grained debugging and analysis,” Advances
pigmented skin lesions. scientific data. 2018; 5: 180161,” Search in, in Neural Information Processing Systems, vol. 35, pp. 18 157–18 167,
vol. 2, 2018. 2022.
[107] C. for Biomedical Image Computing and Analytics, “Multimodal [132] C. Kim et al., “Transparent medical image ai via an image–text
brain tumor segmentation challenge 2020: Data,” MICCAI 2020 foundation model grounded in medical literature,” Nature Medicine,
BraTs, 2020. [Online]. Available: https://www.med.upenn.edu/cbica/ pp. 1–12, 2024.
brats2020/data.html [133] M. Groh et al., “Evaluating deep neural networks trained on clinical
[108] A. E. Kavur et al., “Chaos challenge - combined (ct-mr) healthy images in dermatology with the fitzpatrick 17k dataset,” in Proceed-
abdominal organ segmentation,” Medical Image Analysis, vol. 69, ings of the IEEE/CVF Conference on Computer Vision and Pattern
p. 101950, 2021. [Online]. Available: https://www.sciencedirect.com/ Recognition, 2021, pp. 1820–1828.
science/article/pii/S1361841520303145 [134] R. Daneshjou et al., “Disparities in dermatology ai performance on a
[109] H. Li et al., “Pmjaf-net: Pyramidal multi-scale joint attention and diverse, curated clinical image set,” Science advances, vol. 8, no. 31,
adaptive fusion network for explainable skin lesion segmentation,” p. eabq6147, 2022.
Computers in Biology and Medicine, p. 107454, 2023. [135] A. Lucieri et al., “Exaid: A multimodal explanation framework for
[110] R. Souza et al., “An open, multi-vendor, multi-field-strength brain computer-aided diagnosis of skin lesions,” Computer Methods and
mr dataset and analysis of publicly available skull stripping methods Programs in Biomedicine, vol. 215, p. 106620, 2022.
agreement,” NeuroImage, vol. 170, pp. 482–494, 2018. [136] R. Jalaboi et al., “Dermx: An end-to-end framework for explainable
[111] C. Lian et al., “End-to-end dementia status prediction from brain automated dermatological diagnosis,” Medical Image Analysis, vol. 83,
mri using multi-task weakly-supervised attention network,” in Medical p. 102647, 2023.
Image Computing and Computer Assisted Intervention–MICCAI 2019: [137] N. Z. D. Society, “Dermatology images.” [Online]. Available:
22nd International Conference, Shenzhen, China, October 13–17, 2019, https://dermnetnz.org/
Proceedings, Part IV 22. Springer, 2019, pp. 158–167. [138] X. Sun et al., “A benchmark for automatic visual classification of
[112] C. R. Jack Jr et al., “Update on the magnetic resonance imaging clinical skin disease images,” in Computer Vision–ECCV 2016: 14th
core of the alzheimer’s disease neuroimaging initiative,” Alzheimer’s European Conference, Amsterdam, The Netherlands, October 11-14,
& Dementia, vol. 6, no. 3, pp. 212–220, 2010. 2016, Proceedings, Part VI 14. Springer, 2016, pp. 206–222.
[113] P. W. Koh et al., “Concept bottleneck models,” in International
[139] J. Hou, J. Xu, and H. Chen, “Concept-attention whitening for inter-
conference on machine learning. PMLR, 2020, pp. 5338–5348.
pretable skin lesion diagnosis,” arXiv preprint arXiv:2404.05997, 2024.
[114] M. Yuksekgonul, M. Wang, and J. Zou, “Post-hoc concept bottleneck
models,” arXiv preprint arXiv:2205.15480, 2022. [140] I. Kim et al., “Concept bottleneck with visual concept filtering for
[115] R. Jain et al., “Extending logic explained networks to text classifica- explainable medical image classification,” in International Conference
tion,” arXiv preprint arXiv:2211.09732, 2022. on Medical Image Computing and Computer-Assisted Intervention.
[116] A. Tan, F. Zhou, and H. Chen, “Explain via any concept: Concept Springer, 2023, pp. 225–233.
bottleneck model with open vocabulary concepts,” arXiv preprint [141] C. Patrı́cio, L. F. Teixeira, and J. C. Neves, “Towards concept-based
arXiv:2408.02265, 2024. interpretability of skin lesion diagnosis using vision-language models,”
[117] B. Kim et al., “Interpretability beyond feature attribution: Quantitative in 2024 IEEE International Symposium on Biomedical Imaging (ISBI).
testing with concept activation vectors (tcav),” in International confer- IEEE, 2024, pp. 1–5.
ence on machine learning. PMLR, 2018, pp. 2668–2677. [142] W. Pang et al., “Integrating clinical knowledge into concept bottleneck
[118] J. R. Clough et al., “Global and local interpretability for cardiac models,” in International Conference on Medical Image Computing
mri classification,” in International Conference on Medical Image and Computer-Assisted Intervention (MICCAI), 2024.
Computing and Computer-Assisted Intervention. Springer, 2019, pp. [143] S. Tsutsui, W. Pang, and B. Wen, “Wbcatt: a white blood cell dataset
656–664. annotated with detailed morphological attributes,” Advances in Neural
[119] M. Graziani et al., “Concept attribution: Explaining cnn decisions to Information Processing Systems, vol. 36, 2024.
physicians,” Computers in biology and medicine, vol. 123, p. 103865, [144] R. Marcinkevičs et al., “Interpretable and intervenable ultrasonography-
2020. based machine learning models for pediatric appendicitis,” Medical
[120] R. Achtibat et al., “From attribution maps to human-understandable Image Analysis, vol. 91, p. 103042, 2024.
explanations through concept relevance propagation,” Nature Machine [145] G. Zhao et al., “Diagnose like a radiologist: Hybrid neuro-probabilistic
Intelligence, vol. 5, no. 9, pp. 1006–1019, 2023. reasoning for attribute-based medical image diagnosis,” IEEE Trans-
[121] E. Poeta et al., “Concept-based explainable artificial intelligence: A actions on Pattern Analysis and Machine Intelligence, vol. 44, no. 11,
survey,” arXiv preprint arXiv:2312.12936, 2023. pp. 7400–7416, 2021.
AUTHOR et al.: PREPARATION OF PAPERS FOR IEEE TRANSACTIONS AND JOURNALS (FEBRUARY 2017) 23
[146] S. G. Armato III et al., “The lung image database consortium (lidc) [172] I. Biederman, “Recognition-by-components: a theory of human image
and image database resource initiative (idri): a completed reference understanding.” Psychological review, vol. 94, no. 2, p. 115, 1987.
database of lung nodules on ct scans,” Medical physics, vol. 38, no. 2, [173] G. Singh and K.-C. Yow, “An interpretable deep learning model for
pp. 915–931, 2011. covid-19 detection with chest x-ray images,” Ieee Access, vol. 9, pp.
[147] Z. Fang et al., “Concept-based explanation for fine-grained images and 85 198–85 208, 2021.
its application in infectious keratitis classification,” in Proceedings of [174] J. P. Cohen, P. Morrison, and L. Dao, “Covid-19 image data collection,”
the 28th ACM international conference on Multimedia, 2020, pp. 700– arXiv preprint arXiv:2003.11597, 2020.
708. [175] S. Mohammadjafari et al., “Using protopnet for interpretable
[148] Y. Xu et al., “Deep sequential feature learning in clinical image alzheimer’s disease classification.” in Canadian Conference on AI,
classification of infectious keratitis,” Engineering, vol. 7, no. 7, pp. 2021.
1002–1010, 2021. [176] D. S. Marcus et al., “Open access series of imaging studies (oasis):
[149] C. Wen et al., “Concept-based lesion aware transformer for inter- cross-sectional mri data in young, middle aged, nondemented, and
pretable retinal disease diagnosis,” IEEE Transactions on Medical demented older adults,” Journal of cognitive neuroscience, vol. 19,
Imaging, 2024. no. 9, pp. 1498–1507, 2007.
[150] Y. Zhou et al., “A benchmark for studying diabetic retinopathy: seg- [177] A. J. Barnett et al., “A case-based interpretable deep learning model
mentation, grading, and transferability,” IEEE Transactions on Medical for classification of mass lesions in digital mammography,” Nature
Imaging, vol. 40, no. 3, pp. 818–828, 2020. Machine Intelligence, vol. 3, no. 12, pp. 1061–1070, 2021.
[151] T. Li et al., “Diagnostic assessment of deep learning algorithms for [178] G. Carloni et al., “On the applicability of prototypical part learning
diabetic retinopathy screening,” Information Sciences, vol. 501, pp. in medical images: breast masses classification using protopnet,” in
511–522, 2019. International Conference on Pattern Recognition. Springer, 2022, pp.
[152] M. Kong et al., “Attribute-aware interpretation learning for thyroid 539–557.
ultrasound diagnosis,” Artificial Intelligence in Medicine, vol. 131, p. [179] R. S. Lee et al., “A curated mammography data set for use in computer-
102344, 2022. aided detection and diagnosis research,” Scientific data, vol. 4, no. 1,
[153] J. Liu et al., “A chatgpt aided explainable framework for zero-shot pp. 1–9, 2017.
medical image diagnosis,” arXiv preprint arXiv:2307.01981, 2023. [180] C. Wang et al., “Knowledge distillation to ensemble global and
[154] S. Jaeger et al., “Two public chest x-ray datasets for computer-aided interpretable prototype-based mammogram classification models,” in
screening of pulmonary diseases,” Quantitative imaging in medicine International Conference on Medical Image Computing and Computer-
and surgery, vol. 4, no. 6, p. 475, 2014. Assisted Intervention. Springer, 2022, pp. 14–24.
[155] P. Porwal et al., “Indian diabetic retinopathy image dataset (idrid): [181] C. Cui et al., “The chinese mammography database (cmmd): An online
a database for diabetic retinopathy screening research,” Data, vol. 3, mammography database with biopsy confirmed types for machine
no. 3, p. 25, 2018. diagnosis of breast,” The Cancer Imaging Archive, vol. 1, 2021.
[156] Y. Gao et al., “Aligning human knowledge with visual concepts [182] Y. Wei, R. Tam, and X. Tang, “Mprotonet: A case-based interpretable
towards explainable medical image classification,” arXiv preprint model for brain tumor classification with 3d multi-parametric magnetic
arXiv:2406.05596, 2024. resonance imaging,” in Medical Imaging with Deep Learning. PMLR,
[157] J. N. Kather, N. Halama, and A. Marx, “100,000 histological images 2024, pp. 1798–1812.
of human colorectal cancer and healthy tissue,” Zenodo10, vol. 5281,
[183] B. H. Menze et al., “The multimodal brain tumor image segmentation
no. 9, 2018.
benchmark (brats),” IEEE transactions on medical imaging, vol. 34,
[158] W. Al-Dhabyani et al., “Deep learning approaches for data augmenta-
no. 10, pp. 1993–2024, 2014.
tion and classification of breast masses using ultrasound images,” Int.
[184] L. S. Hesse and A. I. Namburete, “Insightr-net: interpretable neural net-
J. Adv. Comput. Sci. Appl, vol. 10, no. 5, pp. 1–11, 2019.
work for regression using similarity-based comparisons to prototypical
[159] M. Espinosa Zarlenga et al., “Concept embedding models: Beyond
examples,” in International Conference on Medical Image Computing
the accuracy-explainability trade-off,” Advances in Neural Information
and Computer-Assisted Intervention. Springer, 2022, pp. 502–511.
Processing Systems, vol. 35, pp. 21 400–21 413, 2022.
[160] Z. Chen, Y. Bei, and C. Rudin, “Concept whitening for interpretable [185] C. H. Foundation, “Eyepacs,” 2015. [Online]. Available: https:
image recognition,” Nature Machine Intelligence, vol. 2, no. 12, pp. //www.kaggle.com/c/diabetic-retinopathy-detection/data
772–782, 2020. [186] I. B. d. A. Santos and A. C. de Carvalho, “Protoal: Interpretable deep
[161] S. Lapuschkin et al., “Unmasking clever hans predictors and assessing active learning with prototypes for medical imaging,” arXiv preprint
what machines really learn,” Nature communications, vol. 10, no. 1, p. arXiv:2404.04736, 2024.
1096, 2019. [187] E. Decencière et al., “Feedback on a publicly distributed image
[162] A. Ghorbani et al., “Towards automatic concept-based explanations,” database: The messidor database. image anal & stereology 33: 231–
Advances in neural information processing systems, vol. 32, 2019. 234,” 2014.
[163] A. Sarkar et al., “A framework for learning ante-hoc explainable [188] L. S. Hesse, N. K. Dinsdale, and A. I. L. Namburete, “Prototype
models via concepts,” in Proceedings of the IEEE/CVF Conference on learning for explainable brain age prediction,” in Proceedings of the
Computer Vision and Pattern Recognition, 2022, pp. 10 286–10 295. IEEE/CVF Winter Conference on Applications of Computer Vision
[164] C.-K. Yeh et al., “On completeness-aware concept-based explanations (WACV), January 2024, pp. 7903–7913.
in deep neural networks,” Advances in neural information processing [189] https://brain-development.org/ixi-dataset/.
systems, vol. 33, pp. 20 554–20 565, 2020. [190] A. T. Papageorghiou et al., “International standards for fetal growth
[165] Y. Yang et al., “Language in a bottle: Language model guided concept based on serial ultrasound measurements: the fetal growth longitudinal
bottlenecks for interpretable image classification,” in Proceedings of the study of the intergrowth-21st project,” The Lancet, vol. 384, no. 9946,
IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 869–879, 2014.
2023, pp. 19 187–19 197. [191] D. Rymarczyk et al., “Protopshare: Prototypical parts sharing for simi-
[166] T. Brown et al., “Language models are few-shot learners,” Advances in larity discovery in interpretable image classification,” in Proceedings of
neural information processing systems, vol. 33, pp. 1877–1901, 2020. the 27th ACM SIGKDD Conference on Knowledge Discovery & Data
[167] A. Radford et al., “Learning transferable visual models from natural Mining, 2021, pp. 1420–1430.
language supervision,” in International conference on machine learn- [192] D. Rymarczyk et al., “Interpretable image classification with differen-
ing. PMLR, 2021, pp. 8748–8763. tiable prototypes assignment,” in European Conference on Computer
[168] T. Oikarinen et al., “Label-free concept bottleneck models,” arXiv Vision. Springer, 2022, pp. 351–368.
preprint arXiv:2304.06129, 2023. [193] J. Donnelly, A. J. Barnett, and C. Chen, “Deformable protopnet:
[169] Y. Bie et al., “Xcoop: Explainable prompt learning for computer-aided An interpretable image classifier using deformable prototypes,” in
diagnosis via concept-guided context optimization,” arXiv preprint Proceedings of the IEEE/CVF conference on computer vision and
arXiv:2403.09410, 2024. pattern recognition, 2022, pp. 10 265–10 275.
[170] E. Kim et al., “Xprotonet: diagnosis in chest radiography with global [194] J. Wang et al., “Interpretable image recognition by constructing trans-
and local explanations,” in Proceedings of the IEEE/CVF conference parent embedding space,” in Proceedings of the IEEE/CVF interna-
on computer vision and pattern recognition, 2021, pp. 15 719–15 728. tional conference on computer vision, 2021, pp. 895–904.
[171] C. Chen et al., “This looks like that: deep learning for interpretable im- [195] B. Wang et al., “Learning bottleneck concepts in image classification,”
age recognition,” Advances in neural information processing systems, in Proceedings of the ieee/cvf conference on computer vision and
vol. 32, 2019. pattern recognition, 2023, pp. 10 962–10 971.
24 GENERIC COLORIZED JOURNAL, VOL. XX, NO. XX, XXXX 2017
[196] P. Hase et al., “Interpretable image recognition with hierarchical proto- [216] J. T. Wu et al., “Chest imagenome dataset for clinical reasoning,”
types,” in Proceedings of the AAAI Conference on Human Computation in Thirty-fifth Conference on Neural Information Processing Systems
and Crowdsourcing, vol. 7, 2019, pp. 32–40. Datasets and Benchmarks Track (Round 2).
[197] Y. Ukai et al., “This looks like it rather than that: Protoknn for [217] Q. Li et al., “Anatomical structure-guided medical vision-language pre-
similarity-based classifiers,” in The Eleventh International Conference training,” arXiv preprint arXiv:2403.09294, 2024.
on Learning Representations, 2022. [218] T. Tanida et al., “Interactive and explainable region-guided radiology
[198] A. Bontempelli et al., “Concept-level debugging of part-prototype report generation,” in Proceedings of the IEEE/CVF Conference on
networks,” in The Eleventh International Conference on Learning Computer Vision and Pattern Recognition, 2023, pp. 7433–7442.
Representations, 2023. [219] L. Wang et al., “An inclusive task-aware framework for radiology report
[199] O. Li et al., “Deep learning for case-based reasoning through proto- generation,” in International Conference on Medical Image Computing
types: A neural network that explains its predictions,” in Proceedings and Computer-Assisted Intervention. Springer, 2022, pp. 568–577.
of the AAAI Conference on Artificial Intelligence, vol. 32, no. 1, 2018. [220] D. Demner-Fushman et al., “Design and development of a multimodal
[200] Q. Huang et al., “Evaluation and improvement of interpretability biomedical information retrieval system,” Journal of Computing Sci-
for self-explainable part-prototype networks,” in Proceedings of the ence and Engineering, vol. 6, no. 2, pp. 168–177, 2012.
IEEE/CVF International Conference on Computer Vision, 2023, pp. [221] S. Singh et al., “From chest x-rays to radiology reports: a multimodal
2011–2020. machine learning approach,” in 2019 Digital Image Computing: Tech-
[201] M. Nauta et al., “Pip-net: Patch-based intuitive prototypes for in- niques and Applications (DICTA). IEEE, 2019, pp. 1–8.
terpretable image classification,” in Proceedings of the IEEE/CVF [222] G. Spinks and M.-F. Moens, “Justifying diagnosis decisions by deep
Conference on Computer Vision and Pattern Recognition, 2023, pp. neural networks,” Journal of biomedical informatics, vol. 96, p.
2744–2753. 103248, 2019.
[202] C. Ma et al., “This looks like those: Illuminating prototypical con- [223] Y. Kim et al., “Adversarially regularized autoencoders for generating
cepts using multiple visualizations,” Advances in Neural Information discrete structures,” arXiv preprint arXiv:1706.04223, vol. 2, p. 12,
Processing Systems, vol. 36, 2024. 2017.
[203] M. Nauta, R. Van Bree, and C. Seifert, “Neural prototype trees for [224] G. Liu et al., “Clinically accurate chest x-ray report generation,” in
interpretable fine-grained image recognition,” in Proceedings of the Machine Learning for Healthcare Conference. PMLR, 2019, pp. 249–
IEEE/CVF conference on computer vision and pattern recognition, 269.
2021, pp. 14 933–14 943. [225] Z. Chen et al., “Generating radiology reports via memory-driven
[204] A. Tan, Z. Fengtao, and H. Chen, “Post-hoc part-prototype networks,” transformer,” in Proceedings of the 2020 Conference on Empirical
in Forty-first International Conference on Machine Learning. Methods in Natural Language Processing (EMNLP), 2020, pp. 1439–
[205] J. Kim, M. Kim, and Y. M. Ro, “Interpretation of lesional detection 1449.
via counterfactual generation,” in 2021 IEEE International Conference [226] Z. Wang et al., “Metransformer: Radiology report generation by trans-
on Image Processing (ICIP). IEEE, 2021, pp. 96–100. former with multiple learnable expert tokens,” in Proceedings of the
[206] P. Pino et al., “Clinically correct report generation from chest x- IEEE/CVF Conference on Computer Vision and Pattern Recognition,
rays using templates,” in Machine Learning in Medical Imaging: 2023, pp. 11 558–11 567.
12th International Workshop, MLMI 2021, Held in Conjunction with [227] J. Yuan et al., “Automatic radiology report generation based on multi-
MICCAI 2021, Strasbourg, France, September 27, 2021, Proceedings view image fusion and medical concept enrichment,” in Medical Image
12. Springer, 2021, pp. 654–663. Computing and Computer Assisted Intervention–MICCAI 2019: 22nd
[207] C. E. Lipscomb, “Medical subject headings (mesh),” Bulletin of the International Conference, Shenzhen, China, October 13–17, 2019,
Medical Library Association, vol. 88, no. 3, p. 265, 2000. Proceedings, Part VI 22. Springer, 2019, pp. 721–729.
[208] H.-C. Shin et al., “Learning to read chest x-rays: Recurrent neural [228] H. Lee, S. Kim, and Y. Ro, “Generation of multimodal justification
cascade model for automated image annotation,” in Proceedings of the using visual word constraint model for explainable computer-aided
IEEE conference on computer vision and pattern recognition, 2016, diagnosis,” in Interpretability of Machine Intelligence in Medical Image
pp. 2497–2506. Computing and Multimodal Learning for Clinical Decision Support.
[209] A. Gasimova, “Automated enriched medical concept generation for Springer, 2019.
chest x-ray images,” in Interpretability of Machine Intelligence in [229] Z. Zhang et al., “Pathologist-level interpretable whole-slide cancer
Medical Image Computing and Multimodal Learning for Clinical diagnosis with deep learning,” Nature Machine Intelligence, vol. 1,
Decision Support: Second International Workshop, iMIMIC 2019, and no. 5, pp. 236–245, 2019.
9th International Workshop, ML-CDS 2019, Held in Conjunction with [230] N. C. Institute, “The cancer genome atlas program,” 2006. [Online].
MICCAI 2019, Shenzhen, China, October 17, 2019, Proceedings 9. Available: https://www.cancer.gov/tcga
Springer, 2019, pp. 83–92. [231] X. Wang et al., “Tienet: Text-image embedding network for com-
[210] I. Rodin et al., “Multitask and multimodal neural network model for mon thorax disease classification and reporting in chest x-rays,” in
interpretable analysis of x-ray images,” in 2019 IEEE International Proceedings of the IEEE conference on computer vision and pattern
Conference on Bioinformatics and Biomedicine (BIBM). IEEE, 2019, recognition, 2018, pp. 9049–9058.
pp. 1601–1604. [232] B. Jing, P. Xie, and E. Xing, “On the automatic generation of medical
[211] Z. Zhang et al., “Mdnet: A semantically and visually interpretable med- imaging reports,” in Proceedings of the 56th Annual Meeting of the
ical image diagnosis network,” in Proceedings of the IEEE conference Association for Computational Linguistics (Volume 1: Long Papers),
on computer vision and pattern recognition, 2017, pp. 6428–6436. 2018, pp. 2577–2586.
[212] Z. Zhang et al., “Tandemnet: Distilling knowledge from medical [233] K. N. Jones et al., “Peir digital library: Online resources and authoring
images using diagnostic reports as optional semantic references,” in system,” in Proceedings of the AMIA Symposium. American Medical
Medical Image Computing and Computer Assisted Intervention- MIC- Informatics Association, 2001, p. 1075.
CAI 2017: 20th International Conference, Quebec City, QC, Canada, [234] X. Zeng et al., “Generating diagnostic report for medical image by
September 11-13, 2017, Proceedings, Part III 20. Springer, 2017, pp. high-middle-level visual information incorporation on double deep
320–328. learning models,” Computer methods and programs in biomedicine,
[213] K. Ma et al., “A pathology image diagnosis network with visual vol. 197, p. 105700, 2020.
interpretability and structured diagnostic report,” in Neural Information [235] J. Tian et al., “A diagnostic report generator from ct volumes on liver
Processing: 25th International Conference, ICONIP 2018, Siem Reap, tumor with semi-supervised attention mechanism,” in Medical Image
Cambodia, December 13–16, 2018, Proceedings, Part VI 25. Springer, Computing and Computer Assisted Intervention–MICCAI 2018: 21st
2018, pp. 282–293. International Conference, Granada, Spain, September 16-20, 2018,
[214] X. Wang et al., “A computational framework towards medical image Proceedings, Part II 11. Springer, 2018, pp. 702–710.
explanation,” in Artificial Intelligence in Medicine: Knowledge Rep- [236] P. Bilic et al., “The liver tumor segmentation benchmark (lits),” Medical
resentation and Transparent and Explainable Systems: AIME 2019 Image Analysis, vol. 84, p. 102680, 2023.
International Workshops, KR4HC/ProHealth and TEAAM, Poznan, [237] O. Thawkar et al., “Xraygpt: Chest radiographs summarization using
Poland, June 26–29, 2019, Revised Selected Papers. Springer, 2019, medical vision-language models,” arXiv preprint arXiv:2306.07971,
pp. 120–131. 2023.
[215] W. Gale et al., “Producing radiologist-quality reports for interpretable [238] J. Zhou et al., “Pre-trained multimodal large language model enhances
deep learning,” in 2019 IEEE 16th international symposium on biomed- dermatological diagnosis using skingpt-4,” Nature Communications,
ical imaging (ISBI 2019). IEEE, 2019, pp. 1275–1279. vol. 15, no. 1, p. 5649, 2024.
AUTHOR et al.: PREPARATION OF PAPERS FOR IEEE TRANSACTIONS AND JOURNALS (FEBRUARY 2017) 25
[239] kaggle, “Dermnet.” [Online]. Available: https://www.kaggle.com/ and machine explanations,” KI-Künstliche Intelligenz, vol. 34, no. 2,
datasets/shubhamgoel27/dermnet pp. 193–198, 2020.
[240] M. Moor et al., “Med-flamingo: a multimodal medical few-shot [266] R. Sayres et al., “Using a deep learning algorithm and integrated
learner,” in Machine Learning for Health (ML4H). PMLR, 2023, gradients explanation to assist grading for diabetic retinopathy,” Oph-
pp. 353–367. thalmology, vol. 126, no. 4, pp. 552–564, 2019.
[241] W. Lin et al., “Pmc-clip: Contrastive language-image pre-training using [267] V. Petsiuk, A. Das, and K. Saenko, “Rise: Randomized input
biomedical documents,” in International Conference on Medical Image sampling for explanation of black-box models,” arXiv preprint
Computing and Computer-Assisted Intervention. Springer, 2023, pp. arXiv:1806.07421, 2018.
525–536. [268] S. Hooker et al., “A benchmark for interpretability methods in deep
[242] C. Li et al., “Llava-med: Training a large language-and-vision assistant neural networks,” Advances in neural information processing systems,
for biomedicine in one day,” Advances in Neural Information Process- vol. 32, 2019.
ing Systems, vol. 36, 2024. [269] J. Hou et al., “Diabetic retinopathy grading with weakly-supervised
[243] J. J. Lau et al., “A dataset of clinically generated visual questions and lesion priors,” in ICASSP 2023-2023 IEEE International Conference
answers about radiology images,” Scientific data, vol. 5, no. 1, pp. on Acoustics, Speech and Signal Processing (ICASSP). IEEE, 2023,
1–10, 2018. pp. 1–5.
[244] S. He et al., “Meddr: Diagnosis-guided bootstrapping for large-scale [270] A. J. Barnett et al., “Interpretable mammographic image classifica-
medical vision-language learning,” arXiv preprint arXiv:2404.15127, tion using case-based reasoning and deep learning,” arXiv preprint
2024. arXiv:2107.05605, 2021.
[245] J. Chen et al., “Huatuogpt-vision, towards injecting medical vi- [271] S. M. Muddamsetty, M. N. Jahromi, and T. B. Moeslund, “Expert level
sual knowledge into multimodal llms at scale,” arXiv preprint evaluations for explainable ai (xai) methods in the medical domain,”
arXiv:2406.19280, 2024. in International Conference on Pattern Recognition. Springer, 2021,
[246] X. Zhang et al., “Pmc-vqa: Visual instruction tuning for medical visual pp. 35–46.
question answering,” arXiv preprint arXiv:2305.10415, 2023. [272] A.-p. Nguyen and M. R. Martı́nez, “On quantitative aspects of model
[247] S. Kang et al., “Wolf: Large language model framework for cxr interpretability,” arXiv preprint arXiv:2007.07584, 2020.
understanding,” arXiv preprint arXiv:2403.15456, 2024. [273] K. Papineni et al., “Bleu: a method for automatic evaluation of
[248] S. Bae et al., “Ehrxqa: A multi-modal question answering dataset for machine translation,” in Proceedings of the 40th annual meeting of
electronic health records with chest x-ray images,” Advances in Neural the Association for Computational Linguistics, 2002, pp. 311–318.
Information Processing Systems, vol. 36, 2024. [274] C.-Y. Lin, “Rouge: A package for automatic evaluation of summaries,”
[249] Z. Chen et al., “Chexagent: Towards a foundation model for chest x-ray in Text summarization branches out, 2004, pp. 74–81.
interpretation,” arXiv preprint arXiv:2401.12208, 2024. [275] R. Vedantam, C. Lawrence Zitnick, and D. Parikh, “Cider: Consensus-
[250] K. Schutte et al., “Using stylegan for visual interpretability of deep based image description evaluation,” in Proceedings of the IEEE
learning models on medical images,” arXiv preprint arXiv:2101.07563, conference on computer vision and pattern recognition, 2015, pp.
2021. 4566–4575.
[276] S. Singla et al., “Explaining the black-box smoothly—a counterfactual
[251] I. Goodfellow et al., “Generative adversarial nets,” Advances in neural
approach,” Medical Image Analysis, vol. 84, p. 102721, 2023.
information processing systems, vol. 27, 2014.
[277] S. Tonekaboni et al., “What clinicians want: contextualizing explain-
[252] T. Laugel et al., “Issues with post-hoc counterfactual explanations: a
able machine learning for clinical end use,” in Machine learning for
discussion,” arXiv preprint arXiv:1906.04774, 2019.
healthcare conference. PMLR, 2019, pp. 359–380.
[253] H. Guo, T. H. Nguyen, and A. Yadav, “Counternet: End-to-end training
[278] K. Singhal et al., “Towards expert-level medical question answering
of prediction aware counterfactual explanations,” in Proceedings of the
with large language models,” arXiv preprint arXiv:2305.09617, 2023.
29th ACM SIGKDD Conference on Knowledge Discovery and Data
[279] C. Wu et al., “Pmc-llama: toward building open-source language
Mining, 2023, pp. 577–589.
models for medicine,” Journal of the American Medical Informatics
[254] V. Guyomard et al., “Vcnet: A self-explaining model for realistic Association, p. ocae045, 2024.
counterfactual generation,” in Joint European Conference on Machine [280] Y. Gu et al., “Domain-specific language model pretraining for biomed-
Learning and Knowledge Discovery in Databases. Springer, 2022, ical natural language processing,” ACM Transactions on Computing for
pp. 437–453. Healthcare (HEALTH), vol. 3, no. 1, pp. 1–23, 2021.
[255] M. Wilms et al., “Towards self-explainable classifiers and regressors [281] M. Moor et al., “Foundation models for generalist medical artificial
in neuroimaging with normalizing flows,” in International Workshop intelligence,” Nature, vol. 616, no. 7956, pp. 259–265, 2023.
on Machine Learning in Clinical Neuroimaging, 2021, pp. 23–33. [282] X. Ye and G. Durrett, “Can explanations be useful for calibrating
[256] U. Johansson, R. König, and L. Niklasson, “The truth is in there- black box models?” in Proceedings of the 60th Annual Meeting of the
rule extraction from opaque models using genetic programming.” in Association for Computational Linguistics (Volume 1: Long Papers),
FLAIRS, 2004, pp. 658–663. 2022, pp. 6199–6212.
[257] H. Lakkaraju et al., “Faithful and customizable explanations of black [283] X. Wu et al., “From language modeling to instruction following:
box models,” in Proceedings of the 2019 AAAI/ACM Conference on Understanding the behavior shift in llms after instruction tuning,” in
AI, Ethics, and Society, 2019, pp. 131–138. Proceedings of the 2024 Conference of the North American Chapter
[258] W. Jin et al., “Guidelines and evaluation of clinical explainable ai in of the Association for Computational Linguistics: Human Language
medical image analysis,” Medical Image Analysis, vol. 84, p. 102684, Technologies (Volume 1: Long Papers), 2024, pp. 2341–2369.
2023. [284] J. Wei et al., “Chain-of-thought prompting elicits reasoning in large
[259] E. Lughofer et al., “Explaining classifier decisions linguistically for language models,” Advances in neural information processing systems,
stimulating and improving operators labeling behavior,” Information vol. 35, pp. 24 824–24 837, 2022.
Sciences, vol. 420, pp. 16–36, 2017. [285] Y. Shi et al., “Mededit: Model editing for medical question answering
[260] M. Robnik-Šikonja and M. Bohanec, “Perturbation-based explanations with external knowledge bases,” arXiv preprint arXiv:2309.16035,
of prediction models,” Human and Machine Learning: Visible, Explain- 2023.
able, Trustworthy and Transparent, pp. 159–175, 2018. [286] K. Singhal et al., “Large language models encode clinical knowledge,”
[261] A. Adadi and M. Berrada, “Explainable ai for healthcare: from black Nature, vol. 620, no. 7972, pp. 172–180, 2023.
box to interpretable models,” in Embedded systems and artificial [287] C. Zhao et al., “Automated natural language explanation of deep visual
intelligence: proceedings of ESAI 2019, Fez, Morocco. Springer, 2020, neurons with large models,” arXiv preprint arXiv:2310.10708, 2023.
pp. 327–337. [288] Y. Yang et al., “Language in a bottle: Language model guided concept
[262] X. Wu et al., “Usable xai: 10 strategies towards exploiting explainabil- bottlenecks for interpretable image classification,” in Proceedings of the
ity in the llm era,” arXiv preprint arXiv:2403.08946, 2024. IEEE/CVF Conference on Computer Vision and Pattern Recognition,
[263] F. Doshi-Velez and B. Kim, “Towards a rigorous science of inter- 2023, pp. 19 187–19 197.
pretable machine learning,” arXiv preprint arXiv:1702.08608, 2017. [289] C. Singh et al., “Augmenting interpretable models with large language
[264] J. Zhou et al., “Evaluating the quality of machine learning explanations: models during training,” Nature Communications, vol. 14, no. 1, p.
A survey on methods and metrics,” Electronics, vol. 10, no. 5, p. 593, 7913, 2023.
2021. [290] S. Bills et al., “Language models can explain neurons in language
[265] A. Holzinger, A. Carrington, and H. Müller, “Measuring the quality models,” URL https://openaipublic. blob. core. windows. net/neuron-
of explanations: the system causability scale (scs) comparing human explainer/paper/index. html.(Date accessed: 14.05. 2023), vol. 2, 2023.
26 GENERIC COLORIZED JOURNAL, VOL. XX, NO. XX, XXXX 2017