Next Article in Journal
Method for Augmenting Side-Scan Sonar Seafloor Sediment Image Dataset Based on BCEL1-CBAM-INGAN
Next Article in Special Issue
Comparison of Visual and Quantra Software Mammographic Density Assessment According to BI-RADS® in 2D and 3D Images
Previous Article in Journal
Three-Dimensional Reconstruction of Indoor Scenes Based on Implicit Neural Representation
Previous Article in Special Issue
Decoding Breast Cancer: Using Radiomics to Non-Invasively Unveil Molecular Subtypes Directly from Mammographic Images
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Enhancing Deep Learning Model Explainability in Brain Tumor Datasets Using Post-Heuristic Approaches

by
Konstantinos Pasvantis
and
Eftychios Protopapadakis
*,†
Department of Applied Informatics, University of Macedonia, Egnatia 156, 546 36 Thessaloniki, Greece
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
J. Imaging 2024, 10(9), 232; https://doi.org/10.3390/jimaging10090232
Submission received: 30 July 2024 / Revised: 2 September 2024 / Accepted: 5 September 2024 / Published: 18 September 2024

Abstract

:
The application of deep learning models in medical diagnosis has showcased considerable efficacy in recent years. Nevertheless, a notable limitation involves the inherent lack of explainability during decision-making processes. This study addresses such a constraint by enhancing the interpretability robustness. The primary focus is directed towards refining the explanations generated by the LIME Library and LIME image explainer. This is achieved through post-processing mechanisms based on scenario-specific rules. Multiple experiments have been conducted using publicly accessible datasets related to brain tumor detection. Our proposed post-heuristic approach demonstrates significant advancements, yielding more robust and concrete results in the context of medical diagnosis.

Graphical Abstract

1. Introduction

Deep learning models’ capacity to model complicated patterns on various fields, such as medical imaging, has demonstrated great potential for prognostic and diagnostic purposes [1,2,3,4,5,6]. Deep learning’s extensive capabilities have made progress in medical image analysis possible, allowing for the more accurate and efficient diagnosis of a variety of illnesses.
However, the deployment of deep learning models in medical image analysis is not without challenges [7,8]. One major issue with these models is their lack of explainability. Because deep neural networks are complex, their decision-making processes are not frequently transparent, which makes it difficult for medical experts to understand and accept the outcomes. This problem is particularly important for medical applications, because misinterpretation might have serious consequences.
Researchers have looked into a number of ways to improve the interpretability of deep learning models in order to close the explainability gap, especially when it comes to medical imagery [9,10,11]. Such approaches could include (a) Model-specific methods, e.g., saliency maps or activation maximization [12,13], or (b) Model-agnostic methods, e.g., partial dependence plots or surrogate models [14,15,16,17]. Even though progress has been achieved in this area, there is always opportunity for enhancement, particularly when it comes to improving the results’ interpretability for certain applications.
This paper explores the relationship between medical image analysis and deep learning, specifically focusing on enhancing the explainability of classification models for detecting brain tumors in MRI images. Understanding the difficulties in obtaining results that are transparent, we use an explainability method that is specific to the complexities of medical image analysis. Most importantly, we contribute to a later improvement of this method with the goal of offering more accurate and useful information on the existence of brain tumors.

2. Related Work

Various CNN architectures, such as AlexNet [18] and VGGNet [19], have been influential in classifying medical images into distinct categories. Transfer learning methods have also made it easier to apply previously learned CNN models to medical image classification tasks, which has enhanced diagnostic capabilities. AlexNet, for example, has shown success in classifying skin lesions in dermatological images, contributing to the early identification of skin cancers [20].
The application field may vary, from suspicious and abnormal ardiotocographic recordings [21] to glaucoma detection [22]. The former case is important for monitoring the health of both the mother and the fetus during pregnancy. The proposed method improved time complexity, a crucial factor in clinical settings, by combining AlexNet with support vector machines (SVMs) at the fully connected layers. The latter case involved the development of image analysis diagnostic tools.
Another notable example of transfer learning regarding classification is the adaptation of a pre-trained ResNet model for the detecting COVID-19 in chest X-ray images [23]. The model was evaluated using both binary classification and multi-class classification methods, and all of the results indicate that the use of the model can assist medical experts in precisely detecting COVID-19 cases.
In the field of neuroscience, in order to mitigate the serious health risks that brain tumors offer, fast and accurate brain tumor detection is essential. Early detection has a major impact on treatment outcomes, in addition to allowing timely intervention. In this context, deep learning approaches have transformed the processing of brain tumor images. In particular, architectures like U-Net have proven useful in accurately identifying brain tumors from MRI scans, supplying vital data for treatment planning [24,25]. Moreover, the categorization of brain images has been improved with the use of custom CNN architectures and transfer learning methods.
For example, a study utilized transfer learning, using the advantages of existing pretrained models, in order to accurately classify brain tumors from MRI images [26]. In addition, there are multiple research projects that have focused on utilizing transfer learning methods in order to correctly identify brain tumors from MRI scans, resulting in more correct predictions than the traditional use of CNNs [27,28,29].
As deep learning techniques are still transforming medical image processing, there is an increasing need for these complicated models’ decisions to be transparent and understandable. Explainability, also known as the interpretability of artificial intelligence (AI) systems, is becoming more and more important, especially in healthcare applications where it is critical to comprehend the reasoning behind a diagnosis [30,31]. Explainable AI (XAI) aims to provide information about how and why specific conclusions are made from complicated datasets, simplifying the decision-making processes of deep learning models in the larger context of image processing.
CNN’s use of attention mechanisms is a popular example of explainability in AI. Attention mechanisms give models the ability to concentrate on particular areas of an image, offering a type of interpretability by highlighting the input elements that the model considers most important to its judgment. There are many algorithms that utilize attention mechanisms in a medical imaging context, and there is a comperative study among them examining their results [32].
Additionally, saliency maps are frequently used to show which areas of a picture contribute most to the output of the model. Saliency-based methods are much more often used in medical imaging, given the fact that the practitioner can understand the reasoning behind the prediction for every image.
Regarding brain tumor detection from images, there are many saliency-based methods used for explainability, including gradient-based or perturbation-based [33]. Local Interpretable Model-agnostic Explanations (LIMEs) [34] is a perturbation-based method that finds the segments contributing to a model’s prediction for a given image. There are already research projects that have focused on explaining decisions made from pre-trained deep learning models [35,36] using this method, but this method alone may not produce a proper explanation to make things better for the user.

2.1. Research Challenges

Using the LIMEs library for medical image interpretation poses several significant challenges. A primary obstacle is creating segments that correspond well with the actual content of the image [37]. In medical imaging, where precise details hold crucial diagnostic information, segments that fail to capture relevant features may erroneously appear as significant as those containing critical information. This mismatch can lead to confusion among users attempting to gain insights from the model’s predictions, compromising the trustworthiness of the explanations provided.
Furthermore, LIMEs is highly sensitive to even minor changes in the input image. The stability of LIMEs’ explanations is highly susceptible to issues such as the introduction of noise, which can cause significant changes in the explanations. This instability introduces uncertainty into the interpretation process, potentially leading to inconsistencies in outcomes and reducing the trustworthiness of LIMEs’ interpretations, especially in the context of medical image analysis. As a result, establishing the reliability and consistency of LIMEs’ explanations is critical for building trust in its ability to facilitate accurate and reliable diagnoses in medical imaging applications.
In addition to the challenges associated with the LIME library, the integration of post-refinement mechanisms introduces its own set of complexities. Another notable challenge is the selection and optimization of appropriate post-refinement techniques adapted to the specific properties of medical images. Given the diverse nature of medical imaging data, which ranges from changes in imaging modalities to variations in anatomical structures and diseases, establishing efficient refining methods that transfer well across different datasets poses a substantial challenge.
Furthermore, the computational strain associated with post-refinement procedures should be carefully considered, especially in real-time clinical contexts where rapid diagnostic choices are critical. Balancing the need for enhanced interpretability with the computing efficiency required for practical deployment is still a major challenge in the development of post-refinement processes for medical image analysis.

2.2. Our Contribution

Building on the existing literature, our study addresses a critical gap by introducing a novel end-to-end architecture specifically designed to enhance the explainability of deep learning models using image post-processing techniques. Unlike current frameworks, our approach systematically improves the interpretability of image-based explanations through customized mechanisms. This research provides a reliable solution by integrating a refined post-processing step into the explainability pipeline.
Our primary contribution is the creation of a robust refinement approach designed to address the challenges inherent in existing frameworks, particularly in the context of medical image analysis. By incorporating this refinement step, we want to improve the interpretability and reliability of the explanations given by machine learning models, allowing for more informed decision making in clinical contexts.
Furthermore, our findings emphasize the relevance of end-to-end solutions in deep learning explainability, highlighting the need for the efficient integration of post-processing approaches to improve model interpretation consistency. Through our proposed architecture, we seek to establish a standardized approach for refining explainability results, ultimately advancing the transparency and trustworthiness of deep learning models in medical image analysis and beyond.

3. Proposed Methodology

Let I Z w × h be a grayscale image, originating from an MRI scanner, and  t { 0 , 1 } is the decision of a deep learning model (Section 3.1) regarding the existence or not of a tumor. Then, a model-agnostic technique (Section 3.2) generates a heatmap, H R w × h , over image I. The heatmap indicates regions of interest, over I, which have contributed to generating the output t. Ideally, the most prominent regions should include a high portion of the tumor area, giving the physician a proper explanation.
The adopted approach introduces an additional refinement mechanism (Section 3.4), R ( I , H ) , which considers both I and H, as well as eliminates non-informative segments of H, based on a combination of image morphology operations and post-processing heuristics so that R ( I , H ) H ( R ) . H ( R ) is the refined version of H, retaining the most appropriate segments related to brain and tumor geometry after using the techniques explained (Section 3.3). Figure 1 demonstrates the process.

3.1. Employed Deep Learning Architectures

In this work, we handle the brain tumor detection as a binary classification problem using as input a grayscale image, say I. We try to establish a prediction model, f ( x ) t , t { 0 , 1 } , so that given the image I,
f ( I ) = 0 , if I has not brain tumor 1 , if I has brain tumor
The process incorporates the paradigm of transfer learning, since it can provide significant advantages in medical applications [38]. In particular, three deep learning pre-trained models were used: InceptionV3 [39], ResNet50V2 [40], and NasNetLarge [41].
InceptionV3 is a CNN architecture designed for efficient and accurate image classification tasks. Its modules include parallel convolutional operations, allowing the network to capture features at various scales. The architecture is characterized by the use of global average pooling, replacing fully connected layers, which aids in reducing the model’s parameter count. InceptionV3 employs rectified linear unit (ReLU) activation functions to introduce non-linearity. Batch normalization is also integrated, contributing to faster convergence during training.
ResNet50V2, part of the ResNet (Residual Network) architecture family is a deep convolutional neural network created to help with the challenges of training very deep networks. By introducing skip or residual connections, ResNet enables the flow of information directly from one non-adjacent layer to another.
NasNetLarge is a neural network architecture designed through automated architecture search methods. As opposed to manually designed architectures, NasNet is produced by utilizing techniques from reinforcement learning to explore a large search space of possible architectures. It utilizes a combination of normal and reduction cells that are repeatedly stacked to form the overall network structure. Complex patterns and representations in images are successfully captured by the architecture thanks to skip connections and effective utilization of computational resources.
In our study, we used the same parameters for each of these pre-trained models. The pre-trained layers of each model were frozen to retain their weights from the Imagenet Dataset, and a custom head was added in order to adapt to our dataset. This means that, after the first output of each pre-trained model, we pass this in new layers that combine global average pooling and desnse layers concluding with a softmax layer. The training involved utilizing the adam optimizer, as well as categorical crossentropy loss, over 10 epochs. An early stopping criterion was implemented in order to reduce overfitting.

3.2. Model’s Explainability

The Lime Image Explainer (LIE) is employed for creating an explanation, given an image I c , which has been classified as positive to cancer, i.e.,  t c = 1 . At first, given the image I c , the LIE creates a new set of images, say Z = { I c ( 1 ) , . . . , I c ( n ) } , with the same dimensions w × h . The process can be summarized as follows:
(a)
A segmentation algorithm, e.g., QuickShift, Slic, or Felzenszwalb, operates over image I c , generating d segmented areas. This number denotes the number of segments produced by the LIE using one of the above segmentation algorithms, and it can vary from image to image.
(b)
Then, a new image instance, I c ( k ) , is created by maintaining a random number, m < d , of the original segments, and once again, it depends on the number m and on the image that is used at the moment.
(c)
Repeat the process until a predefined number of images are generated. This number is taken as a parameter into the LIE. In this study, we used 1000 new generated images.
The proposed approach generates multiple copies of the image I c with missing areas, corresponding to some of the d segments. Theoretically, such a process may generate up to n = d 0 + d 1 + + d d new image instances.
Then, we use the prediction model, described in Section 3.1 to predict the outcomes using Z as inputs. Each instance I c ( k ) Z is fed to the deep learning model and the corresponding output such that t ( k ) is generated. The LIE calculates the weights corresponding to each segment area, creating a sparse linear model, which approximates the outputs of the deep learning model f ( x ) [42]. That way, the weights highlight the importance of each segment in contributing to the model’s decision, providing a local and interpretable understanding of the black box model’s behavior within a specific region of the input space.
The result is a heatmap is of the form H = { S i : I m p o r t a n c e i , , S j : I m p o r t a n c e j } , where S i represents the ID of the segment, and  I m p o r t a n c e i denotes the corresponding importance value or the weight assigned to that segment by the LIE. Our interest is in identifying the best n areas from the heatmap H provided by the explanation. The number n is selected based on various performance metrics, which are explained in Section 4.2. Figure 2 illustrates the above using a single image as an example, with the best three segments obtained from the heatmap that was produced.

3.3. Brain Area Segmentation

A quick examination of the generated heatmaps, compared to the input images, demonstrates a significant flaw in terms of explainability. It appears that certain segments were outside the bounds of the brain, having no meaningful contribution, from a medical expert perspective, in the generated explanation. As such, a mitigation strategy could be considered. In this study, multiple image operators, i.e., filters, have been considered to improve and refine the explanations’ interpretability.
The core idea lies in the successful segmentation of the brain area, given an image I, using image processing techniques. In this case, the problem at hand could be addressed as an edge detection problem. There are multiple explicit approaches for the identification of edges in an image like Laplace (Section 3.3.1), Sobel (Section 3.3.2), and Canny (Section 3.3.3), or there are implicit ones like thresholding Otsu and Li (Section 3.3.4).
Subsequently, based on the edges provided from the previous algorithms, a new binary brain mask is created. The mask is generated by identifying and extracting the largest contour from the detected edges. This mask provides a rough approximation of the brain area, and it matches the size of the original image. The resulting brain mask, denoted as BrM, serves as a crucial element in our refinement process, providing a clear binary definition of the brain’s spatial extent.

3.3.1. Laplacian Filter

The Laplace filter is one of the many methods used for edge detection.The operator relies on the second derivative to highlight sudden changes in intensity, helping identify important features in the image. The formula for the Laplace filter in image processing is as follows:
2 I ( x , y ) = i = 1 1 j = 1 1 w ( i , j ) · I ( x + i , y + j )
where 2 I ( x , y ) is the intensity of the output pixel ( x , y ) after applying the Laplace filter, I ( x + i , y + j ) represents the intensity of the input pixel at coordinates ( x + i , y + j ) , and  w ( i , j ) are the weights of the Laplacian filter mask. Typically, the weights are defined as follows:
w i , j = 0 1 0 1 4 1 0 1 0
The result of an image after applying the Laplacian filter can be seen in Figure 3.

3.3.2. Sobel

The sobel operator is another commonly used technique, for edge detection, that calculates an approximation of the gradient of the image intensity function. This technique has two kernels: the horizontal G x and the vertical G y . Typically, the two kernels are as follows:
G x = 1 0 1 2 0 2 1 0 1 , G y = 1 2 1 0 0 0 1 2 1
The intensities of each pixel, for  both the horizontal I h o r ( x , y ) and vertical direction I v e r ( x , y ) , are defined using the same formula as the Laplace operator:
I h o r ( x , y ) = i = 1 1 j = 1 1 G x ( i , j ) · I ( x + i , y + j ) , I v e r ( x , y ) = i = 1 1 j = 1 1 G y ( i , j ) · I ( x + i , y + j )
Finally, the gradient magnitude M ( x , y ) for each pixel is computed using the intensities of the image M ( x , y ) , where M ( x , y ) = ( I h o r ( x , y ) ) 2 + ( I v e r ( x , y ) ) 2 , and thresholding is then applied to the gradient magnitude in order to highlight edges. An example of vertical and horizontal sobel filters can be seen below in Figure 4.

3.3.3. Canny Edges

The canny edge detector uses a Gaussian filter to smooth the image, and then it typically uses the sobel operator in order to define the gradient magnitude and gradient direction at each pixel. The direction θ at each pixel is calculated using the I h o r and I v e r using the following formula:
θ ( x , y ) = a r c t a n ( I v e r ( x , y ) , I h o r ( x , y ) ,
where arctan is the arctangent function that returns the angle between the intensity of the pixel in the x axis and the intensity of the pixel in the y axis.
After this step, this technique involves thinning the edges to a single pixel width, iterating over all pixels in the gradient magnitude image and suppressing the gradient values of all pixels except the local maxima in the direction of the gradient. The result of this technique is similar, as shown in Figure 5.

3.3.4. Li’s and Otsu’s Thresholding

A key method in image processing is thresholding, which is used to distinguish objects or areas of interest from the background according to pixel intensity levels. It works by turning color or grayscale images into binary images, in which pixels are categorized as background or foreground (object of interest) based on whether or not they satisfy thresholds or intensity requirements.
The process of thresholding involves setting a threshold value, depending on the intensity of image’s pixels, which acts as a dividing line between the foreground and background pixels. Pixels with intensity values above the threshold are assigned to the foreground, while those below the threshold are assigned to the background. This results in a binary image, where foreground pixels are typically represented as white (or 1) and background pixels as black (or 0).
Mathematically speaking, for a given image I, a threshold t h r based on an algorithm A is found, and the result is a new image I with pixel values I ( x , y ) such that
I ( x , y ) = 0 , if I ( x , y ) < t h r 1 , if I ( x , y ) t h r
In this study, the algorithms used for thresholding were Li’s and Otsu’s thresholding.
Both algorithms use the same principles. Otsu’s thresholding aims to maximize the variance between the foreground F and background B. That is, for a given threshold t h r , the algorithm calculates the variance between the two classes as follows:
V ( t ) = p F ( t h r ) · p B ( t h r ) · [ μ B ( t h r ) μ F ( t h r ) ] 2 ,
where
  • p F ( t h r ) and p B ( t h r ) are the percentage of foreground and background pixels for the given threshold t h r ;
  • μ B ( t h r ) and μ F ( t h r ) are the mean of the background and foreground pixels intensity for the given threshold t h r ,
The threshold t h r that maximizes the variance is selected, and the result is a binary image, as explained before.
Li’s thresholding, aims to minimize the cross-entropy between the intensity values of the foreground pixels and the mean intensity value of the foreground, as well as between the intensity values of the background pixels and the mean intensity value of the background. This essentially means that the threshold should be chosen in such a way that the difference between the intensity values of pixels within a region (foreground or background) and the mean intensity value of that region is minimized.
Mathematically speaking, the algorithm tries to find the optimal threshold t h r using the following:
a r g m i n t ( H F ( t h r ) + H B ( t h r ) ) ,
where the following values are defined:
  • H F ( t h r ) is the cross-entropy between the foreground region and the foreground mean intensity using threshold t h r ;
  • H B ( t h r ) is the cross-entropy between the background region and the background mean intensity using threshold t h r .
We can take a glimpse at the produced results for a specific image after using these two techniques in Figure 6.

3.4. Post-Processing Refinement Mechanisms

After the production of the brain mask for an image, our refinement mechanism introduces a criterion for retaining segments produced by LimeImageExplainer for the same image. The LimeImageExplainer produced a heatmap H (Section 3.2). We then only retained the segments according to the following formula:
S e l e c t e d S e g m e n t s = { S i | P i x e l s ( S i B r M ) 0.8 · P i x e l s ( S i ) }
The result is a new heatmap H ( R ) , which is defined by a new dictionary that is as follows:
H ( R ) = { S i : I m p o r t a n c e i , , S j : I m p o r t a n c e j } ,
where this time, the  I m p o r t a n c e i is the same with the I m p o r t a n c e i if S e g m e n t i was retained, and it is 0 otherwise. This refined heatmap H provides a more accurate representation of the segments contributing to the model’s prediction.
Algorithm 1 summarizes the refinement process. Practically speaking, the refinement process operates over a LIMEs heatmap and recalculates the segments’ importance. Any heatmap segment that does not have a high overlap with the brain mask, produced as explained in Section 3.3, is considered non-informative, and its importance is set to 0. In this study, a threshold overlap value of 80% was selected. If a segment meets this criterion, i.e., 80% of its pixels are brain pixels, it retains its original importance, reflecting its relevance to the model’s prediction.
Algorithm 1 Post-processing refinement mechanism.
 1:
Input: H (LIMEs-generated heatmap), B r M (Brain Mask)
 2:
Output:  H ( R ) (Refined heatmap)
 3:
for each segment S i in H do
 4:
     o v e r l a p Pixels ( S i B r M )
 5:
     t h r e s h o l d 0.8 · Pixels ( S i )
 6:
    if  o v e r l a p t h r e s h o l d  then
 7:
           I m p o r t a n c e i I m p o r t a n c e i
 8:
      else
 9:
           I m p o r t a n c e i 0
10:
  end if
11:
end for
12:
H ( R ) { S i : I m p o r t a n c e i for each segment S i }
13:
return  H ( R )
We conducted a small test with approximately 50 images ranging from 50% to 90% in order to investigate the possible effects that these percentages may have. We found out that when we used a small percentage, the segments that remained were quite uninformative for the algorithm, because there were many segments that were outside of the brain area. Also, when we used a high percentage, the segments that were retained were mostly inside of the brain region, but we were losing many segments that had a small part of them outside of the brain. As such, the 80% value was selected. This method effectively filters out extraneous information, ensuring that only the most pertinent segments contribute to the final explanation. As a result, the refined heatmap provides a clearer and more focused representation of the segments that truly influence the model’s decision-making process.

4. Experimental Setup

The proposed scheme was evaluated using a publicly available dataset, which is described in the end of this article. The objective of the methodology was to refine the LIE-generated heatmaps by removing topology-related inconsistencies. All of the experiments conducted were employed using Python utilizing public libraries (Tensorflow, Skimage, Sklearn, Shapely, Lime, and Matplotlib). The computations were performed in Google Colab with GPU acceleration enabled.
At this point, we need to stress that the setup was arranged to provide robust conclusions about the impact of combining different image segmentation algorithms and edge detection techniques within the LIMEs framework. Our approach aimed to refine and enhance the interpretability of model predictions. The optimal configuration of these methodologies will be explored in future work to further improve the reliability and accuracy of tumor detection.
The following sections present the datasets employed in this study and the approach followed to form training and evaluation sets. Then, we present the details of selecting the classification model that performed best in the specific dataset. Finally, we discuss the obtained results and their implications for improving the interpretability of the deep learning models examined in our study.

4.1. Dataset Pre-Processing

Prior to training the models, the dataset underwent pre-processing steps to ensure consistency and suitability. The images were resized to 224 × 224 pixels and normalized to standardize the pixel values within the range of 0 to 1. To improve the dataset’s quality and diversity, duplicate images were removed, resulting to a total of 4015 images. Furthermore, a Stratified K-Fold validation strategy with five splits was used. This methodology ensures a robust evaluation of the deep learning models’ performance by guaranteeing that each fold retains the same distribution of classes as the original dataset.

4.2. Performance Metrics

Each deep learning model’s performance was evaluated using typical metrics like accuracy, precision, recall, and F1 score. These metrics rely on key elements such as True Positive (TP) values, True Negative (TN) values, False Positive (FP) values, and False Negative (FN) values. These metrics are defined as follows:
A c c u r a c y = T P + T N T P + T N + F P + F N
P r e c i s i o n = T P T P + F P
R e c a l l = T P T P + F N
F 1 S c o r e = 2 · P r e c i s i o n · R e c a l l P r e c i s i o n + R e c a l l
In addition to these metrics, we introduced a new metric to evaluate how well the segments from LimeImageExplainer matched the presence of tumors. Using the VGG Image Annotator, manual annotations of the tumor were performed on 271 images out of the 471 infected instances in the test set. These images were selected for their relative ease of annotation. This custom metric aims to quantify the percentage of the brain tumor included in the explanation’s segments, and we will refer to it as “Tumor Segment Coverage”.
To be more specific, after the use of the VGG Image Annotator, a new mask that represents the location of the tumor was created. This means that a pixel of the original image ( x , y ) belongs in the Tumor Mask if and only if this pixel is inside of the tumor polygon that is created. With this mask, to find the Tumor Segment Coverage (TSC), we calculated the percentage of pixels that were both part of the mask and the tumor annotation. Figure 7 demonstrates such a case with a Tumor Segment Coverage of 51.75%.
In addition we used another metric called “Brain Segment Coverage” (BSC), and it defines the percentage of brain mask covered by the explanation’s segments. This means that, after we produced the refined explanation, we calculated the percentage of the pixels that were part of the explanation and also belonged in the brain mask that was created using an edge detector, as explained in Section 3.4. Figure 8 demonstrates such a case with a brain coverage score of 21.92%.

4.3. Segmentation-Based Refinement Impact

This section demonstrates the possible impact, for the proposed refinment methodology, as discused in Section 3.4. In particular, we started by selecting the best classification model and then produced explanations for the MRI images using the LIE. We then examined the results produced after the use of our refinement method.
Table 1 and Figure 9 demonstrate the performance scores for the models’ predictions for every fold. ResNet50V2 appears to be the most prominent classifier in terms of its F1 score.
To compare the pre-trained models, we conducted a statistical Mann–Whitney test on their F1 scores, considering both the scores obtained during cross-validation and after. The test showed that ResNet50v2 outperformed both InceptionV3 (p = 0.02) and NasNetLarge (p = 0.008). However, there was no significant difference between InceptionV3 and NasNetLarge (p = 0.39). Given these results, we chose ResNet50v2 for further analysis in the explainability part.
As mentioned before, the LimeImageExplainer was employed to provide explanations for model predictions before and after the proposed refinement. We used three segmentation algorithms in order to separate the image into segments: QuickShift, Slic, and Felzenszwalb. We also used five edge detection techniques, as mentioned in Section 3.3. There were two quantitative criterion used for the performance of the adopted approaches: (a) BSC and (b) TSC. The former validates whether the applied image processing techniques correctly identify the brain areas. The latter demonstrates how practical is, for the medical experts, the n most important image segments.
We can see from the Figure 10, that when we used Quickshift as the segmentation algorithm, the LIE produced explanations that had 32.41% TSC on average, without the proposed refinement. We must note here that the LIE took the three most important areas.
Following the introduction of the refinement mechanism, using again the three most important segments, a substantial improvement was observed, with the TSC average increasing to 49.68% across the five edge detection techniques. We must also note here that Otsu’s thresholding produced the best explanations compared to the rest of the edge detection algorithms, as the TSC with this technique averaged 52.74%, when the rest averaged below 49.2%.
To determine the best number of segments for generating meaningful explanations, we explored the impact of selecting the best single, three, and five segments using the refined LIE. Examining the Tumor Segment Coverage, we found that relying on a single segment yielded an average TSC of 27.26%, and employing five segments resulted in a TSC average of 63.22%. All of the results are presented in Table 2. As we can see, Otsu’s thresholding gave the best results in all of the experiments using Quickshift.
In order to check the balance between coverage and specificity, we used the BSC, where one segment covered 10.71%, three segments covered 25.85%, and five segments covered 38.23% of the brain region on average. These results are presented in Table 3 and in Figure 11.
Regarding Felzenszwalb, our refinement seemed to work slightly better than the algorithm without it. To be more specific, when the LIE produced explanations, the average TSC was 36.43%. We must note here that this average was higher than the percentage found regarding Quickshift as the image segmenter. As noted before, the same average with Quickshift was 32.41%.
While this algorithm produced better explanations without the proposed refinement, when we tried to intervene, the new explanations produced had, once again, better TSC percentages, but the results were quite dissapointing. Our refinement showed roughly a 6% higher TSC, with the average being 42.64%.
As we can also see from the Table 4 and the Figure 12, the selection of one and five segments did not bring better results compared to the Quickshift experiment. Selecting the best segment had average TSC of 19.96%, while selecting the best five segments had only a 54.72% TSC. The same percentages in the previous experiment were approximately 10% higher in both selections. The one thing that must be noted here is that, once again, Otsu’s thresholding brought better results in the production of explanations compared to the other methods.
Consistent with the previous results, there was no improvement regarding the BSC. To be more explicit, for every experiment that we made regarding the number of segments, the percentage of the brain used was higher in all of them. As we can see from Table 5 and Figure 13, when we chose one, three, and five segments, the BSC percentages were 17.34 %, 36.23%, and 45.57%, respectively, while the same percentages using Quickshift were 10.71%, 25.85%, and 38.23%.
Not only werethe explanations produced less informative, but they used more brain area; so, in this case, this algorithm was not a better approach than Quickshift. The one thing that we should mention once more here is that, again, the proposed refinement produced better explanations than the original algorithm.
For the last experiment, Slic was used in order to separate the image into areas. This algorithm had the best results in all of the experiments and methods used. We must say here that we only compared this algorithm with Quickshift, having in mind that Felzenszwalb performed worse, so there was no need to compare it.
We can clearly see in Table 6 and Figure 14 that all of the results were elevated. Without the proposed refinement, the LIE itself was capable of producing explanations that had a 46.53% TSC on average. This percentage was more than 10% higher than both of percentages Quickshift and Felzenszwalb.
With the refinement and the selection of the three most important segments, the TSC on the produced explanations rose to 63.77%. This percentage was even higher than the previous experiments when we picked the five most important segments. This means that this algorithm with the combination of our refinement was capable of producing better explanations even with a lower number of segments picked. Indeed, even when we picked only the best segment, the TSC percentage coverage was 34.9% on average, which is a value that is higher than the average TSC percentage when we used the Quickshift algorithm without the refinement and the best three segments (32.41%). The explanations produced with the selection of the five best segments had 74.42% TSC on average, leveraging the highest TSC on all of the experiments conducted.
Examining the BSC of the explantions, as shown in the Figure 15 and the Table 7, we can see that the percentages were slightly higher than Quickshift when selecting the best three and five segments. To be more precise, selecting three and five segments resulted in BSC average percentages of 27.7% and 44.67%, respectively, and in the same cases when using Quickshift, the same percentages were 25.85% and 38.23%.
In contradiction with previous statements, the selection of the best segment had a BSC of 10.3% on average, which is lower than the percentage found from every experiment conducted before in this study. This average, combined with the fact that the average TSC when selecting the best segment was 46.53%, assures that the refinement may produce better explanations than the original algorithm, even with the selection of less segments.

4.4. Statistical Evaluation

In this section, an investigation has been conducted to identify the existence of specific combinatory approaches that outperform others. In particular, we evaluated whether certain combinations of the LIE segmentation algorithms and brain area detectors would produce, statistically speaking, better results. As such, a thorough analysis on the performance of all of our segments/refinement combinations was performed. The performance criterion was considered as the accurate detection of tumor regions. Recall that a desirable outcome would be a high overlap of tumor areas with the n top-ranked areas according to the LIE outcomes.
In Figure 16, we present varius histograms summarizing the difference in tumor segment coverage, for each image, before and after using our refinement method. A positive result indicates that the refinement approach resulted in a higher percentage of tumor-related areas to appear on the top-3 segments provided by the LIE. Each histogram represents a different combination of segment and brain region detector, providing a comprehensive view of the performance across all tested configurations.
It is important to note that a significant subset of images exhibited minimal variation in terms of the TSC, with changes of less than 1%. To better assess the effects of our refinement mechanism, we concentrated on the subset of images that demonstrated a more substantial variation in the TSC, exceeding 1%. Figure 17 demonstrates this case.
The results indicate that all of the edge detectors, in combination with the QuickShift segmentation algorithm, led to a positive improvement regarding the tumor area detection. A similar situation was observed when using Felzenszwalb as the segmenter, with the exception of the Canny edges detector for the refinement; less than 10 images obtained worse results. This is happening due to the weakness of the edge detector to create a brain mask that covers the brain area in some images (see Section 4.5). Something else that should be noticed is that, even when we saw a positive impact regarding the TSC difference, the improvement in the majority of the images ranged only from 0 to 10%.
Now, when we used Slic as the image segmenter, the histograms were quite different. We can see in each of the histograms regarding Slic that there was one or two images that had a negative impact that ranged from 50 to 100%, not being able to produce a proper explanation. The reason that this is happening, as mentioned before, is the inability of the edge detectors to produce a proper brain mask that covers the brain area. We can also see that there were about 15 images that had a negative impact ranging from 0.1 to 5%, but they were just being placed in the bin ranging from −25 to 0. Having this in mind, we can also see that the majority of the explanations had positive impacts, and most of them had an improvement in their TSCs ranging from 75 to 100%, excluding the case using Otsu’s thresholding as the edge detector.
The analysis reveals that all proposed combinations led to a general positive impact on tumor area detection. Also, our refinement approach enhanced the accuracy and reliability of identifying tumor regions, thereby validating the effectiveness of our method. When applying our refinement proccess, the regions of interest contained tumor areas in a higher percentage than before, enhancing the overall interpretability and usefulness of the model’s predictions in a clinical setting.
We also conducted a Kruskal–Wallis statistical test regarding the differences before and after the refinement method for each of the combinations used in order to check if there was a specific combination that would produce better results than the rest.
The finding of this statistical analysis did not show us which combination is better than the rest, but which combinations should be avoided. Checking previous results presented in Section 4.3, we would say that the best algorithm is Slic when talking only about the segmenter. After looking at the results presented in Figure 18 with p-values and mean differences (denoted in the parenthesis) between different combinations, we can come to a conclusion.
The first characters before the underscore denote the edge detection techniques, Canny edges (CEs), Laplacian edges (LAs), Tthresholding LI (LI), Otsu’s thresholding (OTS), and Sobel (SO). The characters after the underscore denote the segmentation algorithm used from the LIE, Slic (SL), Quickshift (QUI), and Felzenszwalb (FEL).
There was not a specific technique that outperformed the rest, but the combinations that had Felzenszwalb as the image segmenter were worse than the others. This means that, statistically speaking, if we pick Slic or Quickshift as image segmenters, we would not have a big difference for the explanations produced. But, if we choose Felzenszwalb as the segmenter, the explanations that will be produced will not give us a clear understanding behind the model’s decision.

4.5. Advantages and Limitations

The obtained results, see Section 4.3, suggest that there is potential in refining LIMEs explanations using image processing approaches. In order to quantify the improvement in interpretability of the black box models, two performance metrics must be considered simultaneously: TSC and BSC. Both values are likely to increase if we maintain more segments for the analysis. Yet, presenting a large area of healthy brain tissue as an explanation for a positive tumor detection is counterintuitive. This scenario is further explained in the text bellow.
The Slic algorithm appears to be the most appropriate image segmentation technique. The average TSC score was higher than the rest of the image segmentors. As shown in Table 6 and Table 7, both the TSC and BSC scored higher when using five segments, with average scores of 74.42% and 44.67%, respectively. Despite the high TSC score, the BSC score of 44.67% advises against using five segments. Practically speaking, almost half of the brain tissue (BSC score of 44.67%) was presented as important for the decision, and approximately one-fourth of the tumor was not included in the suggested areas (1 − TSC = 25.58%).
Considering the above situation, utilizing the top-3 segments appears to be a better alternative. TheTSC and BSC scores were 63.77% and 27.7%, respectively. In other words, the areas presented for explanation cover one-forth of the brain tissue, and approximately one-third of the tumor is not included in the suggested areas (1 − TSC = 36.23%). As such, the use of three segments emerges as an appropriate choice, finding a balance between avoiding the overuse of non-informative brain regions and offering insightful explanations.
While the proposed refinement mechanism has shown improvement in explainability, certain limitations should be acknowledged. First of all, the accuracy and precision of the initial segmentation accomplished by the chosen algorithms directly affects how effective the refinement technique is. Any inaccuracies or inconsistencies in the segmentation output can propagate through the refinement process, compromising the quality of the explanation and possibly resulting in incorrect interpretations.
Factors influencing segmentation quality include the algorithm’s parameter settings, image characteristics (i.e., resolution, contrast), and the presence of noise or artifacts in the input images. In scenarios where the segmentation algorithms fail to accurately define tumor boundaries or distinguish between tumor and non-tumor regions, the subsequent refinement process may struggle to isolate relevant segments, resulting in explanations that are incomplete or erroneous.
Another notable drawback observed in the proposed refinement mechanism is the potential inconsistency in creating the brain mask (see examples in Figure 19). The method relies on the edges detected by the different techniques and subsequently extracting the largest contour as the brain mask. However, in some instances, this process may yield inconsistent results.
One issue arises when the detected edges fail to accurately define the boundaries of the brain, resulting in incomplete or fragmented contours. Consequently, the extracted brain mask may cover only a portion of the actual brain region or extend beyond its boundaries, leading to false interpretations during explanation generation.
To address these challenges, future iterations of the refinement mechanism could explore alternative approaches for brain mask generation, such as incorporating machine learning-based segmentation methods or integrating feedback mechanisms to iteratively refine the mask based on user input. The utilization of multimodal imaging and omics data could further refine the explainability of these models by offering a more detailed and comprehensive understanding of tumor characteristics. This approach may also aid in the identification of biomarkers, enhancing the precision of brain tumor analysis. Additionally, robust preprocessing techniques and parameter tuning may help improve the reliability and consistency of edge detection algorithms, thereby enhancing the overall effectiveness of the refinement process.

5. Conclusions

Throughout this research, our primary objective was to enhance the interpretability of deep learning models in medical image analysis, particularly in the context of brain tumor detection. We aimed to address the challenge of understanding and explaining the predictions generated by some pretrained models, making an effort to bridge the gap between complex algorithmic outputs and human interpretability. Specifically, we sought to investigate the effectiveness of the LIE in providing explanations for model predictions and to propose a refinement mechanism to augment the specificity and accuracy of these explanations.
Through a series of experiments and analyses, we have demonstrated the efficacy of our proposed refinement mechanism in improving the interpretability of deep learning models for brain tumor detection. By integrating the LIE with segmentation algorithms and edge detection techniques, we achieved more precise and informative explanations for model predictions. Our results highlight the importance of refining the initial explanations provided by deep learning models, particularly in complex medical imaging tasks.
Although the results demonstrate the effectiveness of the refining mechanism, it is important to recognize the limits that have been noted, especially with regard to the consistency of brain mask production and image segmentation. To ensure reliable and effective brain mask production, further work should be focused on improving the refinement, possibly incorporating machine learning-based segmentation methods or integrating feedback mechanisms as stated before.
Overall, this work is a positive step toward improving the interpretability and transparency of deep learning models used in medical image analysis. The development of trustworthiness and the ease of integrating these models into clinical decision-making processes will depend on continuous efforts to improve explainability mechanisms as the field progresses.

Author Contributions

Conceptualization, E.P. and K.P.; methodology, E.P. and K.P.; software, E.P. and K.P.; validation, E.P. and K.P.; formal analysis, E.P. and K.P.; investigation, E.P. and K.P.; resources, E.P. and K.P.; data curation, E.P. and K.P.; writing—original draft preparation, E.P. and K.P.; writing—review and editing, E.P. and K.P.; visualization, E.P. and K.P.; supervision, E.P. and K.P.; project administration, E.P. and K.P.; funding acquisition, E.P. and K.P. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The dataset utilized in this study is the “Brain Tumor Dataset” https://www.kaggle.com/datasets/preetviradiya/brian-tumor-dataset (accessed on 10 August 2023) obtained from Kaggle. This publicly available dataset consists of 4602 MRI images capturing random instances of the brain. The images are categorized based on the presence or absence of a brain tumor, offering a wide range of samples for testing and training. The dataset includes various perspectives like axial, coronal, and sagittal views. To maintain consistency, only JPEG images in grayscale format were retained for analysis, ensuring a standardized input for the models and enhancing the reliability of the study’s outcomes.

Acknowledgments

This paper is the result of research conducted as part of the “MSc in Artificial Intelligence and Data Analytics” program at the Department of Applied Informatics, University of Macedonia.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
AIArtificial Intelligence
MRIMagnetic Resonance Imaging
CNNConvolutional Neural Network
CTGCardiotocographic
SVMSupport Vector Machine
XAIExplainable Artificial Intelligence
LIMELocal Interpretable Model-Agnostic Explanations
ReLURectified Linear Unit
ResNetResidual Network
LIELime Image Explainer
TPTrue Positive
TNTrue Negative
FPFalse Positive
FNFalse Negative
TSCTumor Segment Coverage
BSCBrain Segment Coverage

References

  1. Tran, K.A.; Kondrashova, O.; Bradley, A.; Williams, E.D.; Pearson, J.V.; Waddell, N. Deep learning in cancer diagnosis, prognosis and treatment selection. Genome Med. 2021, 13, 152. [Google Scholar] [CrossRef] [PubMed]
  2. Zhu, W.; Xie, L.; Han, J.; Guo, X. The application of deep learning in cancer prognosis prediction. Cancers 2020, 12, 603. [Google Scholar] [CrossRef] [PubMed]
  3. Katsamenis, I.; Protopapadakis, E.; Voulodimos, A.; Doulamis, A.; Doulamis, N. Transfer learning for COVID-19 pneumonia detection and classification in chest X-ray images. In Proceedings of the 24th Pan-Hellenic Conference on Informatics, Athens, Greece, 20–22 November 2020; pp. 170–174. [Google Scholar]
  4. Huang, B.; Tian, S.; Zhan, N.; Ma, J.; Huang, Z.; Zhang, C.; Zhang, H.; Ming, F.; Liao, F.; Ji, M.; et al. Accurate diagnosis and prognosis prediction of gastric cancer using deep learning on digital pathological images: A retrospective multicentre study. EBioMedicine 2021, 73, 103631. [Google Scholar] [CrossRef]
  5. Voulodimos, A.; Doulamis, N.; Doulamis, A.; Protopapadakis, E. Deep learning for computer vision: A brief review. Comput. Intell. Neurosci. 2018, 2018, 7068349. [Google Scholar] [CrossRef]
  6. Voulodimos, A.; Protopapadakis, E.; Katsamenis, I.; Doulamis, A.; Doulamis, N. Deep learning models for COVID-19 infected area segmentation in CT images. In Proceedings of the the 14th Pervasive Technologies Related to Assistive Environments Conference, Corfu, Greece, 29 June–2 July 2021; pp. 404–411. [Google Scholar]
  7. Altaf, F.; Islam, S.M.; Akhtar, N.; Janjua, N.K. Going deep in medical image analysis: Concepts, methods, challenges, and future directions. IEEE Access 2019, 7, 99540–99572. [Google Scholar] [CrossRef]
  8. Razzak, M.I.; Naz, S.; Zaib, A. Deep learning for medical image processing: Overview, challenges and the future. In Classification in BioApps: Automation of Decision Making; Springer: Berlin/Heidelberg, Germany, 2018; pp. 323–350. [Google Scholar]
  9. Uzunova, H.; Ehrhardt, J.; Kepp, T.; Handels, H. Interpretable explanations of black box classifiers applied on medical images by meaningful perturbations using variational autoencoders. In Proceedings of the Medical Imaging 2019: Image Processing Conference, San Diego, CA, USA, 16–21 February 2019; Volume 10949, pp. 264–271. [Google Scholar]
  10. Zhang, Z.; Xie, Y.; Xing, F.; McGough, M.; Yang, L. Mdnet: A semantically and visually interpretable medical image diagnosis network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 6428–6436. [Google Scholar]
  11. Dravid, A.; Schiffers, F.; Gong, B.; Katsaggelos, A.K. medxgan: Visual explanations for medical classifiers through a generative latent space. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 2936–2945. [Google Scholar]
  12. Katzmann, A.; Taubmann, O.; Ahmad, S.; Mühlberg, A.; Sühling, M.; Groß, H.M. Explaining clinical decision support systems in medical imaging using cycle-consistent activation maximization. Neurocomputing 2021, 458, 141–156. [Google Scholar] [CrossRef]
  13. Ann, K.; Jang, Y.; Shim, H.; Chang, H.J. Multi-Scale Conditional Generative Adversarial Network for Small-Sized Lung Nodules Using Class Activation Region Influence Maximization. IEEE Access 2021, 9, 139426–139437. [Google Scholar] [CrossRef]
  14. Kim, J.; Kang, S. Model-Agnostic Post-Processing Based on Recursive Feedback for Medical Image Segmentation. IEEE Access 2021, 9, 157035–157042. [Google Scholar] [CrossRef]
  15. Grassucci, E.; Sigillo, L.; Uncini, A.; Comminiello, D. GROUSE: A Task and Model Agnostic Wavelet-Driven Framework for Medical Imaging. IEEE Signal Process. Lett. 2023, 30, 1397–1401. [Google Scholar]
  16. Yang, R. Who dies from COVID-19? Post-hoc explanations of mortality prediction models using coalitional game theory, surrogate trees, and partial dependence plots. MedRxiv 2020, preprint. [Google Scholar]
  17. Peng, J.; Zou, K.; Zhou, M.; Teng, Y.; Zhu, X.; Zhang, F.; Xu, J. An explainable artificial intelligence framework for the deterioration risk prediction of hepatitis patients. J. Med. Syst. 2021, 45, 61. [Google Scholar] [CrossRef] [PubMed]
  18. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Adv. Neural Inf. Process. Syst. 2012, 25. [Google Scholar] [CrossRef]
  19. Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. In Proceedings of the 3rd International Conference on Learning Representations, ICLR 2015, San Diego, CA, USA, 7–9 May 2015; Conference Track Proceedings. Bengio, Y., LeCun, Y., Eds.; ICLR: Appleton, WI, USA, 2015. [Google Scholar]
  20. Hosny, K.M.; Kassem, M.A.; Fouad, M.M. Classification of skin lesions into seven classes using transfer learning with AlexNet. J. Digit. Imaging 2020, 33, 1325–1334. [Google Scholar] [CrossRef] [PubMed]
  21. Muhammad Hussain, N.; Rehman, A.U.; Othman, M.T.B.; Zafar, J.; Zafar, H.; Hamam, H. Accessing artificial intelligence for fetus health status using hybrid deep learning algorithm (AlexNet-SVM) on cardiotocographic data. Sensors 2022, 22, 5103. [Google Scholar] [CrossRef]
  22. Gandhi, V.C.; P Gandhi, P. Glaucoma Eyes Disease Identification: Using Vgg16 Model throughDeep Neural Network. Int. J. Comput. Digit. Syst. 2024, 16, 1–10. [Google Scholar]
  23. Hamlili, F.Z.; Beladgham, M.; Khelifi, M.; Bouida, A. Transfer learning with Resnet-50 for detecting COVID-19 in chest X-ray images. Indones. J. Electr. Eng. Comput. Sci. 2022, 25, 1458–1468. [Google Scholar] [CrossRef]
  24. Ilhan, A.; Sekeroglu, B.; Abiyev, R. Brain tumor segmentation in MRI images using nonparametric localization and enhancement methods with U-net. Int. J. Comput. Assist. Radiol. Surg. 2022, 17, 589–600. [Google Scholar] [CrossRef]
  25. Allah, A.M.G.; Sarhan, A.M.; Elshennawy, N.M. Edge U-Net: Brain tumor segmentation using MRI based on deep U-Net model with boundary information. Expert Syst. Appl. 2023, 213, 118833. [Google Scholar] [CrossRef]
  26. Özkaraca, O.; Bağrıaçık, O.İ.; Gürüler, H.; Khan, F.; Hussain, J.; Khan, J.; Laila, U.e. Multiple brain tumor classification with dense CNN architecture using brain MRI images. Life 2023, 13, 349. [Google Scholar] [CrossRef]
  27. Srinivas, C.; KS, N.P.; Zakariah, M.; Alothaibi, Y.A.; Shaukat, K.; Partibane, B.; Awal, H. Deep transfer learning approaches in performance analysis of brain tumor classification using MRI images. J. Healthc. Eng. 2022, 2022, 3264367. [Google Scholar] [CrossRef]
  28. Anaya-Isaza, A.; Mera-Jiménez, L. Data augmentation and transfer learning for brain tumor detection in magnetic resonance imaging. IEEE Access 2022, 10, 23217–23233. [Google Scholar] [CrossRef]
  29. Chelghoum, R.; Ikhlef, A.; Hameurlaine, A.; Jacquir, S. Transfer learning using convolutional neural network architectures for brain tumor classification from MRI images. In Proceedings of the IFIP International Conference on Artificial Intelligence Applications and Innovations, Neos Marmaras, Greece, 5–7 June 2020; Springer: Berlin/Heidelberg, Germany, 2020; pp. 189–200. [Google Scholar]
  30. Tjoa, E.; Guan, C. A survey on explainable artificial intelligence (xai): Toward medical xai. IEEE Trans. Neural Netw. Learn. Syst. 2020, 32, 4793–4813. [Google Scholar] [CrossRef] [PubMed]
  31. de Vries, B.M.; Zwezerijnen, G.J.; Burchell, G.L.; van Velden, F.H.; Menke-van der Houven van Oordt, C.W.; Boellaard, R. Explainable artificial intelligence (XAI) in radiology and nuclear medicine: A literature review. Front. Med. 2023, 10, 1180773. [Google Scholar] [CrossRef] [PubMed]
  32. Li, X.; Li, M.; Yan, P.; Li, G.; Jiang, Y.; Luo, H.; Yin, S. Deep learning attention mechanism in medical image analysis: Basics and beyonds. Int. J. Netw. Dyn. Intell. 2023, 2, 93–116. [Google Scholar] [CrossRef]
  33. Zeineldin, R.A.; Karar, M.E.; Elshaer, Z.; Coburger, J.; Wirtz, C.R.; Burgert, O.; Mathis-Ullrich, F. Explainability of deep neural networks for MRI analysis of brain tumors. Int. J. Comput. Assist. Radiol. Surg. 2022, 17, 1673–1683. [Google Scholar] [CrossRef]
  34. Ribeiro, M.T.; Singh, S.; Guestrin, C. “Why should i trust you?” Explaining the predictions of any classifier. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; pp. 1135–1144. [Google Scholar]
  35. Gaur, L.; Bhandari, M.; Razdan, T.; Mallik, S.; Zhao, Z. Explanation-driven deep learning model for prediction of brain tumour status using MRI image data. Front. Genet. 2022, 13, 448. [Google Scholar] [CrossRef]
  36. Haque, R.; Hassan, M.M.; Bairagi, A.K.; Shariful Islam, S.M. NeuroNet19: An explainable deep neural network model for the classification of brain tumors using magnetic resonance imaging data. Sci. Rep. 2024, 14, 1524. [Google Scholar] [CrossRef]
  37. Hryniewska, W.; Grudzień, A.; Biecek, P. LIMEcraft: Handcrafted superpixel selection and inspection for Visual eXplanations. Mach. Learn. 2022, 113, 3143–3160. [Google Scholar] [CrossRef]
  38. Maganaris, C.; Protopapadakis, E.; Bakalos, N.; Doulamis, N.; Kalogeras, D.; Angeli, A. Evaluating transferability for Covid 3D localization using CT SARS-COV-2 segmentation models. In Proceedings of the 15th International Conference on PErvasive Technologies Related to Assistive Environments, Corfu, Greece, 29 June–1 July 2022; pp. 615–621. [Google Scholar]
  39. Szegedy, C.; Vanhoucke, V.; Ioffe, S.; Shlens, J.; Wojna, Z. Rethinking the inception architecture for computer vision. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 2818–2826. [Google Scholar]
  40. He, K.; Zhang, X.; Ren, S.; Sun, J. Identity mappings in deep residual networks. In Proceedings of the Computer Vision–ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016; Proceedings, Part IV 14. Springer: Berlin/Heidelberg, Germany, 2016; pp. 630–645. [Google Scholar]
  41. Zoph, B.; Vasudevan, V.; Shlens, J.; Le, Q.V. Learning transferable architectures for scalable image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 8697–8710. [Google Scholar]
  42. Rabold, J.; Deininger, H.; Siebers, M.; Schmid, U. Enriching visual with verbal explanations for relational concepts–combining LIME with Aleph. In Proceedings of the Machine Learning and Knowledge Discovery in Databases: International Workshops of ECML PKDD 2019, Würzburg, Germany, 16–20 September 2019; Proceedings, Part I. Springer: Berlin/Heidelberg, Germany, 2020; pp. 180–192. [Google Scholar]
Figure 1. Proposed methodology.
Figure 1. Proposed methodology.
Jimaging 10 00232 g001
Figure 2. Demonstrating the overlap of the 3 most important segments (right) given an input image (left) and the LIE heatmap (center).
Figure 2. Demonstrating the overlap of the 3 most important segments (right) given an input image (left) and the LIE heatmap (center).
Jimaging 10 00232 g002
Figure 3. Edge Detection example using Laplacian Filter.
Figure 3. Edge Detection example using Laplacian Filter.
Jimaging 10 00232 g003
Figure 4. Edge detection example using sobel filters.
Figure 4. Edge detection example using sobel filters.
Jimaging 10 00232 g004
Figure 5. Edge detection example using canny filter.
Figure 5. Edge detection example using canny filter.
Jimaging 10 00232 g005
Figure 6. Generated binary masks using Li’s and Otsu’s thresholding.
Figure 6. Generated binary masks using Li’s and Otsu’s thresholding.
Jimaging 10 00232 g006
Figure 7. An example of tumor coverage calculation.
Figure 7. An example of tumor coverage calculation.
Jimaging 10 00232 g007
Figure 8. An example of brain coverage calculation.
Figure 8. An example of brain coverage calculation.
Jimaging 10 00232 g008
Figure 9. Classification performance scores for the utilized approaches.
Figure 9. Classification performance scores for the utilized approaches.
Jimaging 10 00232 g009
Figure 10. Tumor Segment Coverage average using Quickshift.
Figure 10. Tumor Segment Coverage average using Quickshift.
Jimaging 10 00232 g010
Figure 11. Brain Segment Coverage average using Quickshift.
Figure 11. Brain Segment Coverage average using Quickshift.
Jimaging 10 00232 g011
Figure 12. Tumor Segment Coverage average using Felzenszwalb.
Figure 12. Tumor Segment Coverage average using Felzenszwalb.
Jimaging 10 00232 g012
Figure 13. Brain Segment Coverage average using Felzenszwalb.
Figure 13. Brain Segment Coverage average using Felzenszwalb.
Jimaging 10 00232 g013
Figure 14. Tumor Segment Coverage Average using Slic.
Figure 14. Tumor Segment Coverage Average using Slic.
Jimaging 10 00232 g014
Figure 15. Brain Segment Coverage average using Slic.
Figure 15. Brain Segment Coverage average using Slic.
Jimaging 10 00232 g015
Figure 16. Improvements over the tumor segment coverage before and after the refinement process.
Figure 16. Improvements over the tumor segment coverage before and after the refinement process.
Jimaging 10 00232 g016
Figure 17. Improvements over the Tumor Segment Coverage before and after the refinement process for images with absolute difference values greater that 0.01.
Figure 17. Improvements over the Tumor Segment Coverage before and after the refinement process for images with absolute difference values greater that 0.01.
Jimaging 10 00232 g017
Figure 18. Statistical measurements between techniques, with respective p-values and mean differences.
Figure 18. Statistical measurements between techniques, with respective p-values and mean differences.
Jimaging 10 00232 g018
Figure 19. Instances of wrong brain mask production.
Figure 19. Instances of wrong brain mask production.
Jimaging 10 00232 g019
Table 1. Performance metrics for each pre-trained model after k-fold validation.
Table 1. Performance metrics for each pre-trained model after k-fold validation.
InceptionResNet50v2NasNetLarge
Precision0.93970.98260.9473
Accuracy0.95970.95960.9711
Recall0.94020.96640.9536
F1 Score0.94960.9710.9591
Table 2. Average metrics for Tumor Segment Coverage using Quickshift as image segmenter.
Table 2. Average metrics for Tumor Segment Coverage using Quickshift as image segmenter.
Tumor Segment Coverage %
Edge DetectorNo Refinement1 Segments3 Segments5 Segments
Canny32.4126.4148.962.7
Laplace32.4126.949.2163.11
Li32.4127.1449.2062.82
Otsu32.4129.5852.7465.72
Sobel32.4126.348.3861.74
Average32.4127.2649.6863.22
Table 3. Average metrics for Brain Segment Coverage using Quickshift as image segmenter.
Table 3. Average metrics for Brain Segment Coverage using Quickshift as image segmenter.
Brain Segment Coverage %
Edge Detector1 Segments3 Segments5 Segments
Canny10.4425.3337.62
Laplace10.9426.4439.15
Li10.2524.8136.92
Otsu11.6827.7140.34
Sobel10.2224.9737.13
Average10.7125.8538.23
Table 4. Tumor Segment Coverage using Felzenszwalb as image segmenter.
Table 4. Tumor Segment Coverage using Felzenszwalb as image segmenter.
Tumor Segment Coverage %
Edge DetectorNo Refinement1 Segments3 Segments5 Segments
Canny36.4318.7541.755.08
Laplace36.4318.241.3152.45
Li36.4320.3243.3356.17
Otsu36.4324.4346.7459.42
Sobel36.4318.140.1150.51
Average36.4319.9642.6454.72
Table 5. Average metrics for Brain Segment Coverage using Felzenszwalb as image segmenter.
Table 5. Average metrics for Brain Segment Coverage using Felzenszwalb as image segmenter.
Brain Segment Coverage %
Edge Detector1 Segments3 Segments5 Segments
Canny16.635.0944.18
Laplace17.3235.8644.47
Li16.8735.0444.63
Otsu18.6439.9749.97
Sobel17.2535.1844.61
Average17.3436.2345.57
Table 6. Tumor Segment Coverage using Slic as image segmenter.
Table 6. Tumor Segment Coverage using Slic as image segmenter.
Tumor Segment Coverage %
Edge DetectorNo Refinement1 Segments3 Segments5 Segments
Canny46.5335.4563.9874.41
Laplace46.5333.9663.9674.34
Li46.5334.7963.7474.22
Otsu46.5337.5464.6775.37
Sobel46.5332.7562.5273.76
Average46.5334.963.7774.42
Table 7. Average metrics for Brain Segment Coverage using Slic as image segmenter.
Table 7. Average metrics for Brain Segment Coverage using Slic as image segmenter.
Brain Segment Coverage %
Edge Detector1 Segments3 Segments5 Segments
Canny9.8326.8343.63
Laplace10.5728.0745.1
Li9.3525.9342.83
Otsu11.9131.248.86
Sobel9.8626.4642.91
Average10.327.744.67
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Pasvantis, K.; Protopapadakis, E. Enhancing Deep Learning Model Explainability in Brain Tumor Datasets Using Post-Heuristic Approaches. J. Imaging 2024, 10, 232. https://doi.org/10.3390/jimaging10090232

AMA Style

Pasvantis K, Protopapadakis E. Enhancing Deep Learning Model Explainability in Brain Tumor Datasets Using Post-Heuristic Approaches. Journal of Imaging. 2024; 10(9):232. https://doi.org/10.3390/jimaging10090232

Chicago/Turabian Style

Pasvantis, Konstantinos, and Eftychios Protopapadakis. 2024. "Enhancing Deep Learning Model Explainability in Brain Tumor Datasets Using Post-Heuristic Approaches" Journal of Imaging 10, no. 9: 232. https://doi.org/10.3390/jimaging10090232

APA Style

Pasvantis, K., & Protopapadakis, E. (2024). Enhancing Deep Learning Model Explainability in Brain Tumor Datasets Using Post-Heuristic Approaches. Journal of Imaging, 10(9), 232. https://doi.org/10.3390/jimaging10090232

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop