Next Article in Journal
Generative AI Agents for Bedside Sleep Apnea Detection and Sleep Coaching
Previous Article in Journal
SkinVisualNet: A Hybrid Deep Learning Approach Leveraging Explainable Models for Identifying Lyme Disease from Skin Rash Images
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Multi-Criteria Decision-Making Approach for the Selection of Explainable AI Methods

Department of Cybernetics and Artificial Intelligence, Faculty of Electrical Engineering and Informatics, Technical University of Kosice, Letna 9, 040 01 Košice, Slovakia
*
Author to whom correspondence should be addressed.
Mach. Learn. Knowl. Extr. 2025, 7(4), 158; https://doi.org/10.3390/make7040158
Submission received: 31 October 2025 / Revised: 25 November 2025 / Accepted: 27 November 2025 / Published: 1 December 2025

Abstract

The growing trend of using artificial intelligence models in many areas increases the need for a proper understanding of their functioning and decision-making. Although these models achieve high predictive accuracy, their lack of transparency poses major obstacles to trust. Explainable artificial intelligence (XAI) has emerged as a key discipline that offers a wide range of methods to explain the decisions of models. Selecting the most appropriate XAI method for a given application is a non-trivial problem that requires careful consideration of the nature of the method and other aspects. This paper proposes a systematic approach to solving this problem using multi-criteria decision-making (MCDM) techniques: ARAS, CODAS, EDAS, MABAC, MARCOS, PROMETHEE II, TOPSIS, VIKOR, WASPAS, and WSM. The resulting score is an aggregation of the results of these methods using Borda Count. We present a framework that integrates objective and subjective criteria for selecting XAI methods. The proposed methodology includes two main phases. In the first phase, methods that meet the specified parameters are filtered, and in the second phase, the most suitable alternative is selected based on the weights using multi-criteria decision-making and sensitivity analysis. Metric weights can be entered directly, using pairwise comparisons, or calculated objectively using the CRITIC method. The framework is demonstrated on concrete use cases where we compare several popular XAI methods on tasks in different domains. The results show that the proposed approach provides a transparent and robust mechanism for objectively selecting the most appropriate XAI method, thereby helping researchers and practitioners make more informed decisions when deploying explainable AI systems. Sensitivity analysis confirmed the robustness of our XAI method selection: LIME dominated 98.5% of tests in the first use case, and Tree SHAP dominated 94.3% in the second.

1. Introduction

In an era of increasingly widespread use of artificial intelligence in practice, it is important to ensure that its use is safe, fair, and non-discriminatory. The danger lies in generating and using results that are not justified, legitimate, or that do not allow for detailed explanations of their behavior [1]. While the very first artificial intelligence systems were easy to interpret, in recent years, more complex models have been on the rise, which are becoming opaque. There is a need to better understand the functioning of such models and their results.
The solution is provided by explainable artificial intelligence (XAI), which offers explanations of model predictions. Data scientists are and will increasingly be confronted with stakeholder demands for explanations beyond the traditional focus on the predictive performance of models.
The terms interpretability and explainability are often used interchangeably by researchers [2]. These concepts are very closely related, and some works identify their differences and distinguish between the two. It is difficult to define explainability and interpretability mathematically. Miller [3] offers a non-mathematical definition of interpretability, where interpretability is the degree to which a human can understand the reason for a decision. Another definition states that interpretability is the degree to which a human can consistently predict the outcome of a model [4]. One of the most popular definitions of interpretability is that of Doshi-Velez and Kim, who in their work [5] define it as “the ability to explain or present to a human in terms that are understandable”. These authors consider interpretability to be a broader concept than explainability. Gilpin et al. [6] describe interpretability as a necessary condition for explainability. According to them, the interpretability of an explanation describes how understandable the explanation is to humans. In our opinion, the clearest definition is that of the authors of [7], according to which interpretability refers to the passive characteristic of a model, which refers to the level at which the model makes sense to humans. This property is also expressed as transparency. Explainability can be considered as an active characteristic of a model, denoting any action or procedure that a model performs with the intention of explaining or detailing its internal workings.
The main advantages of using explainable artificial intelligence methods include solving the “black box” problem of modern complex models (such as deep neural networks), making their predictions understandable, transparent, and trustworthy. Research has so far revealed various goals that can be achieved by creating an explainable model. The primary goal of XAI is to obtain human-interpretable models because domain experts require assistance in solving issues more efficiently, but they also want to receive meaningful output so they can comprehend and trust those solutions. Additionally, researchers use XAI to achieve a variety of goals, such as enhanced justification, control, improvement, and discovery [8]. Other authors have specialized these goals to the following [9]:
  • Empower individuals to combat any negative consequences of automated decision-making.
  • Help individuals make more informed decisions.
  • Detect and prevent security vulnerabilities.
  • Integrate algorithms with human values.
  • Improve industry standards for the development of AI-based products, thereby increasing consumer and business trust.
  • Promote a Right of Explanation policy.
Arrieta et al. [7] identified nine main goals: trustworthiness, causality, portability, informativeness, reliability, fairness, accessibility, interactivity, and privacy awareness. They assigned these goals to their main target group, as the area of explainability is closely linked to users and their desiderata.
As is evident, the primary role of XAI goes beyond the technical interpretation of models; XAI serves as a crucial bridge for ensuring human understanding and trust in increasingly complex automated decision-making systems. The broad set of identified objectives—from improving control and knowledge discovery to ensuring compliance with the right to explanation—clearly demonstrates that XAI is a cornerstone for responsible, ethical, and acceptable AI development. Ultimately, integrating causality, reliability, and fairness into AI through XAI is not just a technical matter but a strategic imperative for its successful and safe deployment in critical domains.
In recent years, a number of methods have been proposed to explain machine learning models and their decisions. Explainability methods differ in many criteria and are used in a wide range of tasks [10]. The main division of methods is offered by a basic taxonomy that focuses on different characteristics. The question is the right choice of method for a specific use case. As the number of methods increases, it becomes increasingly difficult for stakeholders to choose the right explainability method for a specific use scenario. This carries the risk of failing to achieve reliable interpretability of the model [11]. Our solution to this problem is the proposed framework for selecting the right method that takes into account all important aspects.
The structure of this work is designed to systematically move from theoretical foundation to applied solution: Section 2 presents a basic taxonomy of XAI methods, while Section 3 critically analyzes related works in the field of XAI and MCDM methodology, including XAI benchmarks and metrics. Subsequently, Section 4 defines the methods used. These tools are integrated in Section 5, which presents a proposed solution for objective weighting and ranking of XAI methods. The whole work is concluded in Section 6, Discussion, where the key findings, their contribution, and implications for the responsible deployment of artificial intelligence are interpreted.

2. Basic Taxonomy of XAI Methods

Various and significantly different methods have been developed in the field of explainability [2,12,13]. As explainable methods have become more popular, the number of such methods has also increased, making it more difficult to navigate among them. There are different perspectives on dividing the wide range of explainability methods. The basic pillar is the taxonomy described in the following section. This basic taxonomy is shown in Figure 1.

2.1. Intrinsic and Post hoc Methods

The first criterion is the phase in which the explanation is generated. Accordingly, we divide methods into those in which the explanation is generated during the creation of the model (intrinsic) and those in which the explanation is generated later (post hoc) [13]. Simple models that are interpretable due to their uncomplicated structure, such as decision trees or linear models, belong to the intrinsic group. This form of interpretability is also defined as model transparency and describes how the model works [14]. Post hoc explainability refers to explanatory methods that are applied after a model has been trained and include techniques that allow the conversion of an uninterpretable model into an explainable one.

2.2. Model-Specific and Model-Agnostic Methods

Post hoc methods are further divided into model-specific and model-agnostic methods based on whether they are applicable to specific or arbitrary models. The first group of methods is based on the intrinsic properties of a particular type of model, which limits their applicability. Model-agnostic methods are applicable to any machine learning model after it has been trained. These methods analyze the input and output values and do not have access to the inner workings of the models. Typical representatives of model-specific methods include Grad-CAM [15], LRP [16], and DeepLift [17]. On the other hand, there are model-agnostic methods such as LIME [18], SHAP [19], and Anchors [20].

2.3. Local and Global Methods

If a method explains an individual prediction, it is a local method. It usually tries to approximate the behavior of the model around the instance that the user wants to explain in order to gain information about how the model works. Global methods describe the average behavior of the model as a whole and are often expressed as expected values based on the data distribution. Local methods include LIME [18], Saliency maps [21], and IntGrad [22]. Global methods include ICE [23], PDP [24], and ALE [25].

2.4. Methods by Input Data

Along with the previous criteria, several explainability studies [9,26,27] have also mentioned input and output data as essential factors. Explainable methods differ based on the type of input data, such as images, texts, or tabular data. Each of the input modalities requires different procedures for creating an explanation.

2.5. Methods by Explanation Type

The output of explainability methods is explanations, which can have different formats, such as numerical data, textual data, visualizations, rules, or a combination of the aforementioned [28]. Arrieta et al. [7] describe different types of explanations that differ in their characteristics. They mention textual and visual explanations, but other types are explanation by example and explanation by simplification.

3. Related Work

The aim of this section is to analyze the existing research that deals with the key areas of this study: the selection of the right Explainable Artificial Intelligence method and the application of multi-criteria decision-making. A sequential literature review of these two domains will allow us to comprehensively identify the current state of knowledge and, at the same time, point out the gaps. Although there are many works dedicated to the development of new XAI methods, only a few of them deal with the correct selection of XAI according to the criteria. Therefore, this section serves as a basis that justifies the need for the proposed solution that combines the selection of XAI methods with MCDM, thereby providing a transparent and robust framework for the selection of the optimal XAI method.

3.1. Selecting the XAI Method

An interesting approach to developing a methodology for the correct method selection was introduced by Vermeire et al. [11]. In their research, they argue that a methodology is needed to bridge the gap between stakeholder needs and explainability methods. They designed a card that provides a detailed overview of stakeholder needs. To identify these needs, they created a questionnaire designed to capture stakeholder requirements.
Typical categorizations of stakeholders are based on their role in the organization, their experience with machine learning, or a combination of both. Different methods are implemented to collect needs from stakeholders in a particular use case. Most of them use an approach based on information systems research and software development, where user requirements gathering is a well-known and studied problem. Sometimes, data scientists may have difficulty discussing XAI solutions directly with stakeholders, and therefore, they suggest focusing more on what the stakeholder wants to achieve with explainability [29,30]. They try to understand the background, capabilities, and goals of the stakeholder well. In this way, they create textual and/or visual scenarios that describe where and when explainability is needed.
XAI methods differ significantly in terms of the output/explanation and the way in which these explanations are generated. Therefore, they may be more or less suitable for a particular use case and/or stakeholder. There is therefore a risk that explainability will not be used appropriately. This problem has been noted in other works [31,32], but the existing literature lacks a specific methodology. On the other hand, various studies are being produced devoted to the characterization of XAI methods [33]. These allow us to create a generalized view of explainability methods, which can serve as documentation and a means of comparing them. Several experts [34,35] have proposed a framework of XAI method characteristics that can be supplemented for a specific explainability method. Hall et al. [36] proposed an approach that consists of characteristics that are divided into dimensions: efficiency, versatility, limitations, types and categories of XAI methods, explanation properties, and personal considerations. Sokol and Flach [34] created a framework with five dimensions of the so-called XAI method requirements. Functional requirements determine whether it is practically possible to use the explanation method for a specific use case. Operational requirements concern the interaction of users with the system. Usability requirements consist of properties that are important for the recipient of the explanations. Security and validation requirements focus on aspects such as privacy, security, and validation of the explanation method. It is clear that the dimensions in both studies not only cover general properties of explainability methods but also, to some extent, the needs of stakeholders. According to the authors of [11], there is a lack of research in the professional literature on how to perform a mapping between explainability needs and the properties of the explanation method.
To address the challenge of choosing the right method, a novel framework, AutoXAI [35], was conceptualized and developed. The primary objective of AutoXAI was to automate these complex tasks, thereby assisting data scientists in selecting the optimal XAI solutions based on their specific context. This context was defined as a set of critical elements, including the dataset characteristics, the machine learning model architecture, and the user’s specific XAI needs and constraints. The design of AutoXAI drew upon two distinct methodological domains. Firstly, it leveraged strategies from context-aware recommender systems [37]. This approach allowed for the integration of the user’s context across three distinct phases: contextual prefiltering (to select an initial candidate subset), contextual modeling (to embed context into the recommendation process), and contextual postfiltering (to adjust the final recommendation). A significant preliminary challenge encountered was the lack of unified formalization for XAI elements, which necessitated a rigorous formal definition of the context elements within the framework itself.
Secondly, AutoXAI adapted optimization and evaluation strategies from the field of automated machine learning (AutoML) [38]. This inspiration was crucial because suggesting a reliable explanation required verifying multiple properties of interest simultaneously, which was achieved by optimizing corresponding XAI evaluation metrics. Given the computational expense inherent in hyperparameter optimization using multiple metrics, AutoML techniques were adopted to implement time-saving strategies for efficient evaluation. This combined approach ensured that the recommended XAI solution was both adapted to the user’s context and validated against desired properties. The advantage of this solution is the optimization of hyperparameters of the methods, but the disadvantage is the small number of implemented methods (only LIME, SHAP, and Protodash).
In response to this methodological gap, eXplego [39] was developed and introduced as an interactive, tree-structured tool designed to streamline the XAI method selection process. It operates as a decision tree toolkit that provides developers and practitioners with guided, interactive assistance. Through a number of useful requirements that users must take into account when choosing XAI techniques, eXplego offers navigation to a variety of XAI techniques. The tool recommends an XAI approach that is in line with the user’s explainability requirements after two to five questions. These questions focused on the desired type of explanation, the scope of the explanation, and the nature of the data/model. The final leaf of the decision tree recommended the XAI method deemed most appropriate for the stated use case. It contains 13 different methods that are designed for tabular data. The advantage of this solution is the descriptions of each method, along with sources referring to articles and software implementations. Beyond mere recommendation, the tool provided detailed supporting information, including simplified descriptions of the method’s functionality, practical examples, key points to consider during implementation, and links to the original academic paper and software resources. The disadvantage is its sole focus on tabular input data and the relatively low number of recommended methods.
In [33], Arya et al. presented the AI Explainability 360 software tool, along with a taxonomy of explainable AI methods. The authors proposed a taxonomy of explanations that helps users navigate different approaches and decide which method is appropriate for their context. They also created a guidance taxonomy tree for better orientation. The project aims to address the growing demands of various stakeholders (regulators, users, developers, domain experts) for explanations of AI system decisions. The tool includes twenty-one different explanation methods and two metrics for evaluating explanations. Supports tabular, text, images, and time series data and has an extensible architecture so that new methods and metrics can be added.
In the study by Nauta et al. [40], the authors systematically reviewed more than 600 scientific papers published between 2014 and 2020 that deal with explainable artificial intelligence methods, 312 of which were original XAI approaches. They developed a filtering tool that serves to sort and select relevant scientific papers according to predefined criteria. These criteria include input data type, model type, method type, explanation type, problem type, and task type. Additional criteria include publication source, year of publication, and text search for title, authors, or abstract. The result is a list of relevant studies that can be exported as a list in JSON format. The tool is useful for exploring different methods, but does not offer any further evaluation.
In summary, the research to date has yielded a rich range of theoretical taxonomies and diverse approaches to the characterization and selection of XAI methods. Although these works have significantly contributed to the transparency of AI models, all of the proposed frameworks for selecting the optimal XAI method have their shortcomings. As a result, the decision-making process for selecting an XAI technique often remains incomplete and lacks the required transparency. This is where multi-criteria decision-making intervenes, which has the potential to fill this gap.
Identical predictions can have conflicting and diverse explanations from many explainable methods [41]. To improve explanation quality, some methods aggregate multiple explanations to create a more robust explanation [42]. In [43], the authors proposed a method combining MCDM and XAI. They proposed a method to aggregate multiple explanation models, aiming to enhance the overall robustness of explanations. To maintain purely rank-based procedures, they developed rank-based versions of existing XAI metrics for complexity, stability, and faithfulness. Once the explanations from the component models were evaluated against these metrics, an MCDM algorithm was used to quantify the performance of each component using scalar weights, which were later combined by a rank aggregation algorithm to form a single explanation. Eight MCDM algorithms were considered: EDAS [44], TOPSIS [45], COPRAS [46], PROMETHEE II [47], ARAS [48], COCOSO [49], CODAS [50], and MABAC [51]. Experiments comparing MCDM and rank aggregation algorithms revealed TOPSIS and the weighted sum method to be the best candidates for this use case. A comprehensive experimental analysis across five datasets demonstrated the technique’s effectiveness in enhancing explanation robustness. The study combined the results of several XAI methods and compared several MCDM methods. However, the result is not the selection of the best method, but the ranking of individual features.
Another study that combines XAI methods and MCDM is [52]. The study deals with the development of a sophisticated clinical decision support system (CDSS) that addresses the critical problem of trust in artificial intelligence models in complex domains such as medicine. The authors designed and implemented the CDSS-EQCM framework, which uses an axiomatic integration of explainable AI and multi-criteria decision-making methods to achieve this goal. Specifically, the system generated transparent explanations of models using the model-agnostic techniques LIME and SHAP, and these models were subsequently evaluated according to a set of multiple criteria. To aggregate this multi-criteria evaluation and determine the final ranking of the models, an MCDM approach with TOPSIS and Borda count [53] methods was used. This approach allowed the system to effectively recommend not only the most accurate, but also the most reliable and understandable model to the end user, thereby increasing transparency and the likelihood of AI adoption in clinical practice. The benefit of the work is on-site manual acceptance testing, but the disadvantage is the small number of compared XAI methods and low variability in the evaluation results.
With its ability to systematically process and objectively weigh multiple conflicting criteria and derive a ranking of alternatives, MCDM offers a robust and transparent framework for selecting the optimal XAI method. The aforementioned studies combine MCDM and XAI methods, but they are either focused on evaluating the results of explanations or include only a small number of methods. We propose a solution that will be suitable for directly selecting the most appropriate XAI method in a given context and can be used to compare a larger number of methods.

3.2. Evaluation of XAI Methods

Evaluating explainable artificial intelligence methods requires defining key features that allow us to assess the quality, reliability, and practical utility of the generated explanations. These features provide a foundation for designing metrics to evaluate the extent to which explanations improve model understanding, increase user trust, and facilitate transparent decision-making.
Nauta et al. [40] proposed a scheme of 12 conceptual features (Co-12) to serve as a comprehensive catalogue of criteria for evaluating the quality of an explanation, going beyond fidelity and robustness. These features provide researchers with specific criteria that should be quantified. The authors divided the Co-12 explanation quality properties into three groups according to their focus:
  • Content: Correctness, Completeness, Consistency, Continuity, Contrastivity, Covariate complexity;
  • Presentation: Compactness, Composition, Confidence;
  • User: Context, Coherence, Controllability.
The study critically analyzes evaluation practices in over 300 papers and reveals that a high percentage (approximately one in three papers) rely exclusively on anecdotal evidence to evaluate XAI methods, i.e., a subjective selection of “good” examples of explanations that seem plausible. The paper calls for a paradigm shift toward objective, quantifiable, and comparable evaluation methods.
A systematic approach to the assessment of explainability was introduced by Doshi-Velez and Kim [5], who presented a multi-level view of the XAI assessment that integrates objective and subjective aspects of explainability. This framework still forms the theoretical basis for most recent XAI evaluations. They divided the assessment into three main categories: application-grounded evaluation, human-grounded evaluation, and functionally grounded evaluation. This division clearly shows that no type of assessment is universally sufficient—to achieve reliable explainability, it is necessary to combine functional metrics with user studies and validate explanations in specific application contexts.
  • Application-grounded evaluation: This involves implementing models and testing them on a real-world task by running experiments with end users. The best way to show that a model works is to evaluate it on the task for which it was created.
  • Human-grounded evaluation: This level is also tested in practice, the difference being that these experiments are not performed with domain experts, but with laypeople. Since no domain experts are needed, the experiments are cheaper, and it is easier to find more testers.
  • Functionally grounded evaluation: This level does not require human evaluation. Instead, it uses a formal definition of explainability as a model of the quality of explanation. This evaluation is most appropriate when we have models that have already been validated, e.g., through human-based experiments.
We can combine application-grounded evaluation and human-grounded evaluation into one group of human-centered evaluations. These evaluations are inherently subjective because the results of the methods depend on the selected sample of people.
Functionally grounded assessment cannot replace human-centered assessment, but it is objective and can help guide the selection of explanations that experts can use in a user study.

3.2.1. Objective XAI Metric

As highlighted in the previous section, part of the responsible deployment of XAI is the use of objective metrics that can independently measure the quality and practical feasibility of explanations. These metrics translate abstract properties into measurable numerical values. The aim of this subsection is to introduce some basic objective metrics.
One of the basic properties is fidelity (also called faithfulness, correctness or truthfulness) [40,54,55,56], which determines the extent to which the explanation faithfully reflects the internal behavior of the model. An explanation with high fidelity should be consistent with the model—that is, if the model changes its output, the explanation should respond appropriately. This property is often considered the most important technical characteristic of XAI methods, as it determines their credibility in terms of correctly representing the decision process.
Another key property is stability (also called robustness, continuity, or sensitivity) [28,57,58], which describes how consistent the explanations are with small changes in the input data. If two similar instances lead to significantly different explanations, the method loses credibility. Stability is therefore essential for the practical deployment of XAI in sensitive areas such as healthcare, finance, or autonomous systems.
Consistency [40] is closely related to stability, but it evaluates whether identical inputs have identical explanations. In practice, this property addresses the extent to which the explanation method is deterministic. Low consistency indicates that the explainer is sensitive to noise, and its outputs may not be trustworthy for the decision maker in practice.
Computation complexity [59] is one of the basic technical metrics for evaluating XAI methods, which assesses the practical feasibility of explainability in real systems. This metric determines how much computational resources (time, memory, number of model calls) are required to generate an explanation for a given input. In the context of explainable AI, it is important because many popular methods are conceptually simple but computationally intensive, which limits their use in systems that require fast response or processing of large data.
A related metric is speed [60], which represents one of the most practically important, but less formalized in the literature, criteria for evaluating methods (XAI). It expresses the time required to generate an explanation. In contrast to computational complexity, which theoretically describes the behavior of an algorithm, the speed metric focuses on empirical performance in a real environment, i.e., how quickly an XAI method can provide an interpretation in practice.

3.2.2. Subjective XAI Metric

On the user side, the so-called human-centered properties are evaluated, such as interpretability, understandability, trustworthiness, usefulness, and user satisfaction. These properties measure whether the explanation is understandable to a human, whether it helps to better understand the behavior of the model, and whether it supports rational decision-making in the context of the task.
Interpretability [61] measures the extent to which the user understands the principles or reasons for the model’s decision. It is assessed, for example, using questionnaires, comprehension tests, or qualitative interviews.
Understandability (Comprehensibility) [13] measures how easily and quickly a user can understand the logic and content of the explanation. Another definition describes it as the capacity of a method to make a model understandable [28].
Trustworthiness [7] determines how the explanation affects the user’s trust in the model. This metric is often measured by a subjective scale or through decision-making experiments, in which it is observed whether the user prefers to rely on the model or on their own judgment.
Usefulness [54] assesses whether the explanation helps the user to make an effective decision or improve performance on a given task. This metric tests the practical value of the explanation in real-world scenarios.
User Satisfaction [62] determines the subjective satisfaction of users with the form and content of the explanation. It can be supplemented with qualitative responses about the clarity or visual appropriateness of the explanation. This is the overall emotional and cognitive acceptance of the explanation.

3.2.3. Benchmarks and Libraries Focused on Evaluating XAI Methods

In recent years, several comprehensive benchmarks and libraries have emerged aimed at systematically evaluating and comparing explainable artificial intelligence methods. Together, these cover a wide range of approaches to XAI evaluation. Their comparison is summarized in Table 1. Extensibility may mean the ability to add new models, methods, or metrics.
The M4 benchmark [63] provides a unified framework for comparing faithfulness across different modalities (text, image, tabular data). Quantus [64] is similarly multimodal, providing an open library of over thirty quantitative metrics grouped into six categories: faithfulness, robustness, localization, randomization (sensitivity), complexity, and axiomatic. It also supports a large number of methods from the Captum, Zennit, and tf.explain libraries. The BEExAI benchmark [65] focuses on faithfulness, robustness, and complexity metrics. They leverage 50 popular datasets covering regression and classification tasks to provide a rigorous and repeatable benchmark. Newer approaches, such as FUNCXAI-11 and XAI-Units, extend the assessment to include functional and cognitive dimensions. The functionally grounded benchmark framework developed by the developers of FUNCXAI-11 [66] is hierarchically structured and offers a clear and layered organization of properties (organizing 24 properties into 11 separate property categories). The metrics examined include speed, stability, truthfulness, fidelity, and interactivity. The XAI-Units benchmark [67] is an open benchmark framework designed to objectively and reliably evaluate feature attribution (FA) methods. The main driving force is the fact that several FA techniques frequently yield contradictory significance scores for the same model, and there is no trustworthy “ground truth” to confirm which approach is accurate. By offering a collection of procedurally produced models and synthetic datasets where the internal mechanisms and behavior of the model are understood, XAI-Units tackles this issue in a manner akin to software engineering unit testing. Compare-xAI [68] is a benchmark designed to uniformly quantitatively evaluate and compare different post-hoc methods. It uses a set of functional tests to assess specific aspects of explanation and compresses the results into a single multidimensional score, helping users choose the most appropriate XAI algorithm for their needs. Furthermore, XAIB [69] offers an open and extensible benchmark with an evaluation ontology taken from Co-12, thus contributing to the standardization of XAI evaluation in both academic and industrial contexts. Not all of the 12 properties for each explainer type were addressed during the implementation of metrics; some may be inapplicable, while others have not yet been covered. Correctness, continuity, contrastivity, covariate complexity, compactness, and coherence cases are included. XAI-Bench [70] uses synthetic data with a known ground truth, which allows for efficient and reliable evaluation of the quality of explanations using metrics such as fidelity and monotonicity. OpenXAI [71] is an open source framework designed for systematic, reproducible, and transparent benchmarking of post hoc explainable AI methods, primarily for feature attribution. The library provides a comprehensive collection of real and synthetic datasets, pre-trained models, and quantitative metrics for assessing the faithfulness, stability, and fairness of explanations. Saliency-Bench [72] is specifically designed for visual models. The benchmark includes a standardized and unified evaluation pipeline for assessing faithfulness and alignment of the visual explanation, providing a holistic visual explanation performance assessment. On the other hand, OpenHEXAI [73] represents a human-centered approach that measures factors such as trustworthiness, understandability, and user satisfaction through user experiments, reflecting the growing interest in subjective evaluation of explainability. The authors based their work on OpenXAI [71], and in total, they prepared 16 subjective survey questions. The research by the authors of [74] was also subjectively focused, using a questionnaire to evaluate XAI methods and asking about the metrics of understandability, usefulness, trustworthiness, informativeness, and satisfaction through simple questions and a Likert scale.
Together, these benchmarks and libraries reflect a shift from isolated tests to multidimensional and reproducible evaluation of explainability that combines technical, cognitive, and computational aspects into a unified framework.

4. Methods

4.1. Multi-Criteria Decision-Making

Multi-criteria decision-making, often referred to as multi-criteria decision analysis (MCDA), represents a significant advance in the field of decision-making [75,76]. Instead of relying on a single factor, MCDM immediately considers several different factors—whether these are easily measurable quantitative criteria or qualitative criteria [77]. Finding the optimal solution is therefore inherently dependent on achieving a compromise guided by the explicit preferences of the decision maker. This approach necessitates a clear process for evaluating criteria, which is typically achieved through expert groups providing weighting schemes to reflect the relative importance of each factor in the specific case under study. A wide range of MCDM methods have been developed over recent decades. Each of these methods has its own strengths and weaknesses and differs primarily in several aspects, including the complexity of their algorithms, the specific weighting methodologies they employ, their approach to representing preferences, their capacity to handle uncertain data, and their final data aggregation techniques [78]. The extensive family of MCDM methods is generally divided into two main categories: multi-objective decision-making (MODM) and multi-attribute decision-making (MADM) [79]. Specifically, MODM deals with scenarios where alternatives are not predetermined; instead, it seeks to optimize a set of objective functions subject to constraints, aiming for the most satisfactory and efficient solution where no single objective can be improved without degrading another. In contrast, MADM is applied when a small, predetermined number of alternatives must be evaluated against a predefined set of attributes, which are often challenging to quantify, with the goal of selecting the single best alternative based on comparative analysis and a final compromise. Each of the aforementioned categories contains multiple approaches. For a variety of issues, priority-based, outranking, distance-based, and mixed methods are also used. Every technique has unique properties, and they can be divided into three categories: fuzzy, stochastic, and deterministic, or may be combined. The techniques can be categorized as either single or group decision-making procedures, depending on the number of decision makers.
The practical applications of MCDM span a broad spectrum of fields, demonstrating its versatility and effectiveness in optimizing complex choices, from the evaluation of technology investments [80] to applications in the healthcare [81] and energy [82] industries.

4.1.1. Analytical Hierarchy Process (AHP)

The analytical hierarchy process, conceptualized by Thomas L. Saaty [83,84], stands as a seminal methodology within MCDM, fundamentally designed to structure and solve complex decision problems. At its core, AHP employs hierarchical decomposition, where the overall objective is placed at the top, followed by criteria and sub-criteria, concluding with the available decision alternatives at the lowest level. The methodology mandates a process of pairwise comparison among the elements at any given hierarchy level, assessing their relative importance concerning an element at the immediately superior level. This comparison is quantified using Saaty’s foundational scale of 1–9, where odd values denote discrete levels of preference intensity (1 for equal, 3 for moderately more, 5 for strongly more, 7 for very strongly, and 9 for extremely more importance), and even numbers are utilized for compromise values. The elegance of AHP lies in its use of both ratio scales and verbal assessments, allowing for the effective weighting of both quantifiable and qualitative factors.
Subsequently, the method leverages matrix algebra to compute and aggregate the eigenvectors derived from these comparison matrices. This mathematical procedure culminates in the composite final vector of weight coefficients for all alternatives. The entries within this final vector represent the intrinsic value or relative priority of each alternative relative to the primary goal at the hierarchy’s summit. The decision maker can then utilize this vector, often multiplying it with higher-level weight coefficients in an upward cascading process throughout the hierarchy, to derive the overall weight coefficient for each alternative relative to the ultimate goal. Conventionally, the alternative associated with the maximum final weight coefficient is identified as the optimal choice.
A distinguishing and critical feature of AHP is its capacity to calculate the Inconsistency Index. This metric is generated as the ratio between the decision maker’s derived inconsistency and a corresponding randomly generated index (Random Index). The Inconsistency Index serves as a vital diagnostic tool, providing the decision maker with assurance regarding the internal consistency of their judgments. It is generally accepted that this index should not exceed a threshold of 0.10. While exceeding this limit typically necessitates a re-evaluation of the pairwise comparisons, in certain specific contexts, a non-compliant decision can still be deemed acceptable, although this should be cautiously considered.

4.1.2. Criteria Importance Through Intercriteria Correlation (CRITIC)

The CRITIC [85] method is an objective multi-criteria decision-making method used to determine the weights of individual criteria. Unlike subjective methods (e.g., AHP) that rely on expert judgment, CRITIC derives the weights exclusively from the inherent structure of the data. The key advantage of this method is that it takes into account not only the contrast (variability) of the data within a single criterion (standard deviation), but also the conflict (correlation) between the criteria. A criterion receives a higher weight if it has high variability between alternatives and low correlation with other criteria. This ensures that criteria that provide unique and non-overlapping information receive higher importance in the final decision-making, which makes CRITIC an extremely suitable tool for objectively weighing criteria in complex systems such as the evaluation of XAI methods.

4.1.3. Additive Ratio Assessment (ARAS)

The ARAS method [48] is an MCDM technique that evaluates alternatives based on their relative utility. Its core principle is that an alternative’s overall efficiency is directly proportional to the weighted influence of all criteria used in the assessment. ARAS operates by first normalizing the decision matrix so that all criteria become comparable, eliminating issues arising from differing units or scales. Each normalized value is then multiplied by a criterion weight, ensuring that the relative importance of each metric is fully incorporated into the evaluation.
For every alternative, these weighted normalized values are summed to produce a single complex utility score representing its aggregated performance. The method then introduces an optimal hypothetical alternative, defined as the best possible performance attainable on each criterion. The final utility coefficient for each real alternative is calculated as the ratio between its utility score and that of the optimal solution. These coefficients range between zero and one, with higher values indicating better relative utility and determining the final ranking.

4.1.4. Borda Count

The Borda count method [53] is a classical approach for aggregating rankings of alternatives in multi-criteria decision-making and preference-based systems. Originally developed by Jean-Charles de Borda in the 18th century, the method provides a systematic means of integrating ordinal rankings across multiple criteria or decision-makers into a single, comprehensive ranking.
The fundamental principle of the Borda method involves assigning numerical scores to alternatives based on their relative position within each criterion. Specifically, in a set of n alternatives, the highest-ranked alternative receives n 1 points, the second-ranked n 2 points, and so forth, down to the lowest-ranked alternative receiving zero points. These scores are subsequently summed across all criteria to produce a total Borda count for each alternative. The alternative with the highest cumulative score is then identified as the most preferred.
A notable advantage of the Borda method is its capacity to capture consensus among criteria or evaluators. By considering the complete ordering rather than only top-ranked preferences, it rewards alternatives that consistently perform well across multiple dimensions, providing a more balanced and representative assessment.

4.1.5. Combinative Distance-Based Assessment (CODAS)

The CODAS method [50] is a modern MCDM approach designed to provide a more robust evaluation of alternatives by combining two distance measures. Its core idea is straightforward: an alternative is considered better if it lies farther from the negative ideal solution (NIS), which represents the worst possible performance on each criterion.
The method begins by identifying the NIS and then measuring how far each alternative is from this point. CODAS uses two complementary distance metrics. The primary one is the Euclidean distance, which captures the overall deviation from the NIS across all criteria. It gives a general sense of how well an alternative performs relative to the worst-case scenario. The second metric is the Taxicab (or Manhattan) distance, which sums absolute deviations and provides more sensitivity to individual criterion differences. This second measure becomes important when alternatives are very close to each other in terms of Euclidean distance.
A threshold value determines whether the Taxicab distance should be used. If the Euclidean distances of two alternatives differ enough, the ranking is based directly on that measure. If not, the Taxicab distance serves as a tie-breaker, ensuring fine-grained discrimination.

4.1.6. Evaluation Based on Distance from Average Solution (EDAS)

The EDAS method [44] evaluates decision alternatives by comparing them to the average performance across all criteria, rather than to ideal or worst-case reference points. This makes it different from methods such as TOPSIS or VIKOR, which rely on ideal and anti-ideal solutions.
The process begins by calculating the average solution (AS), which is simply the mean value of all alternatives for each criterion. Each alternative is then assessed in terms of how far it lies above or below this average. Two distances are computed for every criterion: the positive distance (PD), which reflects how much an alternative exceeds the average in benefit criteria (or falls below it in cost criteria), and the negative distance (ND), which shows the degree to which it performs worse than the average in benefit criteria (or better in cost criteria).
These distances are then weighted and summed, producing aggregated PD and ND values for each alternative. The final appraisal score is obtained by combining these weighted distances, and alternatives are ranked according to this score, with higher values indicating better overall performance.

4.1.7. Multi-Attributive Border Approximation Area Comparison (MABAC)

The MABAC method [51] is an intuitive MCDM approach that evaluates alternatives based on how far they lie from a defined reference boundary. Its central idea is the creation of a border approximation area (BAA), which acts as a threshold separating stronger from weaker performance on each criterion.
The BAA is usually derived from the dataset, often calculated as the geometric mean between the best and worst values for each criterion. This produces a reference point that represents a balanced boundary. With this boundary in place, each criterion is conceptually divided into three zones: the boundary area itself, an upper area where performance is better than the BAA, and a lower area where performance is worse.
Once the BAA is defined, each alternative is evaluated by calculating its distance from this boundary. Distances are positive if an alternative lies in the upper (preferred) zone, negative if it falls in the lower zone, and close to zero if it sits near the boundary. These distances are then weighted according to the importance of each criterion. Summing the weighted distances across all criteria produces a final score for each alternative, with higher scores indicating more consistent performance above the boundary.

4.1.8. Measurement of Alternatives and Ranking According to Compromise Solution (MARCOS)

The MARCOS method [86] is a recent and powerful MCDM approach designed to provide a stable and reliable ranking by evaluating alternatives against both ideal and anti-ideal reference points. This dual-reference framework ensures a comprehensive assessment of each option’s relative performance.
MARCOS begins by adding the ideal solution (IS) and anti-ideal solution (AIS) to the decision matrix. Each alternative is then evaluated relative to these references, producing two utility functions: one measuring proximity to the ideal outcome, and the other measuring distance from the worst outcome. These two measures are combined into a single utility degree, which captures both how close an alternative is to the best solution and how far it is from the worst.
The utility degree is then incorporated into a weighting function that aggregates the alternative’s performance across all criteria. This ensures that the final score reflects both absolute performance and strategic positioning within the decision space. Alternatives are ranked in descending order of their final score, with the highest value indicating the preferred choice.

4.1.9. Preference Ranking Organization Method for Enrichment of Evaluations II (PROMETHEE II)

The PROMETHEE II method [47] is a well-established outranking technique within MCDM that differs fundamentally from utility-based or distance-based models. Instead of computing a single aggregated value for each alternative, it builds the ranking through structured pairwise comparisons guided by preference functions.
For each criterion, the decision-maker selects a function—chosen from several standard types—that translates the difference between two alternatives into a degree of preference ranging from zero (no preference) to one (strong preference). This allows PROMETHEE to incorporate meaningful thresholds, such as when small differences should be ignored or when larger gaps reflect clear superiority.
Using these functions, the method calculates two flows for every alternative. The positive flow measures how much an alternative is preferred over all others across the criteria, while the negative flow measures how much it is dominated by the rest. PROMETHEE II then combines these into a net outranking flow, obtained by subtracting the negative flow from the positive one. This net flow produces a complete ranking: the higher the net flow, the stronger the alternative overall.

4.1.10. The Technique for Order Preference by Similarity to Ideal Solutions (TOPSIS)

The technique for order preference by similarity to ideal solution is a powerful and intuitive MCDM method [45]. The fundamental idea behind TOPSIS is simple: the chosen optimal alternative should be geographically closest to the ideal solution (the “best case”) and simultaneously farthest from the negative-ideal solution (the “worst case”) in a multi-dimensional criterion space.
The process begins by formulating a decision matrix containing M alternatives evaluated against N criteria. This matrix is then subjected to normalization and the application of criteria weights to create a weighted decision matrix. Crucially, TOPSIS identifies two theoretical benchmarks: the ideal solution, which represents the maximum performance across all benefit criteria and minimum performance across all cost criteria, and the negative-ideal solution, which represents the inverse. The method then uses Euclidean distance to calculate the separation measure of each real alternative from both the ideal solution and the negative-ideal solution.
Finally, the relative closeness to the Ideal Solution is calculated for every alternative. The alternative achieving the highest closeness score is designated as the best option. TOPSIS is highly valued for its straightforward, distance-based ranking mechanism, making it easily applicable even in common spreadsheet software. The advantage of this method is robustness and reliability [87].

4.1.11. Višekriterijumska Optimizacija I Kompromisno Rešenje (VIKOR)

The VIKOR method [88] is a well-known MCDM technique designed to identify a compromise solution—one that comes closest to the ideal performance while avoiding excessive deviation on any single criterion. Rather than relying solely on distance to the ideal or anti-ideal solution, VIKOR explicitly balances two competing perspectives: overall group satisfaction and individual regret.
At the core of the method are two measures. The first is the group utility measure (S), which reflects the overall performance of an alternative across all criteria. Lower values indicate better collective performance. The second is the individual regret measure (R), which captures the worst deviation of an alternative from the ideal on any single criterion. A smaller R value means the alternative avoids large individual shortcomings.
These two measures are combined into the Compromise Index (Q), which determines the final ranking. A strategy weight governs how much emphasis is placed on group utility versus individual regret: values near 1 favor the majority’s overall benefit, while values near 0 prioritize minimizing the worst-case outcome. A common default is 0.5, representing a balanced compromise. Alternatives are ranked by increasing Q, with the lowest score indicating the recommended compromise solution.

4.1.12. Weighted Aggregated Sum Product Assessment (WASPAS)

The WASPAS method [89] is a modern MCDM technique that combines the strengths of two classical approaches: the Weighted sum model (WSM) and the weighted product model (WPM). This hybrid structure improves ranking accuracy and reliability compared to using either method alone.
WASPAS evaluates alternatives using two complementary aggregation strategies. The WSM component calculates the additive utility by summing weighted normalized scores across all criteria, offering simplicity and a compensatory view of performance. The WPM component calculates the multiplicative utility by taking the weighted product of normalized scores, which emphasizes sensitivity to poor performance in any single criterion. By combining these approaches, WASPAS balances overall utility with attention to individual weaknesses.
The method generates a single combined score, the Joint Generalized Criterion (J), which merges the WSM and WPM results using a generalization parameter. This parameter controls the relative influence of the additive versus multiplicative components: values near 1 favor the WSM, emphasizing overall utility, while values near 0 favor the WPM, emphasizing sensitivity to individual criteria. Alternatives are then ranked by descending J values, with higher scores representing better options.

4.1.13. Weighted Sum Model (WSM)

The weighted sum model [90] is one of the oldest and most widely used MCDM techniques. Its main advantage lies in its simplicity, as it evaluates alternatives by directly aggregating weighted performance scores across all criteria.
WSM is based on the principle that an alternative’s overall performance can be represented as the sum of its weighted scores. It assumes that all criteria are fully compensatory, meaning that a low score on one criterion can be offset by a high score on another.
The method involves two main steps. First, the raw performance scores are normalized to ensure comparability across criteria with different units and scales. Second, each normalized score is multiplied by its criterion weight, and these weighted values are summed to produce a single utility score for each alternative. Alternatives are then ranked in descending order of this total score, with the highest-scoring option considered the best choice.

5. Proposed Tool for Selection of Explainable AI Methods

The following section will describe the proposed solution for selecting explainability methods in a specific case. The tool and all necessary materials are available on the website https://github.com/m-matejova/XAI_selection (accessed on 17 November 2025). The current XAI field faces a diversity problem, and it is extremely challenging for developers and domain experts to objectively and systematically determine which method provides not only technically accurate, but also contextually relevant and credible explanations that meet the specific requirements of a given use case.
Creating a tool that integrates quantitative data from benchmarks and user studies with user preferences and aggregates them through a robust multi-criteria decision-making method directly addresses this challenge. The tool thus serves as a bridge between the technical complexity of XAI and the practical requirements of the end user, thereby increasing trust, transparency, and efficiency in the deployment of AI systems in critical applications. The tool has two parts that can be used separately. The first part is focused on filtering methods. The second phase is focused on selecting the best method based on multi-criteria decision-making.

5.1. Filtering Methods

The taxonomy of explainable artificial intelligence methods provides a systematic framework that helps to navigate the wide range of existing approaches. By combining criteria—according to the scope of explanation, approach to the model, data type, task type, and output form—it is possible to determine which XAI method or methods are most suitable for a given problem.
As a basic list, we used the work of the authors [66], who compiled a list of 249 post-hoc XAI methods. Of these, we kept only 113 methods (see Table 2) that have source code for reproducing the explanations. This list was created based on a survey of relevant scientific articles in the field of XAI from 2015 to 2024. The first column is the name of the method, followed by the year of publication and properties according to the basic taxonomy. If the method is model-specific, the model column specifies the type of model. Another important column is the type of problem being solved, whether it is classification or regression. The methods are ranked according to the original article, where they were ranked by popularity, calculated as a percentage of the reviewed surveys in which each method is mentioned. We supplemented this list with a column describing the output format based on further studies [26,40].
As we can see in Figure 2, first of all, the user enters information about the use case for which they are looking for a suitable XAI method/methods. It is known objective information that results from the basic taxonomy of methods. The tool also includes help texts that describe the filter selection options. These texts briefly explain the differences and describe the selection options with specific examples. (For example, “Select Global if you are interested in understanding the overall system behavior—how the model makes decisions across all possible data points.”) By adding help texts, we ensure that users quickly understand the impact and meaning of each choice. Based on this information, methods from the list of 113 methods are filtered. In this context, a user can be a developer, a domain expert, or even a newcomer to the XAI field. The output is a list of matching methods that the user can view and download in CSV format. The filtered set of methods shall constitute the cornerstone for the subsequent MCDM phase.
Ideally, we aim to secure results for all methods included in this refined list. These results can be derived from existing benchmark studies (Table 1), internally conducted experimental analyses, or dedicated user studies. These data must then be uploaded to the MCDM section in a standardized format, specifically as CSV or XLSX files, to construct the decision matrix (Figure 2). The structure of this input file is crucial:
  • The first column must strictly contain the names of the XAI methods (serving as the alternatives).
  • The subsequent columns must specify the names of the evaluation metrics (serving as the criteria) and the corresponding performance results for each method against these metrics.
Upon successful generation of the decision matrix, the process is prepared to proceed to the second phase, involving the application of MCDM methods for the aggregation and selection of the optimal XAI technique.

5.2. Choosing a Method Using Multi-Criteria Decision-Making

Because it can be difficult or impossible to obtain ground-truth labels and explanations, evaluating an XAI approach is not as easy as analysing the performance of an ML model. The fact that an explanation’s quality is somewhat subjective, depending on how each user interprets it, makes it much more difficult. No XAI method is universally the best.
The use of multi-criteria decision-making methods for selecting a suitable explainable artificial intelligence method represents a significant advance towards the systematization of the decision-making process, which is often complex, subjective and dependent on multiple, interacting factors. The selection of the optimal XAI method rarely depends on just one property, such as accuracy or robustness, but requires taking into account multiple dimensions.
MCDM methods allow for the quantification and weighting of individual criteria according to their importance, thus providing a transparent and reproducible framework for comparing alternative XAI methods. The main advantage of their use is that they can integrate both objective (e.g., fidelity, stability, speed) and subjective metrics (e.g., interpretability, trustworthiness, usability) into a single decision-making model. This overcomes the problem where individual methods excel only in partial areas, but there is no unified way of evaluating them overall.
Another advantage of MCDM approaches is their flexibility and adaptability. Researchers can adapt the weights of the criteria to a specific context—for example, in medicine, the priority may be the understandability and credibility of the explanation, while in industrial applications the speed of calculation dominates. In addition, MCDM techniques support sensitivity analysis, which allows examining how changing the weights will affect the choice of method, which significantly increases the transparency of the decision.
Currently, however, there is no perfect metric or benchmark that compares all existing methods. It is therefore essential that the user uploads a list of methods together with their results from the selected metrics. Next, the user determines the weights of individual metrics or criteria (see Figure 2). There are three options to choose from:
  • Direct Rating—The first option is to enter the criteria weights directly for each of them as a number on a scale from 1 (unimportant) to 10 (very important). This option is suitable when the user has a clear idea of the importance of the metrics.
  • Pairwise comparison—The second option is a pairwise comparison of metrics (based on AHP). This method is not suitable for a higher number of metrics, because in that case, it is necessary to make many comparisons, which can be cognitively demanding and confusing. The recommended number is a maximum of 9 [83], ideally 3 to 7.
  • CRITIC—The third option is the use of the CRITIC method, which serves to objectively determine weights based on variability and relationships between criteria. This method is also suitable for a larger number of metrics.
The metric weights are used as input to the Decision matrix (see Figure 2), which is the input for calculating preferences using MCDM methods: ARAS, CODAS, EDAS, MABAC, MARCOS, PROMETHEE II, TOPSIS, VIKOR, WASPAS, and WSM. The individual preferences/rankings are aggregated using the Borda count method. This process generates a final, comprehensive list of XAI methods, ordered from best to worst based on their cumulative Borda count.

5.3. Sensitivity Analysis

To assess the impact of criterion weights on the final ranking of alternatives, we executed a comprehensive sensitivity analysis. The methodology employed for this procedure draws inspiration from the approach detailed in the study [190]. Sensitivity analysis is also part of the proposed tool.
We generated a total of twenty weight variations for each MCDM method utilized. The weights were systematically varied by adding a value, δ , which ranged between −0.40 and +0.40. The new weight of the criterion W i is computed as defined below (Equation (1)). Here, W i refers to the weight used originally in MCDM methods, δ is as defined above, and n is the number of decision criteria (metrics).
W i = W i + δ ,   f o r   1 i n
Following this adjustment, the modified weights must be normalized (using Equation (2)) to ensure that the sum of the new weights collectively equals one.
W i = W i W i ,   f o r   1   i n
Across the twenty experiments performed for each MCDM method (one for each value δ ), we recalculated the preferences and derived the new ranking of alternatives. This extensive re-calculation allowed us to thoroughly examine how alterations in the criteria’s weights influence the final ranking.

5.4. Use Cases

The following section describes in detail a practical workflow that has been designed for users of our tool to systematically, objectively, and transparently select the most appropriate explainable artificial intelligence method.
1.
Define Context, Metrics, and XAI Methods
The user must first define the problem space, as this determines the necessary trade-offs.
  • Identify XAI candidates: Filter the set of XAI methods to be evaluated (e.g., the 3–5 methods relevant to the user’s specific ML model).
  • Define evaluation metrics: Determine the key performance indicators that the XAI method must satisfy. These should include standard measures and domain-specific needs.
  • Determine metric type: Classify each metric as either a Benefit (higher value is better) or a Cost (lower value is better).
  • Input performance data: The user must run each candidate XAI method and input the performance data into the tool.
2.
Establish Weights
Metric weights can be obtained in several ways. The user must choose one of the options: direct rating, pairwise comparison (AHP), or CRITIC.
3.
Obtaining the Borda count based on MCDM methods results
The tool automatically processes the decision matrix using a suite of diverse MCDM techniques to ensure the quality of the solution.
  • Initial ranking: Review the initial preference scores and the resulting ranking from each method.
  • Borda count: The Borda count provides a balanced overall rank for each XAI method.
4.
Validate Robustness (Sensitivity Analysis)
This is the most critical step for ensuring the chosen XAI method is stable and defensible.
  • Sensitivity tests: Observe how the resulting order of methods changes when changing the weights by a factor of δ .
  • Analyze rank stability: Review the table that shows how often each XAI method retained Rank 1 across all weight perturbations and MCDM methods.
  • Review rank change plots: Examine the visual plots to identify the following:
    Stable Leaders: Lines that stay consistently high.
    Crossover Points: Areas where lines intersect, indicating high instability and a change in the rank when the weights shift marginally.
5.
Final Selection and Documentation
The user makes the final choice based on both initial performance and validated stability.
  • Document Decision: Use the generated MCDM ranking tables and the sensitivity analysis plots as rigorous evidence to document the decision, thereby providing a clear audit trail for compliance and quality control.
To demonstrate the practical use of the proposed solution, we will describe its application in three different use cases. The presented cases demonstrate the flexibility of the proposed approach, which confirms its broad applicability across various domains and tasks. Crucially, the method effectively integrates both objective and subjective metrics and an approach to their weighting.

5.4.1. Choosing the XAI Method in the Field of Hate Speech Detection

As a practical demonstration of the use of the proposed framework for selecting a suitable XAI method, we used the results of an experiment we conducted in the field of hate speech content detection on social networks. The goal was to compare the quality of explanations generated by three popular XAI methods—LIME, SHAP, and Grad-CAM—applied to the ResNet-50 model trained for the image modality on the Hateful Memes dataset [191]. The model classified images containing potentially hateful content, while the explanations of individual methods were subsequently assessed in terms of their comprehensibility and practical value for the user.
The subjective metrics understandability, usefulness, trustworthiness, informativeness, and satisfaction were used to assess the quality of the explanations [74]. The evaluation was carried out using questionnaires and a Likert scale (1–7), where 1 represented the lowest and 7 the highest level of agreement with the explanation statement.
The experiment involved 70 respondents aged 18–22 with varying levels of knowledge in the field of artificial intelligence and visual interpretation. Participants were shown the outputs of all three XAI methods in sequence, and after each viewing, they filled out a short questionnaire evaluating the above metrics.
The questionnaire was structured into distinct sections, with each dedicated to a specific explainability method. To ensure participant comprehension, every section began with accompanying textual instruction and an example explanation. This setup clarified the method’s purpose and established the context for its application. Importantly, all sections utilized the identical set of evaluative statements (metrics noted parenthetically):
  • The explanation contains information that is essential for me to understand the model’s decision. (Understandability)
  • The explanation is useful to me for making better decisions or performing an action. (Usefulness)
  • Based on the explanation, I have more confidence in the model’s decision. (Trustworthiness)
  • The explanation provides sufficient information to explain how the system makes decisions. (Informativeness)
  • I have a satisfied attitude towards the explanation of the model. (Satisfaction)
Users answered to what extent they agreed with the given statements on a scale from 1 to 7. From the responses, we calculated metrics as the average score of the responses.
Imagine a team of researchers working on a project that aims to automatically detect hate speech on a social network—images that spread insults or incite violence. The model they use is a modern ResNet-50 neural network, capable of accurately recognizing visual elements, but its decision-making remains a “black box” for the average user. When the system labels an image as hateful, the social network’s moderators ask themselves: Why this one? What did the model see in it? How do we explain it to users?
To get answers, the team used our tool to filter which methods were suitable for use in this task (model-agnostic, local, image input data, classification, visual output) and selected three popular XAI methods—LIME, SHAP, and Grad-CAM. Each offered a different perspective on how the model thinks. However, they can only deploy one of them.
To verify which method is most understandable and trustworthy for people, the researchers used the research results shown in Table 3.
The ratings thus obtained can then be processed in the proposed decision tool, which allows determining the most appropriate method. After loading the survey results, it is necessary to determine the weights of the individual criteria. With 5 metrics, the team chose the pairwise comparison method, which allows them to compare individual pairs of metrics against each other and thus obtain their weights. The team of researchers created the following Saaty’s matrix by comparing the importance of the metrics (Table 4):
Trustworthiness (weight calculated based on Saaty matrix 0.476) is of the utmost importance in this area, as users need to trust that the explanation is not misleading. Model failure or bias can have serious social or legal consequences. If the system can transparently and accurately explain why certain content has been removed, it increases trust in the platform. Understandability (0.234) and usefulness (0.151) are of moderate importance—people want to understand the output and be able to use it. Informativeness (0.053) and satisfaction (0.088) are complementary but less crucial.
After applying all MCDM methods, the research team obtained preferential results (Table 5) and the Borda count. The LIME method obtained a score of 20, the SHAP method 10, and the Grad-CAM method 0. The team therefore decided to implement the LIME method. This use case demonstrates the possible combination of subjective metrics and pairwise weighting.
The ranking for XAI methods across all MCDM methods is shown in Figure 3. The entire ranking is constant.

5.4.2. Choosing the XAI Method in the Field of Medicine

Imagine a young doctor who works in a diabetes clinic at a university hospital. Every day, she sees dozens of patients who are suspected of having diabetes. In recent years, the hospital has begun experimenting with artificial intelligence systems that can predict the risk of developing the disease based on various data. A model based on a Random Forest showed very promising results, the ability to identify hidden patterns in data that the human eye cannot see.
Nevertheless, the doctor had doubts. If the AI recommends that a patient is at high risk of diabetes, what does it base this conclusion on? What factors were decisive? Was it glucose levels, age, or perhaps BMI? In order for the doctor to trust the decision, it was necessary to understand why the model produced a particular prediction.
In her research, she came across a study [66] that compares 3 methods to explain a Random Forest in a similar problem, but she could not determine which of them would be the best choice for her. Therefore, she will use the proposed framework, where she will load the results (Table 6) from the study.
In the next step, it is necessary to determine the weights for each of the metrics. Due to a lack of experience and a large number of metrics, the doctor decides to use an objective calculation of weights using the CRITIC metric. Since this method calculates variability and some criteria had zero variability (standard deviation 0), the result is their weight of 0 (e.g., selectivity, interactivity, uncertainty). This means that these criteria do not contribute to distinguishing between alternatives in the given matrix. The calculated weights are given in Table 7.
Using the proposed tool, the physician obtained the following Borda count for the methods: LIME 5, Kernel SHAP 6, and Tree SHAP 19. The preference values are given in Table 8. Figure 4 gives us a look at the ranking of XAI methods in the MCDM results, where we can see how the ranking varied. Three SHAP maintained the first place among the nine methods. The doctor decided to implement Tree SHAP to help explain the prognosis to patients. This use case demonstrates the possible combination of objective metrics and weighting.

5.4.3. Choosing the XAI Method in Finance

The authors of the OpenHEXAI benchmark [73] focused their research on creating a framework for human-centered evaluation of XAI methods. They used LIME, SHAP, SmoothGrad, and Integrated Gradients methods that they implemented on a neural network model. When testing the proposed solution, they examined the impact of XAI methods on improving results. The German Credit Dataset [192], a widely used dataset in the fields of machine learning, data science, and financial modeling, was used. Its main goal is to classify loan applicants based on various attributes. Evaluation included both objective (e.g., accuracy) and subjective metrics (answers to various questions on a scale from 1 to 5). They examined 23 metrics in total. With such a large number of metrics, it may not be easy to determine which method is the best. We therefore suggest using our tool, which allows you to find a compromise between the four tested alternatives. After loading the data, it is necessary to determine the weights of the metrics; we recommend using the objective CRITIC metric, which can easily handle a larger number of metrics. The weights obtained in this way are shown in Table 9. Using all MCDM methods, we obtained the following Borda count: LIME: 20, SHAP: 29, SmoothGrad: 9, Integrated Gradients: 2.
Table 10 displays the computed preference scores derived from all MCDM methods. The best choice for implementation in this case is SHAP. This use case demonstrates the possible combination of objective and subjective metrics and objective weighting.
The proposed tool also allows for a simple display of positions in rankings according to MCDM methods for each of the XAI methods (Figure 5). This allows for analysis of changes in positions according to individual MCDM methods. Another useful display is the correlation matrix.
The tool also allows for deeper analysis of the results and the examination of correlations between the results of MCDM methods. The correlation matrix for this third use case is shown in Figure 6. The correlation matrix displays the degree of concordance between the rankings generated by the ten different MCDM methods. The values may range from 0.00 (no agreement) to 1.00 (perfect agreement). Most MCDM methods show a very high degree of correlation (0.80 to 1.00), indicating that for this specific dataset, they largely agree on the final ranking of the alternatives.
In the next section, we will describe the sensitivity analysis for the third use case. According to the analysis design, preferences and rankings were calculated for the data from this use case for various weights with δ increment for all MCDM methods, a total of 200 calculations. The table of weight changes according to the δ value is in Appendix A due to its large scope. The analysis is focused on investigating, in particular, the selection of the best method. Table 11 presents the percentage of methods in first place according to the ranking of MCDM methods. At first glance, it is obvious that the LIME method dominated over all MCDM methods. However, this is in contrast to the results for the original weights, where the SHAP method dominated.
The least robust in this case is the VIKOR method, which replaced 3 methods in first place when changing the weights. A more detailed overview of the ranking change based on the weight change is shown for this method in Figure 7.
The pie chart (Figure 8) shows that among all calculations performed, the preference in this use case was dominated by the LIME method (81.5%); SHAP was in first place only in 15% of cases.
Although SHAP offered a marginally better preference score in the initial setup, the results from the extensive sensitivity analysis clearly establish LIME as the more robust and reliable choice. This suggests that LIME’s performance is stable even when the importance of the XAI metrics is adjusted, making it a more dependable method for application.
Sensitivity analysis for the first and second use cases demonstrated the robustness of the selection of the best XAI method. In the first case, the LIME method dominated in 98.5% of all tests, and in the second case, the Tree SHAP method dominated in 94.3%.

6. Discussion

In this section, we discuss the proposed approach to selecting an appropriate XAI method using MCDM. The complexity and diversity of existing XAI techniques require a systematic framework that would allow users to navigate and select the most appropriate solution based on their specific needs and constraints.
To simplify the selection process and increase its usability, we have created a useful list of XAI methods, which is structured and allows users to filter methods according to key features, such as local vs. global explanation, model-agnostic vs. model-specific, or type of input data (e.g., image, text). To support the evaluation of XAI methods, we have compiled an overview of relevant benchmarks and software libraries that are focused on a quantitative evaluation of XAI methods. The challenge for the future is to add other important aspects that will be more focused on specific domains and user skills, and experiences.
The use of multi-criteria decision-making methods in the context of Explainable Artificial Intelligence is extremely useful and provides solutions to many problems. Compared to previous studies (e.g., AutoXAI or eXplego), we use MCDM to select the right method given the metrics and specifics of the application. When choosing an XAI method, several, often conflicting, requirements need to be considered. For example, a method that is extremely faithful to the model prediction may also be computationally demanding. MCDM allows these different criteria to be weighed, and an optimal balance to be found based on the user’s needs. The proposed tool allows the user to define weights for each metric (e.g., someone will give more weight to Robustness, while another will give more weight to Speed). This makes it a very flexible tool adaptable to specific domain requirements, going beyond the framework of fixed decision trees such as eXplego. Furthermore, the list of XAI methods is easily expandable, which can be more difficult with a tree structure.
While traditional benchmarks often focus only on technical metrics, MCDM provides a structured framework for incorporating the subjective preferences of domain experts. Methods such as AHP or CRITIC allow experts to assign importance (weights) to different metrics, ensuring that the selected XAI method is relevant to their specific work context and requirements. Compared to AutoXAI, our solution does not focus on hyperparameter tuning. Its advantage is that it does not restrict users to predefined metrics; instead, they can select metrics according to their specific use, whereas AutoXAI is limited to the metrics and hyperparameters it supports.
MCDM transforms the complex selection of an XAI method from a manual, intuitive process into a transparent, mathematically justified, and reproducible decision. The proposed solution aggregates the results of ten MCDM methods using the Borda count, thus ensuring the robustness of the result.
MCDM is crucial because the problem of selecting the best XAI method is not a single-criterion problem, but a complex multi-criteria problem requiring a weighted synthesis of objective performance and subjective usability.
The proposed approach includes the possibility of using a wide range of criteria that cover the desired properties of explanations (e.g., fidelity, robustness) as well as technical aspects (e.g., computational complexity, speed). The key is the implementation of a flexible mechanism for calculating the weights of the criteria. Users can enter the weights directly (direct rating method), which is simple and intuitive if they have a clear idea of the importance of the criteria. Alternatively, a pairwise comparison method (based on AHP) is available that takes into account subjective preferences. To ensure objectivity and minimize subjective bias, we integrated the CRITIC method, which determines weights solely based on the internal structure of the data and the contrasting strength of individual criteria. The ability to combine subjective and objective weights represents a significant benefit that ensures robust selection of the most appropriate XAI method.
The presented use cases confirm the broad domain and task applicability of the proposed approach. At the same time, they illustrate the ability of this approach to flexibly obtain weight coefficients using objective and subjective methods and to take into account a wide range of metrics.
Legislative changes pose a challenge for the near future use of artificial intelligence systems. The European Union’s AI Act represents a pioneering regulatory framework designed to ensure that artificial intelligence systems deployed within the EU adhere to safety, transparency, and fundamental rights standards. It utilizes a risk-based approach, classifying AI applications into four tiers—minimal, limited, high, and unacceptable risk—with compliance obligations scaled accordingly. This framework builds upon existing European data protection legislation, notably the General Data Protection Regulation (GDPR), which already mandates strict requirements for the processing of personal data, including the principles of data minimization and the right to explanation for automated decisions.
The EU AI Act inherently relies on explainable AI (XAI) to ensure compliance for systems categorized as High-Risk. The legislation mandates specific outcomes—namely, transparency, human oversight, and verifiable data governance—that necessitate XAI tools. Without effective interpretability techniques, operators cannot adequately understand complex model behavior, audit for bias in training data, or fulfill the requirement for meaningful human oversight. Thus, XAI serves as the crucial technical mechanism for meeting the legal and ethical obligations imposed by the Act.
Since the choice of XAI method involves complex trade-offs, our MCDM tool is crucial for risk mitigation. By systematically evaluating XAI candidates against relevant metrics and employing robust methods to find the optimal compromise, our tool provides a defensible, objective, and stable ranking. Several XAI methods also offer Counterfactual Explanations, which can help with objections.

Author Contributions

Conceptualization, M.M. and J.P.; methodology, M.M.; software, M.M.; validation, M.M.; formal analysis, M.M.; investigation, M.M.; resources, M.M.; data curation, M.M.; writing—original draft preparation, M.M.; writing—review and editing, J.P.; visualization, M.M.; supervision, J.P.; project administration, J.P.; funding acquisition, J.P. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Scientific Grant Agency of the Ministry of Education, Research, Development and Youth of the Slovak Republic and the Slovak Academy of Sciences under grant number 1/0259/24, and by the Slovak Research and Development Agency under contracts No. APVV-22-0414 and APVV-24-0454.

Data Availability Statement

In the use case described in this paper, the German Credit Dataset has been used. It is available on the website here: https://archive.ics.uci.edu/dataset/144/statlog+german+credit+data (accessed on 1 September 2025).

Acknowledgments

During the preparation of this manuscript, the authors used ChatGPT 5 and Gemini 2.5 Flash for the purposes of language translation support, linguistic refinement, and assisting with paraphrasing. The authors have reviewed and edited the output and take full responsibility for the content of this publication.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
AHPAnalytic Hierarchy Process
AIArtificial Intelligence
AISAnti-Ideal Solution
ARASAdditive Ratio Assessment
ASAverage Solution
BAABorder Approximation Area
CDSSClinical Decision Support System
CODASCombinative Distance-based Assessment
CRITICCriteria Importance Through Intercriteria Correlation
EDASEvaluation Based on Distance from Average Solution
ISIdeal Solution
JSONJavaScript Object Notation
MABACMulti-Attributive Border Approximation Area Comparison
MADMMulti-Attribute Decision-Making
MARCOSMeasurement of Alternatives and Ranking according to Compromise Solution
MCDAMulti-Criteria Decision Analysis
MCDMMulti-Criteria Decision-Making
MODMMulti-Objective Decision-Making
NDNegative Distance
NISNegative Ideal Solution
PDPositive Distance
PROMETHEE IIPreference Ranking Organization Method for Enrichment of Evaluations II
TOPSISTechnique for Order of Preference by Similarity to Ideal Solution
VIKORVIšeKriterijumska Optimizacija I Kompromisno Rešenje
WASPASWeighted Aggregated Sum Product Assessment
WSMWeighted Sum Model
XAIExplainable Artificial Intelligence

Appendix A

This section contains a table from the sensitivity analysis for the last use case (Section 5.4.3). The table containing the metric weights and the resulting rankings according to MCDM methods based on the weight change δ has been divided into 4 tables for a clearer display. Table A1 contains the changes in weights W’1 to W’12 according to the δ value. Table A2 contains the weights W’13 to W’23. The resulting ranking according to the ten MCDM methods is in Table A3 and Table A4.
Table A1. Metric weights W’1 to W’12 for sensitivity analysis of the third use case.
Table A1. Metric weights W’1 to W’12 for sensitivity analysis of the third use case.
Weight Change (δ)W’1W’2W’3W’4W’5W’6W’7W’8W’9W’10W’11W’12
−0.40.0360.03680.04340.04420.04210.0450.04140.04510.04530.04510.04530.0437
−0.3580.0350.03590.04340.04430.04190.04530.04120.04530.04550.04530.04550.0437
−0.3160.03370.03480.04340.04440.04160.04550.04080.04550.04580.04560.04580.0437
−0.2740.03190.03320.04340.04450.04130.04590.04030.04590.04630.0460.04620.0438
−0.2320.02930.03090.04330.04480.04080.04650.03960.04650.04690.04660.04680.0438
−0.1890.02520.02720.04330.04520.040.04730.03850.04730.04790.04740.04780.0439
−0.1470.01770.02070.04320.04590.03860.04890.03650.04890.04960.04910.04960.0441
−0.1050.00020.00510.04310.04750.03530.05250.03170.05260.05380.05290.05370.0446
−0.0630.06440.05360.02940.0390.01230.05010.00450.05020.05290.05080.05270.0327
−0.0210.16280.14920.04460.03250.06610.01850.07590.01840.0150.01760.01530.0405
0.0210.08490.08020.04390.03970.05130.03480.05480.03480.03360.03450.03370.0424
0.0630.06860.06570.04370.04120.04820.03820.05030.03820.03750.0380.03760.0428
0.1050.06150.05940.04370.04180.04690.03970.04840.03970.03920.03960.03920.043
0.1470.05750.05590.04360.04220.04610.04050.04730.04050.04010.04040.04020.0431
0.1890.0550.05370.04360.04240.04570.04110.04660.04110.04070.0410.04080.0432
0.2320.05320.05210.04360.04260.04530.04140.04610.04140.04120.04140.04120.0432
0.2740.05190.0510.04360.04270.04510.04170.04580.04170.04150.04170.04150.0433
0.3160.05090.05010.04350.04280.04490.04190.04550.04190.04170.04190.04170.0433
0.3580.05010.04940.04350.04290.04470.04210.04530.04210.04190.0420.04190.0433
0.40.04950.04880.04350.04290.04460.04220.04510.04220.0420.04220.04210.0433
Table A2. Metric weights W’13 to W’23 for sensitivity analysis of the third use case.
Table A2. Metric weights W’13 to W’23 for sensitivity analysis of the third use case.
Weight Change (δ)W’13W’14W’15W’16W’17W’18W’19W’20W’21W’22W’23
−0.40.04550.04510.04570.04420.04520.04470.04550.04470.03840.04270.045
−0.3580.04580.04530.04590.04430.04550.04480.04580.04490.03770.04260.0452
−0.3160.04620.04550.04630.04440.04580.0450.04610.04510.03690.04250.0455
−0.2740.04670.04590.04680.04460.04620.04530.04660.04540.03560.04230.0459
−0.2320.04740.04650.04760.04480.04680.04570.04730.04580.03390.04210.0464
−0.1890.04850.04730.04880.04520.04780.04640.04840.04650.03110.04170.0472
−0.1470.05060.04890.05090.04590.04950.04750.05040.04780.02610.04090.0487
−0.1050.05540.05260.0560.04750.05370.05030.05520.05070.01430.03920.0523
−0.0630.05640.05020.05770.0390.05260.04530.05590.0460.03360.0210.0496
−0.0210.01070.01840.00890.03250.01540.02460.01120.02370.12390.05520.0191
0.0210.03210.03480.03150.03970.03370.03690.03230.03660.07140.04760.035
0.0630.03660.03820.03620.04120.03760.03950.03670.03930.06040.0460.0384
0.1050.03850.03970.03830.04180.03920.04060.03860.04050.05560.04530.0398
0.1470.03960.04050.03940.04220.04020.04130.03970.04110.05290.04490.0406
0.1890.04030.04110.04020.04240.04080.04170.04040.04160.05120.04460.0411
0.2320.04080.04140.04070.04260.04120.04190.04080.04190.050.04440.0415
0.2740.04120.04170.0410.04270.04150.04210.04120.04210.04920.04430.0418
0.3160.04140.04190.04130.04280.04170.04230.04150.04220.04850.04420.042
0.3580.04160.04210.04150.04290.04190.04240.04170.04240.0480.04410.0421
0.40.04180.04220.04170.04290.04210.04250.04180.04250.04750.04410.0422
Table A3. Rank of methods according to ARAS, CODAS, EDAS, MABAC, MARCOS, considering sensitivity analysis for the third use case.
Table A3. Rank of methods according to ARAS, CODAS, EDAS, MABAC, MARCOS, considering sensitivity analysis for the third use case.
Weight Change (δ)ARAS RankCODAS RankEDAS RankMABAC RankMARCOS Rank
−0.4A1 > A3 > A2 > A4A1 > A3 > A2 > A4A1 > A3 > A2 > A4A1 > A3 > A2 > A4A1 > A3 > A2 > A4
−0.358A1 > A3 > A2 > A4A1 > A3 > A2 > A4A1 > A3 > A2 > A4A1 > A3 > A2 > A4A1 > A3 > A2 > A4
−0.316A1 > A3 > A2 > A4A1 > A3 > A2 > A4A1 > A3 > A2 > A4A1 > A3 > A2 > A4A1 > A3 > A2 > A4
−0.274A1 > A3 > A2 > A4A1 > A3 > A2 > A4A1 > A3 > A2 > A4A1 > A3 > A2 > A4A1 > A3 > A2 > A4
−0.232A1 > A3 > A2 > A4A1 > A3 > A2 > A4A1 > A3 > A2 > A4A1 > A3 > A2 > A4A1 > A3 > A2 > A4
−0.189A1 > A3 > A2 > A4A1 > A3 > A2 > A4A1 > A3 > A2 > A4A1 > A3 > A2 > A4A1 > A3 > A2 > A4
−0.147A1 > A3 > A2 > A4A1 > A3 > A2 > A4A1 > A3 > A2 > A4A1 > A3 > A2 > A4A1 > A3 > A2 > A4
−0.105A1 > A3 > A2 > A4A1 > A3 > A2 > A4A1 > A3 > A2 > A4A1 > A3 > A2 > A4A1 > A3 > A2 > A4
−0.063A1 > A2 > A3 > A4A1 > A2 > A3 > A4A1 > A3 > A2 > A4A1 > A2 > A3 > A4A1 > A2 > A3 > A4
−0.021A2 > A1 > A3 > A4A2 > A3 > A4 > A1A2 > A1 > A3 > A4A2 > A1 > A4 > A3A2 > A1 > A3 > A4
0.021A2 > A1 > A3 > A4A2 > A3 > A1 > A4A2 > A1 > A3 > A4A1 > A2 > A3 > A4A2 > A1 > A3 > A4
0.063A2 > A1 > A3 > A4A2 > A3 > A1 > A4A2 > A1 > A3 > A4A1 > A2 > A3 > A4A1 > A2 > A3 > A4
0.105A1 > A2 > A3 > A4A2 > A3 > A1 > A4A2 > A1 > A3 > A4A1 > A2 > A3 > A4A1 > A2 > A3 > A4
0.147A1 > A2 > A3 > A4A2 > A3 > A1 > A4A1 > A2 > A3 > A4A1 > A2 > A3 > A4A1 > A2 > A3 > A4
0.189A1 > A2 > A3 > A4A2 > A3 > A1 > A4A1 > A2 > A3 > A4A1 > A2 > A3 > A4A1 > A2 > A3 > A4
0.232A1 > A2 > A3 > A4A2 > A1 > A3 > A4A1 > A2 > A3 > A4A1 > A2 > A3 > A4A1 > A2 > A3 > A4
0.274A1 > A2 > A3 > A4A1 > A2 > A3 > A4A1 > A2 > A3 > A4A1 > A2 > A3 > A4A1 > A2 > A3 > A4
0.316A1 > A2 > A3 > A4A1 > A2 > A3 > A4A1 > A2 > A3 > A4A1 > A2 > A3 > A4A1 > A2 > A3 > A4
0.358A1 > A2 > A3 > A4A1 > A2 > A3 > A4A1 > A2 > A3 > A4A1 > A2 > A3 > A4A1 > A2 > A3 > A4
0.4A1 > A2 > A3 > A4A1 > A2 > A3 > A4A1 > A2 > A3 > A4A1 > A2 > A3 > A4A1 > A2 > A3 > A4
A1: LIME. A2: SHAP. A3: SmoothGrad. A4: Integrated Gradients.
Table A4. Rank of methods according to PROMETHEE_II. TOPSIS. VIKOR. WASPAS. WSM, considering sensitivity analysis for the third use case.
Table A4. Rank of methods according to PROMETHEE_II. TOPSIS. VIKOR. WASPAS. WSM, considering sensitivity analysis for the third use case.
Weight Change (δ)PROMETHEE_II RankTOPSIS RankVIKOR RankWASPAS RankWSM Rank
−0.4A1 > A3 > A2 > A4A1 > A3 > A2 > A4A3 > A2 > A1 > A4A1 > A3 > A2 > A4A1 > A2 > A3 > A4
−0.358A1 > A3 > A2 > A4A1 > A3 > A2 > A4A3 > A2 > A1 > A4A1 > A3 > A2 > A4A1 > A2 > A3 > A4
−0.316A1 > A3 > A2 > A4A1 > A3 > A2 > A4A3 > A2 > A1 > A4A1 > A3 > A2 > A4A1 > A3 > A2 > A4
−0.274A1 > A3 > A2 > A4A1 > A3 > A2 > A4A3 > A2 > A1 > A4A1 > A3 > A2 > A4A1 > A3 > A2 > A4
−0.232A1 > A3 > A2 > A4A1 > A3 > A2 > A4A3 > A2 > A1 > A4A1 > A3 > A2 > A4A1 > A3 > A2 > A4
−0.189A1 > A3 > A2 > A4A1 > A3 > A2 > A4A3 > A2 > A1 > A4A1 > A3 > A2 > A4A1 > A3 > A2 > A4
−0.147A1 > A3 > A2 > A4A1 > A3 > A2 > A4A3 > A2 > A1 > A4A1 > A3 > A2 > A4A1 > A3 > A2 > A4
−0.105A1 > A3 > A2 > A4A1 > A3 > A2 > A4A1 > A3 > A2 > A4A1 > A3 > A2 > A4A1 > A3 > A2 > A4
−0.063A1 > A3 > A2 > A4A1 > A2 > A3 > A4A2 > A1 > A3 > A4A1 > A2 > A3 > A4A1 > A2 > A3 > A4
−0.021A2 > A1 > A4 > A3A2 > A4 > A1 > A3A2 > A4 > A1 > A3A2 > A1 > A3 > A4A2 > A1 > A3 > A4
0.021A1 > A2 > A3 > A4A2 > A1 > A3 > A4A2 > A1 > A4 > A3A2 > A1 > A3 > A4A2 > A1 > A3 > A4
0.063A1 > A2 > A3 > A4A2 > A1 > A3 > A4A2 > A1 > A4 > A3A1 > A2 > A3 > A4A2 > A1 > A3 > A4
0.105A1 > A2 > A3 > A4A1 > A2 > A3 > A4A1 > A2 > A4 > A3A1 > A2 > A3 > A4A1 > A2 > A3 > A4
0.147A1 > A2 > A3 > A4A1 > A2 > A3 > A4A1 > A2 > A4 > A3A1 > A2 > A3 > A4A1 > A2 > A3 > A4
0.189A1 > A3 > A2 > A4A1 > A2 > A3 > A4A1 > A2 > A4 > A3A1 > A2 > A3 > A4A1 > A2 > A3 > A4
0.232A1 > A3 > A2 > A4A1 > A2 > A3 > A4A1 > A2 > A4 > A3A1 > A2 > A3 > A4A1 > A2 > A3 > A4
0.274A1 > A3 > A2 > A4A1 > A2 > A3 > A4A1 > A2 > A4 > A3A1 > A2 > A3 > A4A1 > A2 > A3 > A4
0.316A1 > A3 > A2 > A4A1 > A2 > A3 > A4A1 > A2 > A4 > A3A1 > A2 > A3 > A4A1 > A2 > A3 > A4
0.358A1 > A3 > A2 > A4A1 > A2 > A3 > A4A1 > A2 > A4 > A3A1 > A2 > A3 > A4A1 > A2 > A3 > A4
0.4A1 > A3 > A2 > A4A1 > A2 > A3 > A4A1 > A2 > A4 > A3A1 > A2 > A3 > A4A1 > A2 > A3 > A4
A1: LIME, A2: SHAP, A3: SmoothGrad, A4: Integrated Gradients

References

  1. Gunning, D. Explainable Artificial Intelligence (XAI); Defense Advanced Research Projects Agency (DARPA): Arlington, VA, USA, 2017. [Google Scholar]
  2. Linardatos, P.; Papastefanopoulos, V.; Kotsiantis, S. Explainable AI: A Review of Machine Learning Interpretability Methods. Entropy 2021, 23, 18. [Google Scholar] [CrossRef]
  3. Miller, T. Explanation in Artificial Intelligence: Insights from the Social Sciences. Artif. Intell. 2017, 267, 1–38. [Google Scholar] [CrossRef]
  4. Kim, B.; Khanna, R.; Koyejo, O.O. Examples Are Not Enough, Learn to Criticize! Criticism for Interpretability. Adv. Neural Inf. Process Syst. 2016, 29, 2288–2296. [Google Scholar]
  5. Doshi-Velez, F.; Kim, B. Towards A Rigorous Science of Interpretable Machine Learning. arXiv 2017, arXiv:1702.08608. [Google Scholar] [CrossRef]
  6. Gilpin, L.H.; Bau, D.; Yuan, B.Z.; Bajwa, A.; Specter, M.; Kagal, L. Explaining Explanations: An Overview of Interpretability of Machine Learning. In Proceedings of the 2018 IEEE 5th International Conference on Data Science and Advanced Analytics, Turin, Italy, 1–4 October 2018; pp. 80–89. [Google Scholar] [CrossRef]
  7. Barredo Arrieta, A.; Díaz-Rodríguez, N.; Del Ser, J.; Bennetot, A.; Tabik, S.; Barbado, A.; Garcia, S.; Gil-Lopez, S.; Molina, D.; Benjamins, R.; et al. Explainable Artificial Intelligence (XAI): Concepts, Taxonomies, Opportunities and Challenges toward Responsible AI. Inf. Fusion 2020, 58, 82–115. [Google Scholar] [CrossRef]
  8. Adadi, A.; Berrada, M. Peeking Inside the Black-Box: A Survey on Explainable Artificial Intelligence (XAI). IEEE Access 2018, 6, 52138–52160. [Google Scholar] [CrossRef]
  9. Guidotti, R.; Monreale, A.; Ruggieri, S.; Turini, F.; Giannotti, F.; Pedreschi, D. A Survey of Methods for Explaining Black Box Models. ACM Comput. Surv. 2019, 51, 1–42. [Google Scholar] [CrossRef]
  10. Ullah, N.; Khan, J.A.; De Falco, I.; Sannino, G. Explainable Artificial Intelligence: Importance, Use Domains, Stages, Output Shapes, and Challenges. ACM Comput. Surv. 2024, 57, 1–36. [Google Scholar] [CrossRef]
  11. Vermeire, T.; Laugel, T.; Renard, X.; Martens, D.; Detyniecki, M. How to Choose an Explainability Method? Towards a Methodical Implementation of XAI in Practice. In Machine Learning and Principles and Practice of Knowledge Discovery in Databases (ECML PKDD 2021); Communications in Computer and Information Science; Springer: Cham, Switzerland, 2021; Volume 1524, pp. 521–533. [Google Scholar] [CrossRef]
  12. Molnar, C. Interpretable Machine Learning: A Guide for Making Black Box Models Explainable, 2nd ed. Available online: https://christophm.github.io/interpretable-ml-book/ (accessed on 23 January 2024).
  13. Carvalho, D.V.; Pereira, E.M.; Cardoso, J.S. Machine Learning Interpretability: A Survey on Methods and Metrics. Electronics 2019, 8, 832. [Google Scholar] [CrossRef]
  14. Lipton, Z.C. The Mythos of Model Interpretability. Commun. ACM 2018, 61, 35–43. [Google Scholar] [CrossRef]
  15. Selvaraju, R.R.; Cogswell, M.; Das, A.; Vedantam, R.; Parikh, D.; Batra, D. Grad-Cam: Visual Explanations from Deep Networks via Gradient-Based Localization. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017. [Google Scholar]
  16. Bach, S.; Binder, A.; Montavon, G.; Klauschen, F.; Müller, K.R.; Samek, W. On Pixel-Wise Explanations for Non-Linear Classifier Decisions by Layer-Wise Relevance Propagation. PLoS ONE 2015, 10, e0130140. [Google Scholar] [CrossRef]
  17. Shrikumar, A.; Greenside, P.; Kundaje, A. Learning Important Features Through Propagating Activation Differences. In Proceedings of the 34th International Conference on Machine Learning, Centre, Sydney, 6–11 August 2017. [Google Scholar]
  18. Ribeiro, M.T.; Singh, S.; Guestrin, C. “Why Should I Trust You?”: Explaining the Predictions of Any Classifier. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; KDD ’16. Association for Computing Machinery: New York, NY, USA, 2016; pp. 1135–1144. [Google Scholar] [CrossRef]
  19. Lundberg, S.M.; Lee, S.I. A Unified Approach to Interpreting Model Predictions. In Proceedings of the 31st International Conference on Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; NIPS’17. Curran Associates Inc.: Red Hook, NY, USA, 2017; pp. 4768–4777. [Google Scholar]
  20. Ribeiro, M.T.; Singh, S.; Guestrin, C. Anchors: High-Precision Model-Agnostic Explanations. In Proceedings of the Thirty-Second AAAI Conference on Artificial Intelligence and Thirtieth Innovative Applications of Artificial Intelligence Conference and Eighth AAAI Symposium on Educational Advances in Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018; AAAI Press: New Orleans, Louisiana, USA, 2018; pp. 1527–1535. [Google Scholar]
  21. Simonyan, K.; Vedaldi, A.; Zisserman, A. Deep Inside Convolutional Networks: Visualising Image Classification Models and Saliency Maps. In Proceedings of the 2nd International Conference on Learning Representations, ICLR 2014—Workshop Track Proceedings, Banff, AB, Canada, 14–16 April 2014. [Google Scholar]
  22. Sundararajan, M.; Taly, A.; Yan, Q. Axiomatic Attribution for Deep Networks. In Proceedings of the 34th International Conference on Machine Learning, ICML 2017, Sydney, Australia, 6–11 August 2017; Volume 7, pp. 5109–5118. [Google Scholar]
  23. Goldstein, A.; Kapelner, A.; Bleich, J.; Pitkin, E. Peeking Inside the Black Box: Visualizing Statistical Learning With Plots of Individual Conditional Expectation. J. Comput. Graph. Stat. 2015, 24, 44–65. [Google Scholar] [CrossRef]
  24. Friedman, J.H. Greedy Function Approximation: A Gradient Boosting Machine. Ann. Statist. 2001, 29, 1189–1232. [Google Scholar] [CrossRef]
  25. Apley, D.W.; Zhu, J. Visualizing the Effects of Predictor Variables in Black Box Supervised Learning Models. J. R. Stat. Soc. Ser. B Stat. Methodol. 2016, 82, 1059–1086. [Google Scholar] [CrossRef]
  26. Vilone, G.; Longo, L. Classification of Explainable Artificial Intelligence Methods through Their Output Formats. Mach. Learn. Knowl. Extr. 2021, 3, 615–661. [Google Scholar] [CrossRef]
  27. Wachter, S.; Mittelstadt, B.; Russell, C. Counterfactual Explanations Without Opening the Black Box: Automated Decisions and the GDPR. Harv. J. Law Technol. 2017, 31, 841–887. [Google Scholar] [CrossRef]
  28. Vilone, G.; Longo, L. Explainable Artificial Intelligence: A Systematic Review. arXiv 2020, arXiv:2006.00093. [Google Scholar] [CrossRef]
  29. Cirqueira, D.; Nedbal, D.; Helfert, M.; Bezbradica, M. Scenario-Based Requirements Elicitation for User-Centric Explainable AI: A Case in Fraud Detection. In Machine Learning and Knowledge Extraction. CD-MAKE 2020; Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics); Springer: Cham, Switzerland, 2020; Volume 12279, pp. 321–341. [Google Scholar] [CrossRef]
  30. Wolf, C.T. Explainability Scenarios: Towards Scenario-Based XAI Design. In Proceedings of the International Conference on Intelligent User Interfaces, Proceedings IUI, Marina del Ray, CA, USA, 17–20 March 2019; pp. 252–257, Part. F147615. [Google Scholar] [CrossRef]
  31. Bhatt, U.; Xiang, A.; Sharma, S.; Weller, A.; Taly, A.; Jia, Y.; Ghosh, J.; Puri, R.; Moura, J.M.F.; Eckersley, P. Explainable Machine Learning in Deployment. In Proceedings of the FAT* 2020—Proceedings of the 2020 Conference on Fairness, Accountability, and Transparency, Barcelona, Spain, 27–30 January 2020; Association for Computing Machinery: New York, NY, USA, 2019; pp. 648–657. [Google Scholar] [CrossRef]
  32. Langer, M.; Oster, D.; Speith, T.; Hermanns, H.; Kästner, L.; Schmidt, E.; Sesing, A.; Baum, K. What Do We Want from Explainable Artificial Intelligence (XAI)?—A Stakeholder Perspective on XAI and a Conceptual Model Guiding Interdisciplinary XAI Research. Artif. Intell. 2021, 296, 103473. [Google Scholar] [CrossRef]
  33. Arya, V.; Bellamy, R.K.E.; Chen, P.-Y.; Dhurandhar, A.; Hind, M.; Hoffman, S.C.; Houde, S.; Liao, Q.V.; Luss, R.; Mojsilović, A.; et al. One Explanation Does Not Fit All: A Toolkit and Taxonomy of AI Explainability Techniques. arXiv 2019, arXiv:1909.03012. [Google Scholar] [CrossRef]
  34. Sokol, K.; Flach, P. Explainability Fact Sheets: A Framework for Systematic Assessment of Explainable Approaches. In Proceedings of the FAT* 2020—Proceedings of the 2020 Conference on Fairness, Accountability, and Transparency, Barcelona, Spain, 27–30 January 2020; Association for Computing Machinery: New York, NY, USA, 2020; pp. 56–67. [Google Scholar] [CrossRef]
  35. Cugny, R.; Aligon, J.; Chevalier, M.; Jimenez, G.R.; Teste, O. Why Should I Choose You? AutoXAI: A Framework for Selecting and Tuning EXplainable AI Solutions. In International Conference on Information and Knowledge Management, Proceedings; Association for Computing Machinery: New York, NY, USA, 2022; pp. 315–324. [Google Scholar] [CrossRef]
  36. Hall, M.; Harborne, D.; Tomsett, R.; Galetic, V.; Quintana-Amate, S.; Nottle, A.; Preece, A. A Systematic Method to Understand Requirements for Explainable AI (XAI) Systems. In Proceedings of the IJCAI Workshop on eXplainable Artificial Intelligence (XAI 2019), Macau, China, 11 August 2019; Volume 11. [Google Scholar]
  37. Adomavicius, G.; Mobasher, B.; Ricci, F.; Tuzhilin, A. Context-Aware Recommender Systems. AI Mag. 2011, 32, 67–80. [Google Scholar] [CrossRef]
  38. He, X.; Zhao, K.; Chu, X. AutoML: A Survey of the State-of-the-Art. Knowl. Based Syst. 2021, 212, 106622. [Google Scholar] [CrossRef]
  39. Jullum, M.; Sjødin, J.; Prabhu, R.; Løland, A. EXplego: An Interactive Tool That Helps You Select Appropriate XAI-Methods for Your Explainability Needs. In Proceedings of the xAI (Late-breaking Work, Demos, Doctoral Consortium), CEUR Workshop Proceedings, Aachen, Germany, 26–28 July 2023. [Google Scholar]
  40. Nauta, M.; Trienes, J.; Nguyen, E.; Peters, M.; Schmitt, Y.; Schlötterer, J.; Van Keulen, M.; Seifert, C. From Anecdotal Evidence to Quantitative Evaluation Methods: A Systematic Review on Evaluating Explainable AI.; From Anecdotal Evidence to Quantitative Evaluation Methods: A Systematic Review on Evaluating Explainable AI. ACM J. 2023, 55, 1–42. [Google Scholar] [CrossRef]
  41. Mitruț, O.; Moise, G.; Moldoveanu, A.; Moldoveanu, F.; Leordeanu, M.; Petrescu, L. Clarity in Complexity: How Aggregating Explanations Resolves the Disagreement Problem. Artif. Intell. Rev. 2024, 57, 338. [Google Scholar] [CrossRef]
  42. Rieger, L.; Hansen, L.K. Aggregating Explanation Methods for Stable and Robust Explainability. arXiv 2019, arXiv:1903.00519. [Google Scholar]
  43. Chatterjee, S.; Colombo, E.R.; Raimundo, M.M. Multi-Criteria Rank-Based Aggregation for Explainable AI. arXiv 2025, arXiv:2505.24612. [Google Scholar]
  44. Ghorabaee, M.K.; Kazimieras, Z.; Olfat, L.; Turskis, Z. Multi-Criteria Inventory Classification Using a New Method of Evaluation Based on Distance from Average Solution (EDAS). Informatica 2015, 26, 435–451. [Google Scholar] [CrossRef]
  45. Hwang, C.-L.; Yoon, K. Methods for Multiple Attribute Decision Making; Springer: Berlin/Heidelberg, Germany, 1981; pp. 58–191. [Google Scholar] [CrossRef]
  46. Zavadskas, E.; Kaklauskas, A.; Šarka, V. The New Method of Multicriteria Complex Proportional Assessment of Projects. Technol. Econ. Dev. Econ. 1994, 1, 131–139. [Google Scholar]
  47. Brans, J.P.; Mareschal, B. The Promethee Methods for MCDM; The Promcalc, Gaia and Bankadviser Software. In Readings in Multiple Criteria Decision Aid; Springer: Berlin/Heidelberg, Germany, 1990; pp. 216–252. [Google Scholar] [CrossRef]
  48. Zavadskas, E.K.; Turskis, Z. A New Additive Ratio Assessment (ARAS) Method in Multicriteria Decision-Making. Technol. Econ. Dev. Econ. 2010, 16, 159–172. [Google Scholar] [CrossRef]
  49. Yazdani, M.; Zarate, P.; Zavadskas, E.K.; Turskis, Z. A Combined Compromise Solution (CoCoSo) Method for Multi-Criteria Decision-Making Problems. Manag. Decis. 2019, 57, 2501–2519. [Google Scholar] [CrossRef]
  50. Ghorabaee, M.K.; Zavadskas, E.K.; Turkis, Z.; Antucheviciene, J. A New Combinative Distance-Based Assessment (Codas) Method for Multi-Criteria Decision-Making. Econ. Comput. Econ. Cybern. Stud. Res. 2016, 50, 25–44. [Google Scholar]
  51. Pamučar, D.; Ćirović, G. The Selection of Transport and Handling Resources in Logistics Centers Using Multi-Attributive Border Approximation Area Comparison (MABAC). Expert. Syst. Appl. 2015, 42, 3016–3028. [Google Scholar] [CrossRef]
  52. Anderková, V.; Babič, F.; Paraličová, Z.; Javorská, D. Intelligent System Using Data to Support Decision-Making. Appl. Sci. 2025, 15, 7724. [Google Scholar] [CrossRef]
  53. Behnke, J. Bordas Text „Mémoire Sur Les Élections Au Scrutin “von 1784: Einige Einführende Bemerkungen. In Jahrbuch für Handlungs-und Entscheidungstheorie; VS Verlag für Sozialwissenschaften: Wiesbaden, Germany, 2004; pp. 155–177. [Google Scholar] [CrossRef]
  54. Bodria, F.; Giannotti, F.; Guidotti, R.; Naretto, F.; Pedreschi, D.; Rinzivillo, S. Benchmarking and Survey of Explanation Methods for Black Box Models. Data Min. Knowl. Discov. 2023, 37, 1719–1778. [Google Scholar] [CrossRef]
  55. Burkart, N.; Huber, M.F. A Survey on the Explainability of Supervised Machine Learning. MF Huber J. Artif. Intell. Res. 2021, 70, 245–317. [Google Scholar] [CrossRef]
  56. Alvarez-Melis, D.; Jaakkola, T.S. Towards Robust Interpretability with Self-Explaining Neural Networks. In Proceedings of the 32nd International Conference on Neural Information Processing Systems; NIPS’18; Curran Associates Inc.: Red Hook, NY, USA, 2018; pp. 7786–7795. [Google Scholar]
  57. Moraffah, R.; Karami, M.; Guo, R.; Raglin, A.; Liu, H. Causal Interpretability for Machine Learning-Problems, Methods and Evaluation. ACM SIGKDD Explor. Newsl. 2020, 22, 18–33. [Google Scholar] [CrossRef]
  58. Yeh, C.K.; Hsieh, C.Y.; Suggala, A.S.; Inouye, D.I.; Ravikumar, P. On the (In)Fidelity and Sensitivity for Explanations. Adv. Neural Inf. Process Syst. 2019, 32, 10967–10978. [Google Scholar]
  59. Chuang, Y.-N.; Wang, G.; Yang, F.; Liu, Z.; Cai, X.; Du, M.; Hu, X. Efficient XAI Techniques: A Taxonomic Survey. arXiv 2023, arXiv:2302.03225. [Google Scholar] [CrossRef]
  60. Samek, W.; Montavon, G.; Lapuschkin, S.; Anders, C.J.; Müller, K.R. Explaining Deep Neural Networks and Beyond: A Review of Methods and Applications. Proc. IEEE 2021, 109, 247–278. [Google Scholar] [CrossRef]
  61. Samek, W.; Wiegand, T.; Müller, K.-R. Explainable Artificial Intelligence: Understanding, Visualizing and Interpreting Deep Learning Models. arXiv 2017, arXiv:1708.08296. [Google Scholar] [CrossRef]
  62. Darias, J.M.; Bayrak, B.; Caro-Martínez, M.; Díaz-Agudo, B.; Recio-Garcia, J.A. An Empirical Analysis of User Preferences Regarding XAI Metrics. In Case-Based Reasoning Research and Development. ICCBR 2024; Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics); Springer: Cham, Switzerland, 2024; Volume 14775, pp. 96–110. [Google Scholar] [CrossRef]
  63. Li, X.; Du, M.; Chen, J.; Chai, Y.; Lakkaraju, H.; Xiong, H. M4: A Unified XAI Benchmark for Faithfulness Evaluation of Feature Attribution Methods across Metrics, Modalities and Models. In Proceedings of the 37th International Conference on Neural Information Processing Systems, New Orleans, LA, USA, 10–16 December 2023; NIPS ’23. Curran Associates Inc.: Red Hook, NY, USA, 2023. [Google Scholar]
  64. Hedström, A.; Weber, L.; Bareeva, D.; Krakowczyk, D.; Motzkus, F.; Samek, W.; Lapuschkin, S.; Höhne, M.M.C. Quantus: An Explainable AI Toolkit for Responsible Evaluation of Neural Network Explanations and Beyond. J. Mach. Learn. Res. 2022, 24, 1–11. [Google Scholar]
  65. Sithakoul, S.; Meftah, S.; Feutry, C. BEExAI: Benchmark to Evaluate Explainable AI. Commun. Comput. Inf. Sci. 2024, 2153, 445–468. [Google Scholar] [CrossRef]
  66. Canha, D.; Kubler, S.; Främling, K.; Fagherazzi, G. A Functionally-Grounded Benchmark Framework for XAI Methods: Insights and Foundations from a Systematic Literature Review. ACM Comput. Surv. 2025, 57, 40. [Google Scholar] [CrossRef]
  67. Lee, J.R.; Emami, S.; Hollins, M.D.; Wong, T.C.H.; Villalobos Sánchez, C.I.; Toni, F.; Zhang, D.; Dejl, A. XAI-Units: Benchmarking Explainability Methods with Unit Tests. In ACMF AccT 2025—Proceedings of the 2025 ACM Conference on Fairness, Accountability, and Transparency, Athens, Greece, 23–26 June 2025; ACM: New York, NY, USA, 2025; pp. 2892–2905. [Google Scholar] [CrossRef]
  68. Belaid, M.K.; Hüllermeier, E.; Rabus, M.; Krestel, R. Do We Need Another Explainable AI Method? Toward Unifying Post-Hoc XAI Evaluation Methods into an Interactive and Multi-Dimensional Benchmark. arXiv 2022, arXiv:2207.14160. [Google Scholar]
  69. Moiseev, I.; Balabaeva, K.; Kovalchuk, S. Open and Extensible Benchmark for Explainable Artificial Intelligence Methods. Algorithms 2025, 18, 85. [Google Scholar] [CrossRef]
  70. Liu, Y.; Khandagale, S.; White, C.; Neiswanger, W. Synthetic Benchmarks for Scientific Research in Explainable Machine Learning. arXiv 2021, arXiv:2106.12543. [Google Scholar] [CrossRef]
  71. Agarwal, C.; Ley, D.; Krishna, S.; Saxena, E.; Pawelczyk, M.; Johnson, N.; Puri, I.; Zitnik, M.; Lakkaraju, H. OpenXAI: Towards a Transparent Evaluation of Model Explanations. Adv. Neural Inf. Process. Syst. 2022, 35, 15784–15799. [Google Scholar]
  72. Zhang, Y.; Song, J.; Gu, S.; Jiang, T.; Pan, B.; Bai, G.; Zhao, L. Saliency-Bench: A Comprehensive Benchmark for Evaluating Visual Explanations. In Proceedings of the 31st ACM SIGKDD Conference on Knowledge Discovery and Data Mining V.2, Toronto, ON, Canada, 3–7 August 2023; Volume 1, pp. 5924–5935. [Google Scholar]
  73. Ma, J.; Lai, V.; Zhang, Y.; Chen, C.; Hamilton, P.; Ljubenkov, D.; Lakkaraju, H.; Tan, C. OpenHEXAI: An Open-Source Framework for Human-Centered Evaluation of Explainable Machine Learning. arXiv 2024, arXiv:2403.05565. [Google Scholar]
  74. Aechtner, J.; Cabrera, L.; Katwal, D.; Onghena, P.; Valenzuela, D.P.; Wilbik, A. Comparing User Perception of Explanations Developed with XAI Methods. In Proceedings of the 2022 IEEE International Conference on Fuzzy Systems (FUZZ-IEEE), Padua, Italy, 18–23 July 2022. [Google Scholar] [CrossRef]
  75. Aruldoss, M.; Lakshmi, T.M.; Venkatesan, V.P. A Survey on Multi Criteria Decision Making Methods and Its Applications. Am. J. Inf. Syst. 2013, 1, 31–43. [Google Scholar]
  76. Velasquez, M.; Hester, P.T. An Analysis of Multi-Criteria Decision Making Methods. Int. J. Oper. Res. 2013, 10, 55–66. [Google Scholar]
  77. Ryciuk, U.; Kiryluk, H.; Hajduk, S. Multi-Criteria Analysis in the Decision-Making Approach for the Linear Ordering of Urban Transport Based on TOPSIS Technique. Energies 2021, 15, 274. [Google Scholar] [CrossRef]
  78. Baczkiewicz, A.; Atróbski, J.W.; Kizielewicz, B.; Sałabun, W. Towards Objectification of Multi-Criteria Assessments: A Comparative Study on MCDA Methods. In Proceedings of the 2021 16th Conference on Computer Science and Intelligence Systems, Sofia, Bulgaria, 2–5 September 2021; Volume 25, pp. 417–425. [Google Scholar] [CrossRef]
  79. Baizyldayeva, U.; Vlasov, O.; Kuandykov, A.A.; Akhmetov, T.B. Multi-Criteria Decision Support Systems. Comparative Analysis. Middle-East. J. Sci. Res. 2013, 16, 1725–1730. [Google Scholar]
  80. Habenicht, W.; Scheubrein, B.; Scheubrein, R. Multiple-Criteria Decision Making. Optim. Oper. Res. 2002, 4, 257–279. [Google Scholar]
  81. Chakraborty, S.; Raut, R.D.; Rofin, T.M.; Chakraborty, S. A Comprehensive and Systematic Review of Multi-Criteria Decision-Making Methods and Applications in Healthcare. Healthc. Anal. 2023, 4, 100232. [Google Scholar] [CrossRef]
  82. Pohekar, S.D.; Ramachandran, M. Application of Multi-Criteria Decision Making to Sustainable Energy Planning—A Review. Renew. Sustain. Energy Rev. 2004, 8, 365–381. [Google Scholar] [CrossRef]
  83. Saaty, T.L. Analytic Hierarchy Process Planning, Priority Setting, Resource Allocation; McGraw-Hill, Inc: New York, NY, USA, 1980. [Google Scholar]
  84. Saaty, T.L. Decision Making for Leaders: The Analytic Hierarchy Process for Decisions in a Complex World; RWS Publications: Pittsburgh, PA, USA, 1992. [Google Scholar]
  85. Diakoulaki, D.; Mavrotas, G.; Papayannakis, L. Determining Objective Weights in Multiple Criteria Problems: The Critic Method. Comput. Ops Res. 1995, 22, 763–770. [Google Scholar] [CrossRef]
  86. Stević, Ž.; Pamučar, D.; Puška, A.; Chatterjee, P. Sustainable Supplier Selection in Healthcare Industries Using a New MCDM Method: Measurement of Alternatives and Ranking According to COmpromise Solution (MARCOS). Comput. Ind. Eng. 2020, 140, 106231. [Google Scholar] [CrossRef]
  87. Jiří, M. The Robustness of TOPSIS Results Using Sensitivity Analysis Based on Weight Tuning. IFMBE Proc. 2018, 68, 83–86. [Google Scholar] [CrossRef]
  88. Duckstein, L.; Opricovic, S. Multiobjective Optimization in River Basin Development. Water Resour. Res. 1980, 16, 14–20. [Google Scholar] [CrossRef]
  89. Zavadskas, E.K.; Turskis, Z.; Antucheviciene, J.; Zakarevicius, A. Optimization of Weighted Aggregated Sum Product Assessment. Elektron. Ir Elektrotechnika 2012, 122, 3–6. [Google Scholar] [CrossRef]
  90. Fishburn, P.C.; Murphy, A.H.; Isaacs, H.H. Sensitivity of Decisions to Probability Estimation Errors: A Reexamination. Oper. Res. 1968, 16, 254–267. [Google Scholar] [CrossRef]
  91. Štrumbelj, E.; Kononenko, I. Explaining Prediction Models and Individual Predictions with Feature Contributions. Knowl. Inf. Syst. 2013, 41, 647–665. [Google Scholar] [CrossRef]
  92. Koh, P.W.; Liang, P. Understanding Black-Box Predictions via Influence Functions. In Proceedings of the 34th International Conference on Machine Learning, ICML 2017, Sydney, Australia, 6–11 August 2017; Volume 4, pp. 2976–2987. [Google Scholar]
  93. Kim, B.; Wattenberg, M.; Gilmer, J.; Cai, C.; Wexler, J.; Viegas, F.; Sayres, R. Interpretability Beyond Feature Attribution: Quantitative Testing with Concept Activation Vectors (TCAV). In Proceedings of the 35th International Conference on Machine Learning, ICML 2018, Stockholm, Sweden, 10–15 July 2018; Volume 6, pp. 4186–4195. [Google Scholar]
  94. Montavon, G.; Bach, S.; Binder, A.; Samek, W.; Müller, K.-R. Explaining NonLinear Classification Decisions with Deep Taylor Decomposition. Pattern Recognit. 2015, 65, 211–222. [Google Scholar] [CrossRef]
  95. Zeiler, M.D.; Fergus, R. Visualizing and Understanding Convolutional Networks. In Computer Vision–ECCV 2014. ECCV 2014; Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics); Springer: Cham, Switzerland, 2013; Volume 8689, pp. 818–833. [Google Scholar] [CrossRef]
  96. Smilkov, D.; Thorat, N.; Kim, B.; Viégas, F.; Wattenberg, M. SmoothGrad: Removing Noise by Adding Noise. arXiv 2017, arXiv:1706.03825. [Google Scholar] [CrossRef]
  97. Kindermans, P.J.; Schütt, K.T.; Alber, M.; Müller, K.R.; Erhan, D.; Kim, B.; Dähne, S. Learning How to Explain Neural Networks: PatternNet and PatternAttribution. In Proceedings of the 6th International Conference on Learning Representations, ICLR 2018—Conference Track Proceedings, Vancouver, BC, Canada, 30 April–3 May 2018. [Google Scholar]
  98. Springenberg, J.T.; Dosovitskiy, A.; Brox, T.; Riedmiller, M. Striving for Simplicity: The All Convolutional Net. In Proceedings of the 3rd International Conference on Learning Representations, ICLR 2015—Workshop Track Proceedings, San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
  99. Fong, R.C.; Vedaldi, A. Interpretable Explanations of Black Boxes by Meaningful Perturbation. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 3449–3457. [Google Scholar] [CrossRef]
  100. Xu, K.; Ba, J.L.; Kiros, R.; Cho, K.; Courville, A.; Salakhutdinov, R.; Zemel, R.S.; Bengio, Y. Show, Attend and Tell: Neural Image Caption Generation with Visual Attention. In Proceedings of the 32nd International Conference on Machine Learning, ICML 2015, Lille, France, 6–11 July 2015; Volume 3, pp. 2048–2057. [Google Scholar]
  101. Erhan, D.; Courville, A.; Bengio, Y. Understanding Representations Learned in Deep Architectures; Department Dinformatique et Recherche Operationnelle, University of Montreal: Montreal, QC, Canada, 2010. [Google Scholar]
  102. Zhou, B.; Khosla, A.; Lapedriza, A.; Oliva, A.; Torralba, A. Learning Deep Features for Discriminative Localization. 2016, pp 2921–2929. Available online: http://cnnlocalization.csail.mit.edu (accessed on 23 January 2024).
  103. Lei, T.; Barzilay, R.; Jaakkola, T. Rationalizing Neural Predictions. In EMNLP 2016—Proceedings of the Conference on Empirical Methods in Natural Language Processing, Proceedings, Austin, TX, USA, 1–5 November 2016; Curran Associates Inc.: Red Hook, NY, USA, 2016; pp. 107–117. [Google Scholar] [CrossRef]
  104. Guidotti, R.; Monreale, A.; Ruggieri, S.; Pedreschi, D.; Turini, F.; Giannotti, F. Local Rule-Based Explanations of Black Box Decision Systems. arXiv 2018, arXiv:1805.10820. [Google Scholar] [CrossRef]
  105. Zintgraf, L.M.; Cohen, T.S.; Adel, T.; Welling, M. Visualizing Deep Neural Network Decisions: Prediction Difference Analysis. In Proceedings of the 5th International Conference on Learning Representations, ICLR 2017—Conference Track Proceedings, Toulon, France, 24–26 April 2017. [Google Scholar]
  106. Petsiuk, V.; Das, A.; Saenko, K. RISE: Randomized Input Sampling for Explanation of Black-Box Models. In Proceedings of the British Machine Vision Conference 2018, BMVC 2018, Newcastle, UK, 3–6 September 2018. [Google Scholar]
  107. Dhurandhar, A.; Chen, P.-Y.; Luss, R.; Tu, C.-C.; Ting, P.; Shanmugam, K.; Das, P. Explanations Based on the Missing: Towards Contrastive Explanations with Pertinent Negatives. In Proceedings of the 32nd International Conference on Neural Information Processing Systems, Montreal, QC, Canada, 3–8 December 2018; NIPS’18. Curran Associates Inc.: Red Hook, NY, USA, 2018; pp. 590–601. [Google Scholar]
  108. Van Looveren, A.; Klaise, J. Interpretable Counterfactual Explanations Guided by Prototypes. In Machine Learning and Knowledge Discovery in Databases; Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics); Springer: Cham, Switzerland, 2019; Volume 12976, pp. 650–665. [Google Scholar] [CrossRef]
  109. Mothilal, R.K.; Sharma, A.; Tan, C. Explaining Machine Learning Classifiers through Diverse Counterfactual Explanations. In FAT* 2020—Proceedings of the 2020 Conference on Fairness, Accountability, and Transparency, Barcelona, Spain, 27–30 January 2020; Association for Computing Machinery: New York, NY, USA, 2020; pp. 607–617. [Google Scholar] [CrossRef]
  110. Ghorbani, A.; Wexler, J.; Zou, J.; Kim, B. Towards Automatic Concept-Based Explanations. In Proceedings of the 33rd International Conference on Neural Information Processing Systems, Vancouver, BC, Canada, 8–14 December 2019; Curran Associates Inc.: Red Hook, NY, USA, 2019; pp. 9277–9286. [Google Scholar]
  111. Yeh, C.K.; Kim, B.; Arik, S.; Li, C.L.; Pfister, T.; Ravikumar, P. On Completeness-Aware Concept-Based Explanations in Deep Neural Networks. In Proceedings of the 34th International Conference on Neural Information Processing Systems, Vancouver, BC, Canada, 6–12 December 2020; NIPS ’20. Curran Associates Inc.: Red Hook, NY, USA, 2020; pp. 20554–20565. [Google Scholar]
  112. Frosst, N.; Hinton, G. Distilling a Neural Network into a Soft Decision Tree. In Proceedings of the First International Workshop on Comprehensibility and Explanation in AI and ML 2017 Co-Located with 16th International Conference of the Italian Association for Artificial Intelligence (AI*IA 2017), CEUR Workshop Proceedings, Bari, Italy, 16–17 November 2017; Volume 2071. [Google Scholar]
  113. Främling, K.; Graillot, D. Extracting Explanations from Neural Networks. In Proceedings of the ICANN’95 Conference, Paris, France, 9–13 October 1995. [Google Scholar]
  114. Yosinski, J.; Clune, J.; Nguyen, A.; Fuchs, T.; Lipson, H. Understanding Neural Networks Through Deep Visualization. arXiv 2015, arXiv:1506.06579. [Google Scholar] [CrossRef]
  115. Adler, P.; Falk, C.; Friedler, S.A.; Nix, T.; Rybeck, G.; Scheidegger, C.; Smith, B.; Venkatasubramanian, S. Auditing Black-Box Models for Indirect Influence. Knowl. Inf. Syst. 2016, 54, 95–122. [Google Scholar] [CrossRef]
  116. Bau, D.; Zhu, J.Y.; Strobelt, H.; Zhou, B.; Tenenbaum, J.B.; Freeman, W.T.; Torralba, A. GAN Dissection: Visualizing and Understanding Generative Adversarial Networks. In Proceedings of the 7th International Conference on Learning Representations, ICLR 2019, New Orleans, LA, USA, 6–9 May 2019. [Google Scholar]
  117. Poyiadzi, R.; Sokol, K.; Santos-Rodriguez, R.; De Bie, T.; Flach, P. FACE: Feasible and Actionable Counterfactual Explanations. In AIES 2020—Proceedings of the AAAI/ACM Conference on AI, Ethics, and Society, New York, NY, USA, 7–8 February 2020; Association for Computing Machinery: New York, NY, USA, 2020; pp. 344–350. [Google Scholar] [CrossRef]
  118. Agarwal, R.; Melnick, L.; Frosst, N.; Zhang, X.; Lengerich, B.; Caruana, R.; Hinton, G.E. Neural Additive Models: Interpretable Machine Learning with Neural Nets. Adv. Neural Inf. Process Syst. 2020, 6, 4699–4711. [Google Scholar]
  119. Ming, Y.; Qu, H.; Bertini, E. RuleMatrix: Visualizing and Understanding Classifiers with Rules. IEEE Trans. Vis. Comput. Graph. 2019, 25, 342–352. [Google Scholar] [CrossRef]
  120. Chattopadhay, A.; Sarkar, A.; Howlader, P.; Balasubramanian, V.N. Grad-CAM++: Generalized Gradient-Based Visual Explanations for Deep Convolutional Networks. In Proceedings of the 2018 IEEE Winter Conference on Applications of Computer Vision, WACV 2018, Lake Tahoe, NV, USA, 12–15 March 2018; pp. 839–847. [Google Scholar] [CrossRef]
  121. Chen, J.; Song, L.; Wainwright, M.J.; Jordan, M.I. Learning to Explain: An Information-Theoretic Perspective on Model Interpretation. In Proceedings of the 35th International Conference on Machine Learning, ICML 2018, Stockholm, Sweden, 10–15 July 2018; Volume 2, pp. 1386–1418. [Google Scholar]
  122. Plumb, G.; Molitor, D.; Talwalkar, A. Model Agnostic Supervised Local Explanations. Adv. Neural Inf. Process Syst. 2018, 31, 2515–2524. [Google Scholar]
  123. Altmann, A.; Toloşi, L.; Sander, O.; Lengauer, T. Permutation Importance: A Corrected Feature Importance Measure. Bioinformatics 2010, 26, 1340–1347. [Google Scholar] [CrossRef]
  124. Staniak, M.; Biecek, P. Explanations of Model Predictions with Live and BreakDown Packages. R. J. 2018, 10, 395–409. [Google Scholar] [CrossRef]
  125. Guidotti, R.; Monreale, A.; Matwin, S.; Pedreschi, D. Explaining Image Classifiers Generating Exemplars and Counter-Exemplars from Latent Representations. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; AAAI Press: Palo Alto, CA, USA, 2020; Volume 34, pp. 13665–13668. [Google Scholar] [CrossRef]
  126. Dandl, S.; Molnar, C.; Binder, M.; Bischl, B. Multi-Objective Counterfactual Explanations. In Parallel Problem Solving from Nature –PPSN XVI; Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics); Springer: Cham, Switzerland, 2020; Volume 12269, pp. 448–469. [Google Scholar] [CrossRef]
  127. Martens, D.; Provost, F. Explaining Data-Driven Document Classifications1. Manag. Inf. Syst. Q. 2014, 38, 73–100. [Google Scholar] [CrossRef]
  128. Huang, Q.; Yamada, M.; Tian, Y.; Singh, D.; Chang, Y. GraphLIME: Local Interpretable Model Explanations for Graph Neural Networks. IEEE Trans. Knowl. Data Eng. 2020, 35, 6968–6972. [Google Scholar] [CrossRef]
  129. Setzu, M.; Guidotti, R.; Monreale, A.; Turini, F. Global Explanations with Local Scoring. Commun. Comput. Inf. Sci. 2020, 1167, 159–171. [Google Scholar] [CrossRef]
  130. Blanco-Justicia, A.; Domingo-Ferrer, J.; Martínez, S.; Sánchez, D. Machine Learning Explainability via Microaggregation and Shallow Decision Trees. Knowl. Based Syst. 2020, 194, 105532. [Google Scholar] [CrossRef]
  131. Casalicchio, G.; Molnar, C.; Bischl, B. Visualizing the Feature Importance for Black Box Models. In Machine Learning and Knowledge Discovery in Databases; Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics); Springer: Cham, Switzerland, 2019; Volume 11051, pp. 655–670. [Google Scholar] [CrossRef]
  132. Datta, A.; Sen, S.; Zick, Y. Algorithmic Transparency via Quantitative Input Influence: Theory and Experiments with Learning Systems. In Proceedings of the 2016 IEEE Symposium on Security and Privacy, SP 2016, San Jose, CA, USA, 23–25 May 2016; pp. 598–617. [Google Scholar] [CrossRef]
  133. Lucic, A.; Oosterhuis, H.; Haned, H.; de Rijke, M. FOCUS: Flexible Optimizable Counterfactual Explanations for Tree Ensembles. In Proceedings of the 36th AAAI Conference on Artificial Intelligence, AAAI 2022, Online, 22 February–1 March 2019; Volume 36, pp. 5313–5322. [Google Scholar] [CrossRef]
  134. Mahajan, D.; Tan, C.; Sharma, A. Preserving Causal Constraints in Counterfactual Explanations for Machine Learning Classifiers. arXiv 2019, arXiv:1912.03277. [Google Scholar]
  135. Russell, C. Efficient Search for Diverse Coherent Explanations. In FAT* 2019—Proceedings of the 2019 Conference on Fairness, Accountability, and Transparency, Atlanta, GA, USA, 29–31 January 2019; Association for Computing Machinery: New York, NY, USA, 2019; pp. 20–28. [Google Scholar] [CrossRef]
  136. Ustun, B.; Spangher, A.; Liu, Y. Actionable Recourse in Linear Classification. In FAT* 2019—Proceedings of the 2019 Conference on Fairness, Accountability, and Transparency, Atlanta, GA, USA, 29–31 January 2019; Association for Computing Machinery: New York, NY, USA, 2019; pp. 10–19. [Google Scholar] [CrossRef]
  137. Kanamori, K.; Takagi, T.; Kobayashi, K.; Arimura, H. DACE: Distribution-Aware Counterfactual Explanation by Mixed-Integer Linear Optimization. IJCAI Int. Jt. Conf. Artif. Intell. 2020, 3, 2855–2862. [Google Scholar] [CrossRef]
  138. Karimi, A.H.; Barthe, G.; Balle, B.; Valera, I. Model-Agnostic Counterfactual Explanations for Consequential Decisions. Proc. Mach. Learn. Res. 2019, 108, 895–905. [Google Scholar]
  139. Pawelczyk, M.; Broelemann, K.; Kasneci, G. Learning Model-Agnostic Counterfactual Explanations for Tabular Data. In Web Conference 2020—Proceedings of the World Wide Web Conference, WWW 2020, Taipei, Taiwan, 20–24 April 2020; Association for Computing Machinery: New York, NY, USA, 2020; pp. 3126–3132. [Google Scholar] [CrossRef]
  140. Ramakrishnan, G.; Lee, Y.C.; Albarghouthi, A. Synthesizing Action Sequences for Modifying Model Decisions. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; AAAI Press: Palo Alto, CA, USA, 2020; Volume 34, pp. 5462–5469. [Google Scholar] [CrossRef]
  141. Cheng, F.; Ming, Y.; Qu, H. DECE: Decision Explorer with Counterfactual Explanations for Machine Learning Models. IEEE Trans. Vis. Comput. Graph. 2020, 27, 1438–1447. [Google Scholar] [CrossRef]
  142. Karimi, A.H.; Schölkopf, B.; Valera, I. Algorithmic Recourse: From Counterfactual Explanations to Interventions. In FAccT 2021—Proceedings of the 2021 ACM Conference on Fairness, Accountability, and Transparency, Virtual Event/Toronto, Canada, 3–10 March 2020; Association for Computing Machinery: New York, NY, USA; pp. 353–362. [CrossRef]
  143. Laugel, T.; Lesot, M.-J.; Marsala, C.; Renard, X.; Detyniecki, M. Inverse Classification for Comparison-Based Interpretability in Machine Learning. arXiv 2017, arXiv:1712.08443. [Google Scholar] [CrossRef]
  144. Sharma, S.; Henderson, J.; Ghosh, J. CERTIFAI: A Common Framework to Provide Explanations and Analyse the Fairness and Robustness of Black-Box Models. In AIES 2020—Proceedings of the AAAI/ACM Conference on AI, Ethics, and Society, Madrid, Spain, 20–22 October 2020; Association for Computing Machinery: New York, NY, USA, 2020; pp. 166–172. [Google Scholar] [CrossRef]
  145. Gomez, O.; Holter, S.; Yuan, J.; Bertini, E. ViCE: Visual counterfactual explanations for machine learning models. In Proceedings of the International Conference on Intelligent User Interfaces, Proceedings IUI 2020, Cagliari, Italy, 24–27 March 2020; pp. 531–535. [Google Scholar] [CrossRef]
  146. Lucic, A.; Haned, H.; de Rijke, M. Why Does My Model Fail? Contrastive Local Explanations for Retail Forecasting. In FAT* 2020—Proceedings of the 2020 Conference on Fairness, Accountability, and Transparency, Barcelona, Spain, 27–30 January 2020; Association for Computing Machinery: New York, NY, USA, 2020; pp. 90–98. [Google Scholar] [CrossRef]
  147. Ramon, Y.; Martens, D.; Provost, F.; Evgeniou, T. A Comparison of Instance-Level Counterfactual Explanation Algorithms for Behavioral and Textual Data: SEDC, LIME-C and SHAP-C. Adv. Data Anal. Classif. 2020, 14, 801–819. [Google Scholar] [CrossRef]
  148. White, A.; D’Avila Garcez, A. Measurable Counterfactual Local Explanations for Any Classifier. Front. Artif. Intell. Appl. 2019, 325, 2529–2535. [Google Scholar] [CrossRef]
  149. Ying, R.; Bourgeois, D.; You, J.; Zitnik, M.; Leskovec, J. GNNExplainer: Generating Explanations for Graph Neural Networks. In Proceedings of the 33rd International Conference on Neural Information Processing Systems, Vancouver, BC, Canada, 8–14 December 2019; Curran Associates Inc.: Red Hook, NY, USA, 2019; pp. 9244–9255. [Google Scholar]
  150. Wang, J.; Wiens, J.; Lundberg, S. Shapley Flow: A Graph-Based Approach to Interpreting Model Predictions. Proc. Mach. Learn. Res. 2020, 130, 721–729. [Google Scholar]
  151. Sagi, O.; Rokach, L. Explainable Decision Forest: Transforming a Decision Forest into an Interpretable Tree. Inf. Fusion 2020, 61, 124–138. [Google Scholar] [CrossRef]
  152. Hatwell, J.; Gaber, M.M.; Azad, R.M.A. CHIRPS: Explaining Random Forest Classification. Artif. Intell. Rev. 2020, 53, 5747–5788. [Google Scholar] [CrossRef]
  153. Rajapaksha, D.; Bergmeir, C.; Buntine, W. LoRMIkA: Local Rule-Based Model Interpretability with k-Optimal Associations. Inf. Sci. (N. Y.) 2020, 540, 221–241. [Google Scholar] [CrossRef]
  154. Loor, M.; De Tré, G. Contextualizing Support Vector Machine Predictions. Int. J. Comput. Intell. Syst. 2020, 13, 1483–1497. [Google Scholar] [CrossRef]
  155. Li, H.; Tian, Y.; Mueller, K.; Chen, X. Beyond Saliency: Understanding Convolutional Neural Networks from Saliency Prediction on Layer-Wise Relevance Propagation. Image Vis. Comput. 2019, 83–84, 70–86. [Google Scholar] [CrossRef]
  156. Zafar, M.R.; Khan, N.M. DLIME: A Deterministic Local Interpretable Model-Agnostic Explanations Approach for Computer-Aided Diagnosis Systems. arXiv 2019, arXiv:1906.10263. [Google Scholar]
  157. Mollas, I.; Bassiliades, N.; Tsoumakas, G. LioNets: Local Interpretation of Neural Networks through Penultimate Layer Decoding. Commun. Comput. Inf. Sci. 2020, 1167, 265–276. [Google Scholar] [CrossRef]
  158. Kapishnikov, A.; Bolukbasi, T.; Viegas, F.; Terry, M. XRAI: Better Attributions Through Regions. In Proceedings of the IEEE International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 4947–4956. [Google Scholar] [CrossRef]
  159. Lampridis, O.; Guidotti, R.; Ruggieri, S. Explaining Sentiment Classification with Synthetic Exemplars and Counter-Exemplars. In Discovery Science. DS 2020; Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics); Springer: Cham, Switzerland, 2020; Volume 12323, pp. 357–373. [Google Scholar] [CrossRef]
  160. Hoover, B.; Strobelt, H.; Gehrmann, S. exBERT: A Visual Analysis Tool to Explore Learned Representations in Transformers Models. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics: System Demonstrations, Online, 5–10 July 2020; pp. 187–196. [Google Scholar] [CrossRef]
  161. Jacovi, A.; Shalom, O.S.; Goldberg, Y. Understanding Convolutional Neural Networks for Text Classification. In Proceedings of the EMNLP 2018—2018 EMNLP Workshop BlackboxNLP: Analyzing and Interpreting Neural Networks for NLP, Proceedings of the 1st Workshop, Brussels, Belgium, 1 November 2018; pp. 56–65. [Google Scholar] [CrossRef]
  162. Zhou, Y.; Zhu, Y.; Ye, Q.; Qiu, Q.; Jiao, J. Weakly Supervised Instance Segmentation Using Class Peak Response. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–20 June 2018; pp. 3791–3800. [Google Scholar] [CrossRef]
  163. Qin, Y.; Kamnitsas, K.; Ancha, S.; Nanavati, J.; Cottrell, G.; Criminisi, A.; Nori, A. Autofocus Layer for Semantic Segmentation. In Medical Image Computing and Computer Assisted Intervention–MICCAI 2018; Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics); Springer: Cham, Switzerland, 2018; Volume 11072, pp. 603–611. [Google Scholar] [CrossRef]
  164. Bondarenko, A.; Aleksejeva, L.; Jumutc, V.; Borisov, A. Classification Tree Extraction from Trained Artificial Neural Networks. Procedia Comput. Sci. 2017, 104, 556–563. [Google Scholar] [CrossRef]
  165. Burns, C.; Thomason, J.; Tansey, W. Interpreting Black Box Models via Hypothesis Testing. In FODS 2020—Proceedings of the 2020 ACM-IMS Foundations of Data Science Conference, Virtual Event, 19–20 October 2020; Association for Computing Machinery: New York, NY, USA, 2020; pp. 47–57. [Google Scholar] [CrossRef]
  166. Ibrahim, M.; Modarres, C.; Louie, M.; Paisley, J. Global Explanations of Neural Network: Mapping the Landscape of Predictions. In AIES 2019—Proceedings of the 2019 AAAI/ACM Conference on AI, Ethics, and Society; Association for Computing Machinery: New York, NY, USA, 2019; pp. 279–287. [Google Scholar] [CrossRef]
  167. Lengerich, B.J.; Konam, S.; Xing, E.P.; Rosenthal, S.; Veloso, M. Towards Visual Explanations for Convolutional Neural Networks via Input Resampling. arXiv 2017, arXiv:1707.09641. [Google Scholar] [CrossRef]
  168. Barratt, S. InterpNET: Neural Introspection for Interpretable Deep Learning. arXiv 2017, arXiv:1710.09511. [Google Scholar] [CrossRef]
  169. Chattopadhyay, A.; Manupriya, P.; Sarkar, A.; Balasubramanian, V.N. Neural Network Attributions: A Causal Perspective. In Proceedings of the 36th International Conference on Machine Learning, ICML 2019, Long Beach, CA, USA, 9–15 June 2019; pp. 1660–1676. [Google Scholar]
  170. Panigutti, C.; Perotti, A.; Pedreschi, D. Doctor XAI An Ontology-Based Approach to Black-Box Sequential Data Classification Explanations. In FAT* 2020—Proceedings of the 2020 Conference on Fairness, Accountability, and Transparency, Barcelona, Spain, 27–30 January 2020; Association for Computing Machinery: New York, NY, USA, 2020; pp. 629–639. [Google Scholar] [CrossRef]
  171. Kanamori, K.; Takagi, T.; Kobayashi, K.; Ike, Y.; Uemura, K.; Arimura, H. Ordered Counterfactual Explanation by Mixed-Integer Linear Optimization. Proc. AAAI Conf. Artif. Intell. 2021, 35, 11564–11574. [Google Scholar] [CrossRef]
  172. Kenny, E.M.; Keane, M.T. On Generating Plausible Counterfactual and Semi-Factual Explanations for Deep Learning. Proc. AAAI Conf. Artif. Intell. 2021, 35, 11575–11585. [Google Scholar] [CrossRef]
  173. Wu, T.; Ribeiro, M.T.; Heer, J.; Weld, D.S. Polyjuice: Generating Counterfactuals for Explaining, Evaluating, and Improving Models. In ACL-IJCNLP 2021—59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing, Proceedings of the Conference, Online, 1–6 August 2021; Association for Computational Linguistics: Stroudsburg, PA, USA, 2021; Volume 1, pp. 6707–6723. [Google Scholar] [CrossRef]
  174. Schleich, M.; Geng, Z.; Zhang, Y.; Suciu, D. GeCo: Quality Counterfactual Explanations in Real Time. Proc. VLDB Endow. 2021, 14, 1681–1693. [Google Scholar] [CrossRef]
  175. Fernández, R.R.; Martín de Diego, I.; Aceña, V.; Fernández-Isabel, A.; Moguerza, J.M. Random Forest Explainability Using Counterfactual Sets. Inf. Fusion 2020, 63, 196–207. [Google Scholar] [CrossRef]
  176. Wexler, J.; Pushkarna, M.; Bolukbasi, T.; Wattenberg, M.; Viegas, F.; Wilson, J. The What-If Tool: Interactive Probing of Machine Learning Models. IEEE Trans. Vis. Comput. Graph. 2020, 26, 56–65. [Google Scholar] [CrossRef] [PubMed]
  177. Ghazimatin, A.; Balalau, O.; Roy, R.S.; Weikum, G. Prince: Provider-Side Interpretability with Counterfactual Explanations in Recommender Systems. In WSDM 2020—Proceedings of the 13th International Conference on Web Search and Data Mining, Houston, TX, USA, 3–7 February 2020; Association for Computing Machinery: New York, NY, USA, 2020; pp. 196–204. [Google Scholar] [CrossRef]
  178. Aguilar-Palacios, C.; Munoz-Romero, S.; Rojo-Alvarez, J.L. Cold-Start Promotional Sales Forecasting through Gradient Boosted-Based Contrastive Explanations. IEEE Access 2020, 8, 137574–137586. [Google Scholar] [CrossRef]
  179. Wang, H.; Wang, Z.; Du, M.; Yang, F.; Zhang, Z.; Ding, S.; Mardziel, P.; Hu, X. Score-CAM: Score-Weighted Visual Explanations for Convolutional Neural Networks. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops, Seattle, WA, USA, 14–19 June 2020; pp. 111–119. [Google Scholar] [CrossRef]
  180. Amoukou, S.I.; Brunel, N.J.-B.; Salaün, T. The Shapley Value of Coalition of Variables Provides Better Explanations. arXiv 2021, arXiv:2103.13342. [Google Scholar]
  181. Mishra, S.; Sturm, B.L.; Dixon, S. Local Interpretable Model-Agnostic Explanations for Music Content Analysis. In Proceedings of the 18th International Society for Music Information Retrieval Conference, Suzhou, China, 23–27 October 2017; pp. 537–543. [Google Scholar] [CrossRef]
  182. Welling, S.H.; Refsgaard, H.H.F.; Brockhoff, P.B.; Clemmensen, L.H. Forest Floor Visualizations of Random Forests. arXiv 2016, arXiv:1605.09196. [Google Scholar] [CrossRef]
  183. Rosenbaum, L.; Hinselmann, G.; Jahn, A.; Zell, A. Interpreting Linear Support Vector Machine Models with Heat Map Molecule Coloring. J. Cheminform. 2011, 3, 11. [Google Scholar] [CrossRef]
  184. Akula, A.R.; Wang, S.; Zhu, S.C. CoCoX: Generating Conceptual and Counterfactual Explanations via Fault-Lines. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; AAAI Press: Palo Alto, CA, USA, 2020; Volume 34, pp. 2594–2601. [Google Scholar] [CrossRef]
  185. Bonifazi, G.; Cauteruccio, F.; Corradini, E.; Marchetti, M.; Terracina, G.; Ursino, D.; Virgili, L. A Model-Agnostic, Network Theory-Based Framework for Supporting XAI on Classifiers. Expert. Syst. Appl. 2024, 241, 122588. [Google Scholar] [CrossRef]
  186. Ali, A.; Schnake, T.; Eberle, O.; Montavon, G.; Müller, K.R.; Wolf, L. XAI for Transformers: Better Explanations through Conservative Propagation. Proc. Mach. Learn. Res. 2022, 162, 435–451. [Google Scholar]
  187. Moradi, M.; Samwald, M. Post-Hoc Explanation of Black-Box Classifiers Using Confident Itemsets. Expert. Syst. Appl. 2021, 165, 113941. [Google Scholar] [CrossRef]
  188. Bousselham, W.; Boggust, A.; Chaybouti, S.; Strobelt, H.; Kuehne, H. LeGrad: An Explainability Method for Vision Transformers via Feature Formation Sensitivity. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Paris, France, 1–6 October 2024. [Google Scholar]
  189. Amara, K.; Sevastjanova, R.; El-Assady, M. SyntaxShap: Syntax-Aware Explainability Method for Text Generation. In Proceedings of the Findings of the Association for Computational Linguistics: ACL 2024, Bangkok, Thailand, 11–16 August 2024; Association for Computational Linguistics: Kerrville, TX, USA, 2024; pp. 4551–4566. [Google Scholar] [CrossRef]
  190. Kumar, G.; Parimala, N. A Sensitivity Analysis on Weight Sum Method MCDM Approach for Product Recommendation. In Distributed Computing and Internet Technology. ICDCIT 2019; Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics); Springer: Cham, Switzerland, 2019; Volume 11319, pp. 185–193. [Google Scholar] [CrossRef]
  191. Kiela, D.; Firooz, H.; Mohan, A.; Goswami, V.; Singh, A.; Ringshia, P.; Testuggine, D. The Hateful Memes Challenge: Detecting Hate Speech in Multimodal Memes. Adv. Neural Inf. Process Syst. 2020, 33, 2611–2624. [Google Scholar]
  192. Statlog (German Credit Data)—UCI Machine Learning Repository. Available online: https://archive.ics.uci.edu/dataset/144/statlog+german+credit+data (accessed on 31 October 2025).
Figure 1. Basic taxonomy of XAI methods based on stage, scope, input data type and output format.
Figure 1. Basic taxonomy of XAI methods based on stage, scope, input data type and output format.
Make 07 00158 g001
Figure 2. Flowchart of the proposed tool.
Figure 2. Flowchart of the proposed tool.
Make 07 00158 g002
Figure 3. Constant ranking for XAI methods across all MCDM methods for the first use case.
Figure 3. Constant ranking for XAI methods across all MCDM methods for the first use case.
Make 07 00158 g003
Figure 4. Ranking for XAI methods across all MCDM methods for the second use case.
Figure 4. Ranking for XAI methods across all MCDM methods for the second use case.
Make 07 00158 g004
Figure 5. Position in rankings across all MCDM methods for XAI methods in the third use case.
Figure 5. Position in rankings across all MCDM methods for XAI methods in the third use case.
Make 07 00158 g005
Figure 6. Correlation matrix for the third use case.
Figure 6. Correlation matrix for the third use case.
Make 07 00158 g006
Figure 7. Rank change of methods due to weight change for VIKOR.
Figure 7. Rank change of methods due to weight change for VIKOR.
Make 07 00158 g007
Figure 8. Overall share of 1st rank wins in sensitive analysis for the third use case.
Figure 8. Overall share of 1st rank wins in sensitive analysis for the third use case.
Make 07 00158 g008
Table 1. Comparison of benchmarks and libraries aimed at evaluating XAI methods.
Table 1. Comparison of benchmarks and libraries aimed at evaluating XAI methods.
BenchmarkXAI MethodsMetricsData TypeSupported ModelsExtensible
M4 [63]LIMEFaithfulnessIMGResNetsYes
Integrated Gradient TXTMobileNets
SmoothGrad VGG
GradCAM ViT
Generic Attribution MAE-ViT-base
Bidirectional Explanations BERTs
DistilBERT
ERNIE-2.0-base
RoBERTa
Quantus [64]GradientShapFaithfulnessIMGNNYes
IntegratedGradientsRobustnessTAB
DeepLiftLocalisationTS
DeepLiftShapComplexity
InputXGradientRandomisation (Sensitivity)
SaliencyAxiomatic
Feature Ablation
Deconvolution
Feature Permutation
LIME
Kernel SHAP
LRP
Gradient
Occlusion
Layer GradCam
Guided GradCam
Layer Conductance
Layer Activation
Internal Influence
Layer GradientXActivation
Control Var. Sobel Filter
Control Var. Constant
Control Var. Random Uniform
Vanilla Gradients
Gradients Input
Occlusion Sensitivity
GradCAM
SmoothGrad
BEExAI [65]Feature AblationSensitivityTABLinear RegressionYes
LIMEInfidelity Logistic Regression
Shapley Value SamplingComprehensiveness Random Forest
Kernel SHAPSufficiency Decision Tree
Integrated GradientsFaithfulness Correlation Gradient Boosting
SaliencyAUC-TP XGBoost
DeepLiftMonotonicity Dense Neural Network
InputXGradientComplexity
Sparseness
FUNCXAI-11 [66]Not definedRepresentativenessTABNot definedYes
StructureIMG
SelectivityTXT
Contrastivity
Interactivity
Fidelity
Faithfulness
Truthfulness
Stability
(Un)certainty
Speed
XAI UNITS [67]DeepLIFTInfidelityTABMLPYes
Shapley Value SamplingSensitivityIMGCNN
InputXGradientMSETXTViT
IntegratedGradientsMask Error LLM
LIMEMask Proportion Image
Kernel SHAPMask Proportion Text
Feature Ablation
Gradient SHAP
DeepLIFT SHAP
Saliency
Deconvolution
Guided Backpropagation
Guided GradCAM
Feature Permutation
Occlusion
Compare-xAI [68]Exact Shapley ValuesComprehensibility:TABNot definedYes
Kernel SHAP• Fidelity
LIME• Fragility
MAPLE• Stability
Partition• Simplicity
Permutation• Stress tests
Permutation PartitionPortability
SaabasAverage execution time
SAGE
SHAP Interaction
Shapley Taylor Interaction
Tree SHAP
Tree SHAP Approximation
XAIB [69]ConstantModel Randomization Check (MRC)TABSVCYes
LIMESmall Noise Check (SNC) MLP
SHAPLabel Difference (LD) KNN
KNNDifferent Methods Agreement (DMA)
Sparsity (SP)
Covariate Regularity (CVR)
Target Discriminativeness (TGD)
Same Class Check (SCC)
XAI-Bench [70]LIMEFaithfulnessTABLinear RegressionNo
SHAPMonotonicity Decision Tree
MAPLEROAR MLP
SHAPRGT-Shapley
BF-SHAPInfidelity
L2X
BreakDown
OpenXAI [71]LIMEFaithfulness:TABNNYes
SHAP• Feature Agreement (FA)Logistic regression
Vanilla Gradients• Rank Agreement (RA)
InputXGradient• Sign Agreement (SA)
SmoothGrad• Signed Rank Agreement (SRA)
Integrated Gradients• Rank Correlation (RC)
• Pairwise Rank Agreement (PRA)
• Prediction Gap on Important feature perturbation (PGI)
• Prediction Gap on Unimportant feature perturbation (PGU)
Stability:
• Relative Input Stability (RIS)
• Relative Representation Stability (RRS)
• Relative Output Stability (ROS)
Fairness
Saliency bench [72]Grad CAMmean Intersection over Union (mIoU)IMGCNNYes
GradCAM++Pointing Game (PG) ViT
Integrated GradientsInsertion (iAUC)
InputXGradientPrecision
OcclusionRecall
RISE
OpenHEXAI [73]LIMEAccuracyTABNNPartially
SHAPF1 Logistic regression
Vanilla GradientsAVG Time
InputXGradientOver-Reliance
SmoothGradUnder-Reliance
Integrated GradientsAverage Absolute Odds Difference (AAOD)
Equal Opportunity Difference (EOD)
TAB: tabular, TXT: text, IMG: image, TS: time series, NN: neural network, CNN: convolutional neural network, Vit: vision transformer, LLM: large language model.
Table 2. List of XAI methods.
Table 2. List of XAI methods.
MethodYearPortabilityModelScopeData TypeProblemOutput
LIME [18]2016MA LTAB, IMG, TXTC, RN, V
SHAP [19]2017MA L, GTAB, IMG, TXTC, RN, V
Shapley values [91]2014MA LTABC, RN
LRP [16]2015MSDNNLIMG, TXTCN, V
Saliency Maps [21]2013MSDNNLIMG, TXTCV
Grad-CAM [15]2019MSCNNLIMGCV
IntGrad [22]2017MSDNNLIMG, TXTCN, V
Anchors [20]2018MA LTAB, TXTCRU
DeepLIFT [17]2017MSDNNLIMG, TXTCN, V
Influence Functions [92]2017MA LIMGCN, V
TCAV [93]2017MA GIMGCN
ICE [23]2015MA GTABC, RV
DTD [94]2017MSDNNLIMGCN, V
DeconvNet [95]2013MSCNNLIMGCN, V
SmoothGrad [96]2017MSDNNLIMGCN, V
PDP [24]2001MA GTABC, RV
PatternAttribution [97]2017MSDNNLIMGCN, V
Guided BackProp [98]2014MSDNNLIMGCN, V
Meaningful Perturbation [99]2017MSDNNLIMGCN, V
PatternNet [97]2017MSDNNLIMGCN, V
Show, Attend and Tell [100]2015MSCNNLIMGCT, V
Activation Maximization [101]2010MSDNNLIMGCV
CAM [102]2015MSCNNLIMGCV
Rationales [103]2016MSNLP modelLTXTCT
LORE [104]2018MA LTABCRU
PDA [105]2017MSDNNLIMGCV
RISE [106]2018MA LIMGCN, V
CEM [107]2018MA LIMG, TAB, TXT, GRH, TS, VIDCN, T
Guided Proto [108]2019MA LIMG, TABCV, T
DICE [109]2020MA LTABCN, T
ACE [110]2019MA GIMGCN, T
ConceptSHAP [111]2020MA GIMGCN, V
Soft DT [112]2017MSDecision TreeGIMGCRU
ALE [25]2016MA GTABC, RV
CIU [113]1995MA L, GIMG, TABC, RN
Regularisation [114]2015MS LIMGCN
GFA [115]2016MA GTABC, RN, V
GAN Dissection [116]2018MSGANLIMGCV
FACE [117]2019MA LIMG, TAB, TXT, GRH, TS, VIDCN, T
NAM [118]2020MS LTABC, RV
RuleMatrix [119]2018MA GTABCRU, V
Grad-CAM++ [120]2018MSCNNLIMGCV
L2X [121]2018MA LIMG, TXTCN
MAPLE [122]2018MA LTABC, RN, RU
PIMP [123]2010MA GTABCN
BreakDown [124]2018MA LTABC, RN, T
ABELE [125]2020MA LIMGCN, T
MOC [126]2020MA LTABC, RN, T
SEDC [127]2014MA LTXTCV, RU
LIVE [124]2018MA LTABC, RV
GraphLIME [128]2020MA LGRHCN, V
GLocalX [129]2020MA L, GTABCRU
Privacy-Preserving Explanations [130]2020MA LTABCN
PI, ICI [131]2018MA GTABC, RN
QII [132]2016MA GTABCN
FOCUS [133]2021MA GTABC, RN
EBCF [134]2020MSRecommender SystemsLTABCV, T
DCE [135]2019MA LTABCN, T
Actionable Recourse [136]2019MA LTABCN, T
DACE [137]2020MA LTABCN, T
MACE [138]2020MA LTABC, RN, T
C-CHVAE [139]2020MA LTABCN, T
SYNTH [140]2020MA LTABCV
DECE [141]2020MA L, GTABCN, T
ALG-REC [142]2020MA LTABCN, T
Growing Spheres [143]2017MA LTABCN, T
CERTIFAI [144]2020MA LTAB, IMG, TXTC, RN
ViCE [145]2020MA LTXTCV, T
MC-BRP [146]2020MA LTABC, RRU
LIME-C/SHAP-C [147]2020MA LTABCN, V
CLEAR [148]2019MA LTABC, RT, RU
GNNExplainer [149]2019MA LGRHCN, V
Shapley Flow [150]2020MA L, GTABC, RN, V
FBT [151]2020MSDT EnsamblesGTABCN
CHIRPS [152]2020MSRandom ForestLTABCN, V
LoRMIka [153]2019MA LTABCRU, V
Color-based monogram [154]2016MA GTABCV
SR map [155]2019MSCNNLIMGCRU, V
DLIME [156]2019MA LIMG, TAB, TXT, GRH, TS, VIDC, RN
LioNets [157]2019MSIntrinsically Interpretable DNNLTXTCRU
SkopeR 2020MA L, GTABCRU
XRAI [158]2019MSCNNLIMGCV
XSPELLS [159]2020MA LTXTCRU
exBERT [160]2019MSTransformer modelLTXTCV
Slot Activation Vectors [161]2018MSSlot Attention ModelLTXTCN
Peak Response [162]2018MSCNNLIMGCV
Autofocus-Layer [163]2018MSCNNLIMGCN
NNKX [164]2017MSDNNGTABCN, V
Hypothesis Testing [165]2019MA LIMG, TXTCN
GAM [166]2019MA GIMG, TABCV
Important Neurons and Patches [167]2017MSCNNGIMGCN, V
InterpNET [168]2017MA LIMGCV
ACE [169]2019MSDNNGIMG, TAB, TXT, GRH, TS, VIDCN, T
DoctorXAI [170]2020MA LTABCN, T
ORDCE [171]2021MA LTABCN, T
PIECE [172]2021MSCNNLIMGCV
POLYJUICE [173]2021MA LTXTC, RT
GeCO [174]2021MA LTABC, RN, T
RF-OCSE [175]2020MA LTABC, RRU
What-If [176]2020MA GTAB, IMG, TXTC, RV, N
PRINCE [177]2020MA LGRHCV, T
Gradient Boosted CFs [178]2020MSDT EnsemblesLTABCN, T
Score-CAM [179]2019MSCNNLIMGCV
ACV [180]2022MA LTABC, RN
SoundLIME [181]2017MA LTSCN, V
Forest Floor [182]2016MSDT EnsemblesGTABC, RV
ChemHeatmap [183]2011MA LGRHCV
Cocox [184]2020MA LIMGCN, T
NAFER [185]2024MA L, GTABCN
Improved LRP [186]2022MSDNNLIMG, TXT, GRHCN, V
CIE [187]2021MA LTXT, TABCN
LeGrad [188]2024MA LIMGCV
SyntaxShap [189]2024MSNLP modelLTABCN, T
MA: model-agnostic, MS: model-specific, L: local, G: global, TAB: tabular, TXT: text, IMG: image, GRH: graph, TS: time series, VID: video, C: classification, R: regression, N: numerical, T: textual, V: visual, RU: rules.
Table 3. Average results for subjective metrics of evaluation of XAI methods in the field of hate speech obtained from questionnaires based on a Likert scale (1–7).
Table 3. Average results for subjective metrics of evaluation of XAI methods in the field of hate speech obtained from questionnaires based on a Likert scale (1–7).
MethodUnderstandabilityUsefulnessTrustworthinessInformativenessSatisfaction
LIME5.465.414.995.335.47
SHAP5.465.394.975.235.60
Grad-CAM4.834.764.494.804.81
Table 4. Saaty’s matrix created by researchers. Inconsistency index is 0.068, which is acceptable.
Table 4. Saaty’s matrix created by researchers. Inconsistency index is 0.068, which is acceptable.
UnUsTrIS
Un131/333
Us1/311/333
Tr33175
I1/31/31/711/3
S1/31/31/531
Un: understandability, Us: usefulness, Tr: trustworthiness, I: informativeness, S: satisfaction.
Table 5. Preferences of LIME, SHAP, and Grad-CAM according to all MCDM methods for the first use case.
Table 5. Preferences of LIME, SHAP, and Grad-CAM according to all MCDM methods for the first use case.
MethodLIMESHAPGrad-CAM
ARAS0.99790.99660.8897
CODAS0.16840.1629−0.3313
EDAS10.98110
MABAC0.41110.3919−0.5745
MARCOS0.70330.70230.627
PROMETHEE II0.79560.2044−1
TOPSIS0.97480.96110
VIKOR00.01471
WASPAS0.9980.99650.8897
WSM0.3460.34550.3085
Table 6. Results of 11 metrics for LIME, Kernel SHAP, and Tree SHAP methods [66].
Table 6. Results of 11 metrics for LIME, Kernel SHAP, and Tree SHAP methods [66].
MethodReStSeCoInFiFaTrStaUncSp
LIME117.711.130.71.420.423
Kernel SHAP10611.230.62.521.222
Tree SHAP107.311.2322.531.224
Re: representativeness, St: structure, Se: selectivity, Co: contrastivity, In: interactivity, Fi: fidelity, Fa: faithfulness, Tr: truthfulness, Sta: stability, Unc: uncertainty, Sp: speed.
Table 7. Weights of criteria according to the CRITIC method.
Table 7. Weights of criteria according to the CRITIC method.
CriterionWeight
Representativeness0.2266
Structure0.1397
Selectivity0.0000
Contrastivity0.1267
Interactivity0.0000
Fidelity0.0825
Faithfulness0.1267
Truthfulness0.0849
Stability0.1267
Uncertainty0.0000
Speed0.0862
Table 8. Preferences of LIME, Kernel SHAP, and Tree SHAP according to all MCDM methods.
Table 8. Preferences of LIME, Kernel SHAP, and Tree SHAP according to all MCDM methods.
MethodLIMEKernel SHAPTree SHAP
ARAS0.72510.78550.9674
CODAS−0.4043−0.08430.4886
EDAS0.09890.31290.9726
MABAC−0.0239−0.05910.3013
MARCOS0.59280.63910.767
PROMETHEE II−0.0563−0.2740.3303
TOPSIS0.51780.41940.5542
VIKOR0.451210.5
WASPAS0.71840.78380.9644
WSM0.29080.31540.3938
Table 9. Criterion weights according to the CRITIC method for the German Credit Dataset.
Table 9. Criterion weights according to the CRITIC method for the German Credit Dataset.
CriterionWeight
Accuracy0.105
F10.098
AVG Time0.0441
Over-Reliance0.0378
Under-Reliance0.0551
Average Absolute Odds Difference (AAOD)0.0306
Equal Opportunity Difference (EOD)0.0602
Q10.0305
Q20.0288
Q30.0301
Q40.0289
Q50.0419
Q60.0265
Q70.0305
Q80.0257
Q90.0378
Q100.029
Q110.0338
Q120.0268
Q130.0333
Q140.085
Q150.0495
Q160.0309
Table 10. Preferences of LIME, SHAP, SmoothGrad, and Integrated Gradients according to all MCDM methods.
Table 10. Preferences of LIME, SHAP, SmoothGrad, and Integrated Gradients according to all MCDM methods.
MethodLIMESHAPSmoothGradIntegrated Gradients
ARAS0.88420.90550.8650.8118
EDAS0.60290.76550.51610.149
CODAS−0.01380.13280.0077−0.1267
MABAC0.16190.20.0151−0.1644
PROMETHEE II0.26160.2223−0.0601−0.4238
VIKOR0.31320.03170.75380.5
TOPSIS0.53330.68390.4120.5099
WASPAS0.88830.89930.86570.8152
MARCOS0.68030.69390.66490.6264
WSM0.2550.26160.24950.2339
Table 11. Percentage of 1st Rank across sensitivity analysis experiments.
Table 11. Percentage of 1st Rank across sensitivity analysis experiments.
MethodLIMESHAPSmoothGradIntegrated Gradients
ARAS85.00%15.00%0.00%0.00%
CODAS65.00%35.00%0.00%0.00%
EDAS80.00%20.00%0.00%0.00%
MABAC95.00%5.00%0.00%0.00%
MARCOS90.00%10.00%0.00%0.00%
PROMETHEE II95.00%5.00%0.00%0.00%
TOPSIS85.00%15.00%0.00%0.00%
VIKOR45.00%20.00%35.00%0.00%
WASPAS90.00%10.00%0.00%0.00%
WSM85.00%15.00%0.00%0.00%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Matejová, M.; Paralič, J. A Multi-Criteria Decision-Making Approach for the Selection of Explainable AI Methods. Mach. Learn. Knowl. Extr. 2025, 7, 158. https://doi.org/10.3390/make7040158

AMA Style

Matejová M, Paralič J. A Multi-Criteria Decision-Making Approach for the Selection of Explainable AI Methods. Machine Learning and Knowledge Extraction. 2025; 7(4):158. https://doi.org/10.3390/make7040158

Chicago/Turabian Style

Matejová, Miroslava, and Ján Paralič. 2025. "A Multi-Criteria Decision-Making Approach for the Selection of Explainable AI Methods" Machine Learning and Knowledge Extraction 7, no. 4: 158. https://doi.org/10.3390/make7040158

APA Style

Matejová, M., & Paralič, J. (2025). A Multi-Criteria Decision-Making Approach for the Selection of Explainable AI Methods. Machine Learning and Knowledge Extraction, 7(4), 158. https://doi.org/10.3390/make7040158

Article Metrics

Back to TopTop