Next Article in Journal
Structure Approximation-Based Preconditioning for Solving Tempered Fractional Diffusion Equations
Previous Article in Journal
A Semi-Supervised-Learning-Aided Explainable Belief Rule-Based Approach to Predict the Energy Consumption of Buildings
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Nomological Deductive Reasoning for Trustworthy, Human-Readable, and Actionable AI Outputs

by
Gedeon Hakizimana
and
Agapito Ledezma Espino
*
Department of Computer Science & Engineering, Universidad Carlos III de Madrid, 28911 Leganes, Spain
*
Author to whom correspondence should be addressed.
Algorithms 2025, 18(6), 306; https://doi.org/10.3390/a18060306
Submission received: 21 March 2025 / Revised: 8 May 2025 / Accepted: 20 May 2025 / Published: 23 May 2025
(This article belongs to the Section Algorithms for Multidisciplinary Applications)

Abstract

:
The lack of transparency in many AI systems continues to hinder their adoption in critical domains such as healthcare, finance, and autonomous systems. While recent explainable AI (XAI) methods—particularly those leveraging large language models—have enhanced output readability, they often lack traceable and verifiable reasoning that is aligned with domain-specific logic. This paper presents Nomological Deductive Reasoning (NDR), supported by Nomological Deductive Knowledge Representation (NDKR), as a framework aimed at improving the transparency and auditability of AI decisions through the integration of formal logic and structured domain knowledge. NDR enables the generation of causal, rule-based explanations by validating statistical predictions against symbolic domain constraints. The framework is evaluated on a credit-risk classification task using the Statlog (German Credit Data) dataset, demonstrating that NDR can produce coherent and interpretable explanations consistent with expert-defined logic. While primarily focused on technical integration and deductive validation, the approach lays a foundation for more transparent and norm-compliant AI systems. This work contributes to the growing formalization of XAI by aligning statistical inference with symbolic reasoning, offering a pathway toward more interpretable and verifiable AI decision-making processes.

Graphical Abstract

1. Introduction

Artificial Intelligence (AI) is increasingly being deployed in high-stakes domains such as healthcare, finance, autonomous systems, and law, where decisions have significant real-world consequences. Despite their notable performance, many modern AI models—especially deep-learning-based systems—operate as “black boxes” [1,2], lacking the transparency and traceable logic necessary for human understanding [3,4,5,6]. This opacity undermines user trust, accountability, and interpretability.
Explainable AI (XAI) has thus emerged as a critical area of research, driven by the need for AI systems to generate understandable and justifiable decisions. In domains like healthcare and finance, explainability is essential for professionals to evaluate AI-generated outputs [7,8,9]. Autonomous vehicles face similar scrutiny, where the lack of interpretability hampers public acceptance and regulation [10]. This concern has been institutionalized in policy frameworks such as the EU’s GDPR [11], the OECD’s AI principles [12], and the proposed AI Act [13], with comparable guidelines developed across Asia [14,15,16] and Africa [17,18,19].
Academic progress in XAI has yielded a variety of methods, including model-agnostic techniques like LIME [20] and SHAP [21], and more structured approaches like Anchors [22] and prototype-based explanations [23]. More recently, large language models (LLMs) have been employed to generate human-readable explanations. However, many of these solutions still fall short in providing causally grounded, verifiable reasoning [24], limiting their value in critical contexts.
Rule-based expert systems have historically addressed this gap by embedding human-like reasoning through symbolic inference. Recent hybrid methods have revisited this paradigm, combining statistical learning with structured logic—for instance, via the Shapley–Lorenz approach [25] or confident itemsets [26]. These efforts emphasize how rule-based models can support traceable, actionable decision-making.
Yet, challenges persist. Existing XAI techniques may oversimplify, omit critical details, or overwhelm users with complexity. Furthermore, there remains no consensus on what constitutes a “good” explanation, with effectiveness varying by context [27,28]. Ultimately, XAI must align with human cognitive expectations to foster trust and usability [29,30].
To address these gaps, this paper introduces a novel framework grounded in Nomological Deductive Reasoning (NDR) and built upon Nomological Deductive Knowledge Representation (NDKR). By integrating domain-specific rules with formal logic, the approach aims to produce transparent, law-based explanations that align with normative and epistemic standards.
The rest of the paper is organized as follows: Section 2 outlines the methodology, including a literature review and framework development. Section 3 analyzes existing XAI techniques and their limitations. Section 4 identifies gaps in the explanation clarity and trust. Section 5 details the proposed NDR framework and presents experimental validations, and Section 6 concludes with key contributions and future directions.

2. Methodology

This study combines a systematic literature review with empirical evaluation to develop and assess the Nomological Deductive Reasoning (NDR) framework for explainable Artificial Intelligence (XAI). The approach addresses both theoretical gaps in explainability research and practical challenges in generating trustworthy, traceable AI decisions.

2.1. Conceptual Framework Development

The conceptual design of the NDR framework was informed by a structured literature review of XAI methods published between 2010 and 2024, conducted in accordance with PRISMA guidelines. The review focused on explanation formats, logical structure, and epistemic soundness, emphasizing the importance of producing explanations that are not only interpretable but also grounded in valid reasoning [31,32]. The analysis revealed a prevailing reliance on post hoc and correlation-based interpretability techniques, which often lack formal causal justification. In response, the NDR framework was developed to integrate statistical predictions with symbolic deductive logic, guided by Hempel’s deductive–nomological (D-N) model. In this approach, AI outputs must be derivable from a structured knowledge base of domain-specific rules and antecedent conditions, embedding explainability directly into the reasoning process.

2.2. Architecture and Workflow

The proposed NDR framework comprises five sequential layers: (1) knowledge representation, (2) the AI model, (3) reasoning and deductive inference, (4) explanation generation, and (5) user interaction and trust evaluation. Each layer contributes to generating rule-consistent, human-readable justifications grounded in domain knowledge.

2.3. Experimental Setup

To evaluate the framework, experiments were conducted using the Statlog (German Credit Data) dataset [33], which includes 1000 samples with 20 features related to creditworthiness. Data preprocessing involved the normalization of numerical attributes and the encoding of categorical variables. The dataset was split into training (70%) and test (30%) sets, with a fixed random seed (42) to ensure reproducibility. A Random Forest classifier [34] was trained to generate initial credit-risk predictions. This model was chosen for its robustness and interpretability in structured financial data contexts.

2.4. Deductive Validation and Explanation Generation

Each prediction produced by the classifier was evaluated by the NDR deductive engine. This component queried a formalized financial knowledge base containing symbolic rules such as: “Loan duration > 12 months indicates credit stability” or “Multiple debtors suggest increased financial risk”. The engine determined whether the prediction followed logically from the rules and, if so, generated a corresponding explanation.

2.5. Evaluation Metrics

The framework was assessed along two dimensions: predictive accuracy and interpretability. Specifically, the following parameters were assessed:
  • Accuracy: Classification correctness of the machine learning model.
  • Rule Consistency: Agreement of predictions with symbolic knowledge.
  • Rule Coverage: Percentage of predictions for which valid rule-based explanations were generated.
  • Mismatch Penalty: Count of predictions inconsistent with the knowledge base.
Figure 1 summarizes the methodological flow for developing and evaluating the NDR framework.

2.6. Reproducibility and Future Work

To support transparency and reproducibility, all implementation artifacts—including the source code and the structured knowledge base—are available via a public GitHub version 8 repository. Future work includes a planned user study to assess explanation clarity, trustworthiness, and usability in real-world decision-making scenarios, particularly in credit assessment and clinical diagnostics.

3. Current XAI State of the Art

In this research, the focus was put on the formats of explanations generated by current XAI methods. This angle was chosen for several reasons. First, the way explanations are presented plays a crucial role in how effectively users understand, trust, and act on AI decisions. Second, by examining different explanation formats, bridging the gap between complex AI systems and users with varying technical background becomes easier, ensuring that explanations are both accurate and accessible. Third, as the content and format of explanation are both important in the explanation process, placing the formats of the explanators at the center of the research leads to a better understanding of how the user-interaction cognitive load and decision-making can be enhanced, ultimately improving the overall usability and trustworthiness of AI models across diverse application domains. Therefore, in the context of an exponential growth in the explainable Artificial Intelligence (XAI) field, with more than 200 different methods of explainability on the one hand and a continuously growing need for human-readable explanations on the other hand, the format of various explanations was purposely explored in this paper to better understand why there is still a gap in producing human-readable explanations. XAI methods vary widely in how they present explanations and in the types of AI systems they aim to explain [35,36,37], and it is argued that the understanding of explanation format varieties is key to grasping the current landscape of XAI research. Before exposing the core of this research paper, the following section reviews the explanation formats as per the current state of the art in XAI as broadly classified by various scholars, including Vilone and Longo [38], whose classification appears to be the most comprehensive and structured, with five main types of explanators, namely visual, textual, rule-based, numerical/graphical, and mixed explanations.

3.1. Explanator Formats

In the context of explainable Artificial Intelligence (XAI), explanators refer to the various formats or mechanisms through which explanations of model decisions or behaviors are conveyed to users [28,39]. These explanations aim to make complex AI models, particularly black-box models such as deep learning networks, more understandable and interpretable. The nature of these explanators significantly influence how effectively users can comprehend, trust, and act upon the AI’s outputs [31,40,41]. As per the current XAI state of the art, there is no one-size-fits-all type of explanation, since it is believed that different explanatory formats serve distinct purposes, each tailored to different stakeholders, such as developers, domain experts, or end-users. Next, we review the current formats of explanations generated by different XAI methods.

3.1.1. Visual Explanations

In the XAI field, visual explanations are a type of detail generated by AI algorithms that use visual cues to interpret the reasoning behind the machine learning models, particularly in image processing. This type of explanation is considered particularly useful when interpreting image-based models or in computer-vision tasks. Visual explanations aim to highlight the parts of the input (such as an image or video) that are most relevant to a model’s decision-making process [42]. These explanations are typically in the form of heatmaps, saliency maps, or attention maps that visually guide the user’s understanding of the AI model’s focus areas, as illustrated in Figure 2.
The key methods used to generate visual explanations include the following:
Grad-CAM (Gradient-weighted Class Activation Mapping) [44]: This is one of the most popular methods found in the literature for generating visual explanations. Grad-CAM generates heatmaps by computing the gradient of the output category with respect to the feature maps of the last convolutional layer. This heatmap helps visualize which regions of an image contributed most to the model’s prediction.
Saliency Maps [45]: This method computes the gradient of the predicted class with respect to the input image and then highlights the areas where small changes in the input would most influence the prediction. It is particularly useful for CNNs in image classification. Practically, saliency maps are the images that highlight the region of interest in a given computer-vision model, where a heatmap is generally superimposed over the input image to highlight the pixels of the image that provided the most important contributions in the prediction task. Figure 3 is an example of such images.
Activation Maximization [46]: This technique seeks to visualize the features that maximally activate specific neurons in the convolutional network. By generating inputs that lead to high activation in specific neurons, it provides insight into the types of patterns that a deep learning model is sensitive to.
DeepLIFT (Deep Learning Important FeaTures) [47]: Unlike Grad-CAM, DeepLIFT computes the contribution of each input feature by comparing the activations to a reference activation, offering a clearer breakdown of the model’s reasoning.
Other XAI methods used to generate visual explanations include, but are not limited to, Layer-wise Relevance Propagation (LRP) [48], Integrated Gradients (IG) [49], Guided Back propagation [50], RISE [51], Class-Enhanced Attentive Response (CLEAR) [52], and Vision Transformer (ViT) [53].

3.1.2. Textual Explanations

Textual explanations in Artificial Intelligence refer to sentences presented in natural language, which try to provide human-readable narratives of the model’s decision-making process. This aspect makes them suitable for non-technical stakeholders, as the explanations describe, at a certain level of natural language, why a model made a particular prediction or decision. Technically, the approach provides human users of AI systems with clues about the process behind the decision of any ML model, such as the labels assigned to image parts in an image classification task [54]. Table 1 is an example chosen from the literature of the generation of textual explanations.
The key methods used to generate textual explanations include the following:
LIME (Local Interpretable Model-Agnostic Explanations) [20]: Generating basically mixed explanations, LIME generates local surrogate models that approximate the decision boundary of a black-box model in a specific region of interest. It then provides a text in a natural-language format, aiming to describe the important features in that region. This makes it easier for humans to understand why the model made a particular decision.
SHAP (Shapley Additive Explanations) [21]: Although this method generates more than just texts, as will be shown later, SHAP provides a game-theoretic approach to explain the output of any machine learning model. By computing Shapley values, the method determines the contribution of each feature to a model’s prediction and provides textual explanations that describe the impact of each feature.
Various other methods for generating textual explanations exist. These include, but are not limited to, Rationales [58], InterpNet [59], Relevance Discriminative Loss [60], Most-Weighted-Path, Most-Weighted-Combination and Maximum-Frequency-Difference [61], and Neural-Symbolic Integration [62].

3.1.3. Rule-Based Explanations

Rule-based explanations are a popular type of explanation technique in data mining, machine learning and Artificial Intelligence by which regularities in data are identified through various rule-extraction techniques. These regularities are mostly presented in the form of the IF-THEN rule, sometimes with AND/OR operators, as illustrated in Figure 4, which is an example decision list that was created using the Titanic data set available in R for the task of predicting whether the passenger survived the Titanic tragedy based on his or her features, such as age (adult or child), gender (male or female), passenger class (first, second, third, or crew) [63]. The “if” statements express a set of features comparable to the model’s inputs, and the “then” statements correspond to the predicted outcome corresponding to the conditions stated by “if”.
Among the current XAI methods, this approach provides high interpretability of black boxes [37,64] even if some improvements can be added if analyzed from the holistic perspective of the logic and theory of explanation. The key methods used to provide such explanations are the following:
Decision Trees [65]: One of the oldest and most intuitive methods for model explanation. Decision trees split input data into various branches based on specific features, making them highly interpretable. These trees provide explicit rules that explain how a decision is made.
Anchors [66]: Anchors generate simple, understandable decision rules that are highly accurate in specific regions of the input space. These rules are based on the important features that “anchor” the decision in a given region of the input space.
Scalable Rule Lists and Rule Sets [67,68,69]: These methods extract human-readable decision rules directly from a model. They are often used with ensemble methods like decision forests or other machine learning algorithms that rely on multiple decision-making processes.
Various other methods such as Discriminative Patterns [70], MYCIN [71], Fuzzy Inference Systems [72], Automated Reasoning [73], Genetic Rule Extraction (G-REX) [74], Global to Local eXplainer (GLocalX) [75], Mimic Rule Explanation (MRE) [76] and many others exist, as detailed in [37]; these have the capacity to generate explanatory rules of the decision made by the model, although the form may be in some other type than IF…THEN.

3.1.4. Numerical/Graphical Explanations

As indicated by their name, numerical or graphical explanations provide a quantitative understanding of the model’s predictions or decisions. These explanations often rely on numbers, charts, or statistical representations to help users understand the model’s behavior. The key methods in this category include:
Partial Dependence Plots (PDPs) [77,78]: PDPs provide a graphical representation of the relationship between a feature and the predicted outcome while holding other features constant. This method helps to visualize the effect of a feature on the model’s prediction.
Feature Importance Scores [79]: Feature importance methods rank the features based on their contribution to the prediction. For example, Permutation Feature Importance measures the change in the model’s performance when a feature’s values are shuffled, providing insights into how much that feature impacts the model.
SHAP Values [80]: SHAP values provide a numerical explanation by attributing each feature’s contribution to the final prediction. These values can be visualized through force plots, offering a combination of both numerical and graphical insights into the model’s decision-making process.
ICE (Individual Conditional Expectation) Plots [81]: These plots illustrate how the model’s prediction changes when a particular feature varies while others remain fixed, offering a detailed visualization of the impact of specific features.
Concept Activation Vectors (TCAVs) [82]: Testing with Concept Activation Vectors is an explanation method used to interpret neural networks where the language of explanation is expanded from that of input features to include human-friendly concepts for greater intelligibility.
XAI methods that provide numerical or graphical explanations are abundant, including, but not limited to, Distill-and-Compare [83], Feature Contribution [84,85], Gradient Feature Auditing (GFA) [86], Shapley–Lorenz–Zonoid Decomposition [25], Probes [87], and Contextual Importance and Utility [88]. Illustrations of numerical and graphical explanations can be found in [81].

3.1.5. Mixed (Multiformat) Explanations

Mixed explanations combine two or more of the aforementioned formats in a hybrid approach aimed at providing richer, multi-dimensional insights into the model’s behavior by combining the strengths of various formats, as illustrated in Figure 5.
The key methods in this class include, but are not limited to, the following:
LIME and SHAP: Both methods are inherently model-agnostic and allow for explanations that integrate numerical/graphical and textual components. Specifically, LIME explains a model’s decision by approximating it with a simple, interpretable surrogate model, which can be accompanied by a textual explanation of the prediction along with graphical visualizations of feature importance.
Interactive Dashboards: Some systems such as L2X (Learning to Explain) [90], Partial Dependence Plots (PDP) [75], and DeepLIFT (Deep Learning Important FeaTures) [46] use interactive visualizations that combine graphical components, like feature importance graphs, with textual summaries or natural-language explanations. These hybrid methods are increasingly being used in domains like healthcare, where practitioners require a combination of quantitative insights and verbal explanations to make informed decisions.
Counterfactual Explanations [91]: These explanations involve presenting a user with an alternative scenario, i.e., what would need to change in the input to alter the prediction. These can be accompanied by both textual and graphical explanations showing the differences between the original and counterfactual instances.
There exist other XAI methods in this category, including, but not limited to, Image Caption Generation with Attention Mechanism [92], Bayesian Teaching [93], Neighborhood-based Explanations [94], ExplAIner [95], Sequential Bayesian Quadrature (SBQ) [96], and Pertinent Negatives [97].

3.2. Shortfall in Conveying Structured Human-Readable Explanations That Make Transparent the Reasoning Process

Despite the emergence of over 200 XAI techniques and various explanation formats, current methods often fail to deliver cognitively accessible and structured explanations that clearly reveal the model’s reasoning process. This shortfall undermines user trust and hampers their adoption in critical fields such as healthcare, finance, and autonomous systems [98,99,100].

3.2.1. Lack of Contextual Relevance in Visual Explanations

Visual tools like saliency maps, Grad-CAM, and activation maximization help localize influential input regions, yet they often lack contextual grounding. For example, saliency maps may highlight areas without indicating why these features matter, especially when domain-specific knowledge is required to interpret them [101,102]. Techniques like SmoothGrad and Guided Grad-CAM improve the visual clarity but still lack semantic context integration [42]. Consequently, while useful for model debugging, these methods offer limited support for end-users needing actionable insights. As seen in medical imaging applications (e.g., pneumonia/COVID-19 diagnosis), such visual outputs can be unintelligible without prior expertise, limiting their transparency and auditability.

3.2.2. Ambiguity and Imprecision in Textual Explanations

Textual explanation methods like LIME, SHAP, and Anchors aim for interpretability but often oversimplify complex models or use jargon inaccessible to non-experts [103]. LIME, for instance, approximates black-box behavior with local linear models, potentially misrepresenting global reasoning. SHAP provides feature importance grounded in game theory yet lacks insight into why such features contribute, especially when multiple interactions are involved [26]. These methods often fail to communicate the reasoning process, leaving users unable to trace or verify the model’s logic, which critically undermines trust.

3.2.3. Limitations of Rule-Based Explanations

Rule-based approaches—like decision trees, anchors, and model-agnostic logic rules—are intuitive but struggle with scalability in high-dimensional contexts [69]. As complexity increases, rule sets become large and fragmented, overwhelming users and obscuring global model behavior. Local rule-based methods like Anchors, while interpretable in isolation, often lack generalizability and fail to provide a holistic view of model reasoning [104,105].

3.2.4. Overload from Numerical and Graphical Explanations

Quantitative methods—such as feature-importance rankings, PDPs, ICE plots, and SHAP visualizations—effectively quantify feature contributions but often overwhelm users with data in high-dimensional settings [106]. Graphs like SHAP summary plots demand mathematical literacy to interpret, excluding non-experts. For instance, interpreting Figure 4 requires an understanding of symbolic data on 2D plots, which restricts accessibility. Without simplified, domain-aligned presentations, these tools hinder rather than support comprehension, especially where interpretability and accountability are essential.

3.2.5. Mixed-Explanation Methods and Cognitive Overload

Multiformat approaches combining more than one format, such as textual, visual, graphical and rule-based explanations, offer comprehensive insights but often at the cost of usability. Users must navigate multiple formats, risking redundancy, inconsistency, and cognitive overload. The added complexity can obscure rather than clarify the model’s behavior, raising questions about the necessity and effectiveness of combining multiple explanation modalities.

3.2.6. Confabulation and Lack of a Formal Logical Structure in LLM-Based Explanations

Recent efforts to integrate Large Language Models (LLMs) into XAI frameworks show promise due to their natural-language capabilities. However, LLM-generated explanations often suffer from confabulations and lack a formal logical structure [24]. Their outputs, while fluent, may be inconsistent or unverifiable, particularly in high-stakes domains. For example, in the case of bank loan evaluations, stating that an applicant was approved because they “resembled others who were approved” lacks transparency about the criteria applied in the current case and fails to support trustworthy decision-making.

4. Twofold Need for Improvement in XAI Research

Explainable Artificial Intelligence (XAI) presents a dual-faceted challenge, encompassing both the technical imperative of producing faithful and accurate representations of model behavior and the human-centered need for interpretability, usefulness, and actionability. While significant advances have been made in developing explanation techniques, current methods often struggle to reconcile these two dimensions, as discussed in Section 3. Bridging this gap remains a critical concern for the XAI research community, necessitating solutions that not only uncover the rationale behind algorithmic decisions but also render them transparent and cognitively accessible. As illustrated in Figure 6, this bifurcation reflects the tension between extracting and representing information from inherently opaque models and addressing the cognitive, psychological, and social factors that influence human understanding, trust, and interaction with AI systems [107,108,109,110]. Effective XAI solutions must therefore bridge this gap, integrating algorithmic transparency with human-centered communication strategies.
In light of the construct presented in Figure 6 (the two facets of XAI), two fundamental questions emerge: (1) From a technical standpoint, is it possible to discuss intelligence—or intelligent systems—without invoking the notion of “knowledge”? If not, how adequate are current knowledge representation approaches in existing XAI methods? (2) Can the challenge of explainability be effectively addressed without engaging with the underlying theoretical foundations of the “explanation” concept itself? The intersection of these two questions highlights the need to assign greater importance to both knowledge representation and the theory of explanation as the XAI community continues its efforts to foster effective user trust in AI systems. These issues are further explored in the following paragraphs.

4.1. The Need for Enhanced Knowledge Representation

From cognitive, philosophical, and psychological perspectives, intelligence and knowledge are fundamentally intertwined. Intelligence is broadly defined as the capacity to apply knowledge for problem-solving, adapt to new situations, and understand complex relationships, while knowledge refers to the facts, information, and skills acquired through experience or education. It is the dynamic interaction between the two that underpins intelligent behavior.
Cognitive science provides a structured view of this relationship, emphasizing that intelligence emerges from both the storage and effective application of structured knowledge. Anderson’s cognitive architecture and the ACT-R theory, for instance, illustrate the centrality of schemas and mental representations in enabling intelligent action [111]. Likewise, Newell and Simon’s problem-solving theory posits that intelligent behavior involves navigating a solution space by leveraging prior knowledge [112].
Philosophical discourse reinforces this perspective. Plato, in The Republic, asserts that knowledge is the foundation of intellectual virtue and a prerequisite for achieving truth and understanding [113]. Descartes’ famous assertion, “Cogito, ergo sum” (“I think, therefore I am”), underscores the indivisibility of thought and knowledge, suggesting that clear and distinct thinking is contingent upon well-grounded knowledge [114].
Psychological theories echo this view. Piaget emphasizes that intelligence develops through the assimilation and accommodation of new knowledge into existing mental schemas [115], while Vygotsky argues that cognitive development—and by extension, intelligence—is socially and culturally mediated through knowledge acquisition and transmission [116].
Building on these foundational insights, the formalization of knowledge for artificial systems began in the 1950s, coinciding with the emergence of early AI. Pioneers such as Marvin Minsky [117], John McCarthy [118,119], and Seymour Papert [120], drawing inspiration from classical philosophers like Aristotle and Plato [121], laid the groundwork for what is now recognized as the field of knowledge representation (KR). Contributions from ancient Eastern traditions, including the Nalanda School, also played a critical role in the early development of KR by promoting structured thought as a cognitive tool [122].
Over the past seven decades, KR has matured into a robust scientific discipline—particularly in ontology engineering—featuring a wide array of techniques that model knowledge through symbolic systems, natural language, and imagery to support intelligent and autonomous software systems [123].
However, despite this rich foundation, many contemporary explainable AI (XAI) approaches inadequately integrate core KR principles. This oversight is particularly evident in data-driven methods, where even knowledge-based approaches, which should enhance interpretability, often suffer from misrepresentation. These methods frequently fall short of the established criteria for KR adequacy, resulting in representations that are incomplete, ambiguous, or opaque [124].
To mitigate these issues, hybrid approaches such as neuro-symbolic AI have emerged that combine the learning capacity of neural networks with the interpretability of symbolic reasoning [125]. While promising in theory, these models encounter challenges in aligning sub-symbolic and symbolic components, often leading to inconsistencies. This misalignment can undermine interpretability, scalability, and transparency—especially in complex or high-stakes domains.
Similarly, knowledge graphs have gained popularity for encoding semantic relationships and supporting AI decision-making processes [126]. They are particularly effective for modeling static relationships and offer high interpretability. However, unless integrated with reasoning engines, they lack the logical rigor required for dynamic or causal inference, thus limiting their utility in domains where explainability is critical.
In sum, accurately and explicitly representing knowledge in a reasoning-compatible manner is essential for developing transparent and trustworthy AI systems, ensuring compliance with representational adequacy standards [124].

4.2. The Need for Effective and Efficient Integration of Explanation Theory into Explainability Design

As explicit knowledge representation emerges as a candidate solution to the first facet of explainable AI—enabling AI systems that are inherently capable of being explained, as discussed and illustrated in Figure 6—the theory of explanation presents itself as the complementary solution to the second critical facet: generating human-compatible explanations grounded in explanation constructs.
A critical challenge in current XAI methods is the insufficient integration of formal explanation theories, which contributes to opaque decision-making in AI systems. Without a clear theoretical foundation, these systems often lack transparency, limiting the traceability and auditability of their reasoning processes. While some efforts have incorporated explanation theories—yielding promising results—the absence of a widely accepted, foundational framework within the XAI community remains a major obstacle.
Several scholars have advocated integrating causal explanation theories, particularly Judea Pearl’s framework of causal inference [127], which models relationships using Bayesian networks [128] and supports counterfactual reasoning through directed acyclic graphs (DAGs). These tools enable mathematically rigorous intervention-based explanations. However, their complexity often hinders accessibility for non-experts, limiting their ability to foster trust. The transparency offered by causal models is indirect, relying on a user’s capacity to interpret statistical relationships and manipulate causal structures.
Other approaches, such as Granger causality [129], causal discovery algorithms [130], and structural causal models [131], have also been explored for their potential to enhance interpretability. Granger causality, while useful for detecting temporal dependencies in time-series data, offers limited support for non-temporal causal reasoning and often lacks intuitive interpretability. Similarly, causal discovery methods—though promising in extracting causal graphs from data—depend heavily on assumptions (e.g., no hidden confounders), and the resulting models are typically opaque to end-users without domain expertise.
Collectively, these methods underscore the promise of causal reasoning for improving explainability and trust in AI. Yet their effectiveness is contingent upon translating complex mathematical outputs into accessible, human-understandable explanations—a challenge that remains unresolved in current XAI practices.
In sum, advancing explainability in AI requires a deliberate and systematic integration of explanation theory. In scientific disciplines, theory serves as the backbone of research, guiding experimental design, methodological development, and result interpretation [132,133]. Without it, explainability research risks becoming fragmented and methodologically incoherent. The lack of a unified theoretical framework has led to ad hoc and domain-specific XAI solutions that are often incapable of delivering transparent reasoning or actionable insights.
This is particularly problematic in high-stakes domains such as healthcare, finance, transportation, and law, where understanding why an AI system makes a decision is as crucial as understanding what that decision is. To achieve truly interpretable and end-to-end explainable AI, the field must adopt a general, formal theory of explanation as the foundation for method development and evaluation. Only then can XAI systems consistently produce explanations that are transparent, trustworthy, and aligned with domain knowledge.

5. Nomological Deductive Reasoning (NDR): A Proposed Solution

Building on the foundations laid in Section 4.1 and Section 4.2, we propose the Nomological Deductive Reasoning (NDR) framework—an approach grounded in Nomological Deductive Knowledge Representation (NDKR)—to address the persistent transparency gap in contemporary explainable AI (XAI) methods. NDR bridges this gap by integrating explicit knowledge representation with Hempel’s deductive–nomological model, thereby enabling the generation of explanations that are both logically rigorous and readily interpretable by human users. At its core, the framework is based on the premise that meaningful explanation requires structured, domain-specific knowledge. Consequently, knowledge representation is treated not merely as a design component but as a foundational mechanism for explainability.
Crucially, the NDKR technique introduces a formal constraint: machine learning models must consult a structured knowledge base—comprising general laws and antecedent conditions—prior to making predictions. This constraint ensures that decision-making is not solely data-driven but also aligned with established causal and logical principles. As a result, the model embeds an internal mechanism for consistency checking and epistemic validation.
Moreover, the NDR framework is inspired by Hempel’s covering-law model of explanation [134], a foundational theory in the philosophy of science that continues to shape formal approaches to explanation. Although more recent models have sought to better reflect human cognitive processes [26,135], rule-based methods remain particularly effective in producing explanations that are transparent, trustworthy, and computationally efficient. Accordingly, by integrating causal inference, deductive reasoning, and domain-specific knowledge, NDR enhances system interpretability and fosters user confidence—without compromising algorithmic performance.
In operational terms, the framework constructs explanations through deductive inference: outcomes are logically derived from general laws and specific conditions. This process mirrors human causal reasoning [136,137,138,139], thereby yielding explanations that are both logically coherent and cognitively aligned. Within the deductive–nomological (D-N) model, a valid explanation demonstrates how the explanandum (the phenomenon to be explained) necessarily follows from the explanans (a set of laws and initial conditions).
This methodological choice is further supported by its effectiveness in addressing “why” questions with clarity and precision. As detailed in Appendix A, our analysis of competing explanatory theories reveals a common emphasis on intelligibility, testability, and causal grounding as key criteria for effective explanation. Furthermore, as noted by Hume [140], human knowledge presupposes that the universe operates under consistent and predictable laws—even when such laws are not immediately observable. Therefore, approaching the explainability challenge through domain-specific laws or principles becomes more understandable to human users.
By the virtue of the covering-law model being materialized through the nomological–deductive formula, while there may be different ways of explaining a phenomenon, science has proved that the event under discussion should be explained by subsuming it under general laws, i.e., by showing that it occurred in accordance with those laws, by virtue of the realization of certain specified antecedent conditions. A D-N explanation is therefore a deductive argument such that the explanandum statement is the result of the explanans [134,141]. Formally, the explanans are made up of m     1   u n i v e r s a l   g e n e r a l i z a t i o n s , referred to as laws, and n   1   s t a t e m e n t s   o f   a n t e c e d e n t   c o n d i t i o n s , as summarized by Equation (1):
L 1 , L 2 , , L m C 1 , C 2 , , C n E
where L 1 ,   L 2 ,   , L m represent the universal generalizations or laws in a certain domain of knowledge, and C 1 ,   C 2 ,   , C n represent the antecedent or conditions that must hold for the law(s) to apply. This covering-law approach to the explanation can be incorporated into every single explanation from any domain, such that any explanation can be analyzed through the lens of the explanation premises captured in Table 2.

5.1. Mathematical Model of the NDR Framework

  • Laws ( L )
    Let L represent the set of laws or rules governing a certain real-world domain (healthcare diagnosis, bank credit score, traffic code for mobility applications, criminal justice, etc.). These laws are formalized as logical statements or principles that provide the foundation for reasoning in the system.
    Each law L i L corresponds to a specific rule or law within the system.
    Example (in credit assessment settings):
    • L 1 : “If an applicant has multiple financial obligations and a credit purpose that are considered risky, then they are at high risk of default”.
    • L 2 : “If the loan duration is > 12 months and the applicant has stable employment, then the credit is classified as ‘Good’”.
  • Conditions (C)
    Let C denote the set of antecedents or conditions that must hold true in order for a law to be applicable to a particular data instance. This means that each condition C j C is a prerequisite or condition that must be satisfied for the corresponding law to be activated or relevant.
    Example (in medical settings):
    • C 1 : “Applicant has stable employment”.
    • C 2 : “The loan duration is > 12 months”.
  • Data instances (D)
    Let D = d 1 ,   d 2 ,   , d k represent the set of input data fed into the AI system. Each d i D represents a specific instance or data sample.
    Example (in credit assessment):
    • d 1 : A data sample where the applicant has stable employment and is applying for a loan payable in more than 12 months.
    • d 2 : A data sample where the applicant does not have any job and is asking for a loan payable in 3 months.
  • Hypothesis or Prediction (H)
    Let H = h 1 ,   h 2 ,   , h p represent the set of predictions or outcomes generated by the AI model. Each h i H corresponds to a specific prediction or outcome generated for the instance d i .
    Example (in credit assessment):
    • h 1 : “The applicant is at high risk for loan payment default”
    • h 2 : “The applicant is not at high risk for loan payment default”.
  • Formalized Deductive Inference
    The key goal of the NDR framework is to use deductive reasoning to formalize the process of how the AI model generates a prediction h i based on the combination of conditions C and laws L applied to the input data d i .
    The formalized deductive inference is represented as
    d i D , h i ,   such   that   C 1 C 2 C n L 1 L 2 , L m h i
    where
    • d i D is an input data instance;
    • C 1 ,   C 2 ,   ,   C n are the conditions (e.g., characteristics of the applicant);
    • L 1 ,   L 2 ,   ,   L m are banking principles or laws (e.g., rules about how conditions relate to prediction like good or bad credit);
    • denotes the deductive reasoning process, where the combination of applicant conditions and banking laws leads to the prediction h i for the given instance d i .
  • Formalized Explanation Generation
    Once we have the laws, conditions and input data, the explanation E i for the prediction h i can be expressed as
    E i = f ( L , C , d i ) h i
    where
    • E i is the explanation for the prediction h i ;
    • f is the function that describes how the laws L , conditions C , and data d i combine to generate the outcome h i ;
    • indicates the reasoning flow from the combination of laws and conditions to the prediction.

5.2. NDR Framework Architecture and Operational Integration with Machine Learning

The Nomological Deductive Reasoning (NDR) framework is designed to produce interpretable, actionable decisions by combining data-driven feature extraction with symbolic reasoning grounded in domain knowledge. This section describes the full system architecture, explains how knowledge is encoded, and details the interaction between machine learning and deductive logic. A structured flow and pseudocode are also provided to operationalize the theory in computational terms.

5.2.1. System Architecture Overview

To operationalize the Nomological Deductive Reasoning (NDR) framework, we propose a multi-layered architecture that integrates symbolic knowledge representation, statistical learning, and deductive inference into a unified explanation pipeline. Figure 7 illustrates the overall system design, detailing how each component—from domain knowledge encoding to user interaction—contributes to generating transparent, norm-compliant, and causally grounded explanations. This architecture supports the core objective of aligning machine learning outputs with structured, law-based reasoning to enhance trust and accountability in AI decision-making.
The Knowledge Representation Layer encodes structured domain-specific knowledge—such as legal statutes, clinical guidelines, or financial regulations, etc.—into a semantic ontology using symbolic rules and causal relationships. This layer forms the epistemic foundation for deductive reasoning. The Machine Learning Layer or AI Model Layer performs statistical learning using traditional [78,142] or deep learning models [143]. It extracts salient features from input data and provides probabilistic or deterministic predictions that are then evaluated by the Reasoning and Deductive Inference Engine against the symbolic rules encoded in the knowledge base. Using logical inference and causal reasoning, it verifies whether the AI’s output aligns with normative expectations and domain constraints. The inferred decision is then passed onto the Explanation Generation Layer to produce human-readable, law-based explanations grounded in domain-specific knowledge as encoded in the knowledge base. These explanations are crafted to be both comprehensible and traceable. Finally, the generated explanation is sent to the User Interaction and Trust Evaluation Layer, which interfaces with end-users, supporting transparent decision communication. Each component contributes to ensuring explainability, traceability, and robustness in high-stakes decision-making scenarios such as in healthcare and medical diagnosis, credit-risk assessment, criminal justice, etc.

5.2.2. Knowledge Encoding: Symbolic Rule Definition

The knowledge base in NDR is composed of two components: domain-specific laws (L) and antecedent conditions (C). These are hand-encoded as logical rules formalized using first-order logic and operationalized as IF–THEN statements. The following are example rules (symbolic format) in a financial-use case:
  • IF checking_account = None AND savings < 100 THEN LiquidityRisk = TRUE;
  • IF employment_duration < 1 year AND age < 30 THEN IncomeInstability = TRUE;
  • IF LiquidityRisk = TRUE AND IncomeInstability = TRUE THEN CreditRisk = High.
Each rule maps observed conditions to semantic risk categories. These rules are stored in a structured ontology that defines the hierarchy and relationships between classes (e.g., ApplicantProfile, RiskIndicator) and properties (e.g., hasCreditBehavior, leadsToRecommendation).

5.2.3. Integration of Machine Learning and Deductive Reasoning

In the NDR framework, machine learning (ML) and symbolic reasoning interact sequentially and complementarily. The ML model is responsible for feature extraction—identifying which attributes are most predictive for the task at hand. However, the final decision is not directly adopted from the ML model. Instead, the most influential features are passed onto the reasoning layer for formal evaluation. This decoupling ensures that ML contributes empirical insights by identifying relevant features, and the deductive system ensures interpretability, logical consistency, and compliance with domain laws. This integration follows a two-stage architecture:
  • Stage 1—ML Model Prediction: A supervised ML model (e.g., deep neural network, random forest, or any other model) generates a prediction h i for input d i .
  • Stage 2—Deductive Reasoning Engine: The reasoning module takes the input instance d i , the prediction h i , and the knowledge base L and C and then verifies whether the prediction can be deduced from the knowledge base, such as by
    ( L C d i ) h i
    If the deduction is consistent, the prediction is validated, and a logical explanation is generated. If not, the system raises a trust warning or requests further input.

5.2.4. Inference Flow and Operationalization

The following step-by-step flow describes the full inference pipeline from the raw input to the final recommendation:
  • Input: For example, in a bank loan application scenario, applicant data are provided as structured records and constitute the system input.
  • ML Feature Extraction: A trained model (e.g., Random Forest) identifies the most influential features. These features are passed as symbolic assertions onto the knowledge layer.
  • Symbolic Assertion Mapping: Features are translated into logical predicates (e.g., SavingsLow(Applicant123)).
  • Deductive Inference: The reasoning engine applies the rules using forward chaining (modus ponens) to derive intermediate-risk indicators and final recommendations.
  • Explanation Generation: Activated rules are converted into natural-language templates.
  • Output: A classification decision (e.g., “High Credit Risk”) with justification is presented to the user.
The flowchart of the process is presented in Figure 8.

5.3. Use-Case Implementation of the NDR-Based Explanations

The initial testing of the NDR framework was conducted for credit scoring on the German Credit dataset [33] by integrating NDR with a Random Forest Classifier algorithm. The experiment aimed to evaluate the framework’s performance in explaining the reasoning behind the predicted creditworthiness based on the input data. Algorithm 1 was used.

5.3.1. Knowledge Base and Ontology Construction in NDR

In the Nomological Deductive Reasoning (NDR) architecture, the knowledge base (KB) serves as the epistemic substrate enabling norm-guided reasoning and explainable inference. Built atop a semantically structured ontology, the KB integrates two core components: domain-specific laws (L) and antecedent conditions (C). These components are formalized using first-order logic (FOL) and operationalized as IF–THEN rules, ensuring that decisions made by machine learning models are both interpretable and normatively aligned. This section details how the KB was constructed using the German Credit Data [33] and aligned with the NDR framework to support high-stakes decision-making in credit-risk assessment.

Ontology Design

The Nomological Deductive Reasoning (NDR) ontology serves as a formal semantic framework for representing the entities, attributes, and interrelations essential for reasoning over structured credit assessment data. It facilitates interpretable, logic-based inference by encoding both normative domain knowledge and case-specific facts in a machine-understandable format. This ontology is structured around key concept classes, which include Applicant, CreditApplication, FinancialProfile, CreditHistory, EmploymentStatus, PropertyOwnership, and RiskAssessment. These classes collectively model an applicant’s demographic attributes, financial behavior, loan parameters, and creditworthiness outcome.
Data properties define attributes such as age, loan amount, duration, employment duration, and credit history. For example, hasAge, hasLoanAmount, hasSavings, and hasNumExistingLoans support quantitative and categorical assessment criteria central to credit decision-making logic.
Object properties (e.g., appliesFor, hasProfile, hasEmploymentStatus) capture relationships between entities, enabling integrated reasoning across multiple dimensions of credit evaluation.
To support logical deduction, the ontology defines derived classes through rule-based expressions using thresholds. For instance,
  • Young(x) ≡ hasAge(x) < 25;
  • HighLoan(x) ≡ hasLoanAmount(x) > 10,000;
  • StableEmployment(x) ≡ hasEmploymentDuration(x) > 1;
  • OwnsProperty(x) ≡ ownsProperty(x) = true.
These formally defined concepts underpin the NDR framework’s rule-based laws (e.g., L1–L15 as defined in the Domain-specific Laws’ subsection), enabling precise, explainable reasoning over structured representations aligned with financial norms.

Domain-Specific Laws (L)

Domain-specific laws encode general principles about credit risk that are derived from financial expertise and empirical patterns. In the NDR framework, these laws are formalized in FOL and expressed as deductive rules that define the logical conditions under which an applicant is expected to be high- or low-risk.
L1: High Loan + No Checking Account ⟹ High Risk: Applicants requesting high loan amounts without a checking account are likely to be high-risk:
x ( H i g h L o a n x N o C h e c k i n g A c c o u n t x H i g h R i s k x )
L2: Young Age + Long Loan Duration ⟹ High Risk: Young applicants with long loan durations are more likely to default:
x ( Y o u n g x L o n g L o a n D u r a t i o n x H i g h R i s k x )
L3: Stable Employment + High Savings ⟹ Low Risk: Stable employment and substantial savings indicate financial reliability:
x ( S t a b l e E m p l o y m e n x H i g h S a v i n g s x ¬ H i g h R i s k x )
L4: Multiple Existing Loans ⟹ High Risk: Multiple existing credits increase the financial burden and risk:
x ( N u m E x i s t i n g L o a n s x > 2 H i g h R i s k x )
L5: Long Loan Duration Indicates Lower Risk: Longer loan durations are typically associated with structured repayment plans, which reflect lender confidence.
x ( L o a n D u r a t i o n x > 12 ¬ H i g h R i s k x )
L6: Full Repayment of Previous Credits Implies Trustworthiness: Applicants who have fully repaid past credits are less likely to default.
x ( P a i d P r e v i o u s C r e d i t s x ¬ H i g h R i s k x )
L7: Stable Employment (>1 year) Signals Financial Stability: Consistent employment indicates a reliable income source.
x ( E m p l o y m e n t D u r a t i o n x > 1 S t a b l e E m p l o y m e n t x )
L8: Ownership of a House or Real Estate Reduces Risk: Property ownership suggests financial maturity and collateral availability.
x ( O w n s P r o p e r t y x ¬ H i g h R i s k x )
L9: Very Young Applicants (<25) Have Elevated Default Risk: This reflects the behavioral and income uncertainty that are often correlated with young age groups.
x ( O w n s P r o p e r t y x ¬ H i g h R i s k x )
L10: Long-Term Employment (>2 years) Significantly Reduces Risk: Extended employment history contributes strongly to financial reliability.
x ( E m p l o y m e n t D u r a t i o n x > 2 ¬ H i g h R i s k x )
L11: Absence of a Foreign Worker Status Positively Influences Creditworthiness: Applicants who are not classified as foreign workers exhibit stronger integration and employment stability, which favorably impact credit assessments.
x ( ¬ F o r e i g n W o r k e r x G o o d C r e d i t x )
L12: Certain Personal Status Categories Contribute Positively to Credit Assessment: Applicants falling within specific personal status categories (e.g., single male, divorced female) are associated with slightly improved creditworthiness, potentially due to social or economic correlations.
x ( P e r s o n a l S t a t u s x ( 2,3 ] S l i g h t P o s i t i v e I m p a c t x )
L13: A Low Job Classification Marginally Enhances the Credit Evaluation: Contrary to conventional assumptions, being in the lowest job tier may correlate with basic but stable employment, contributing weakly positively to credit decisions.
x ( J o b L e v e l x 0 S l i g h t P o s i t i v e I m p a c t x )
L14: A Limited Credit History Increases the Risk of Default: Applicants with minimal or unfavorable credit histories demonstrate higher uncertainty in repayment behavior, which slightly increases risk.
x ( C r e d i t H i s t o r y x 2 H i g h R i s k x )
L15: A Low Checking Account Status Slightly Decreases Creditworthiness: Applicants without a robust checking account history tend to receive lower credit assessments due to their perceived lack of financial infrastructure.
x ( C h e c k i n g A c c t S t a t u s x 1 S l i g h t N e g a t i v e I m p a c t x )
Each predicate (e.g., HighLoan, Young, HighRisk) is grounded in ontological definitions, where threshold values and categorical labels (e.g., “young” = age < 25) are specified as concept constraints or semantic annotations.

Antecedent Conditions (C)

Antecedent conditions represent instance-specific facts, derived from input data and expressed as logical assertions. These instantiate the ontology with individual-level information. As an example, applicant A123 has the following information:
-
Age: 22;
-
Checking Account Status: None;
-
Employment duration: 0.5;
-
Credit history: critical;
-
Number of existing loans: 3;
-
Savings: less than 100.
These facts map onto the ontology and activate predicates such as
  • Young(A123), HighLoan(A123), NoCheckingAccount(A123), UnstableEmployment(A123), NumExistingLoans(A123) > 2.

Logical Reasoning and Inference

Using the logical rules (L) and instantiated conditions (C), the NDR inference engine performs deductive reasoning to assess whether the AI model’s prediction aligns with the domain norms. For example, applying laws L1, L2, and L4 to A123:
HighLoan(A123) ∧ NoCheckingAccount(A123) → HighRisk(A123)
Young(A123) ∧ LongLoanDuration(A123) → HighRisk(A123)
NumExistingLoans(A123) > 2 → HighRisk(A123)

5.3.2. Data Processing and Model Training

The German Credit Dataset [33] comprises 1000 credit applicants, each described by 20 financial and demographic attributes, including features such as credit history, loan duration, employment status, savings level, and age. The target variable, Class, indicates the applicant’s creditworthiness, where a value of 1 represents Good Credit and a value of 0 denotes Bad Credit. The dataset contains a mixture of categorical and numerical variables, offering a comprehensive profile of each individual.
Data Preprocessing:
Categorical variables were transformed into numerical representations using label encoding. Numerical attributes were preserved in their original form. To ensure experimental reproducibility, the dataset was partitioned into a training set (70%) and a testing set (30%) using a fixed random seed (random_state = 42).
Model Training:
A Random Forest classifier was trained on the processed training set with the following settings:
  • Number of trees: n_estimators = 100;
  • Criterion: default Gini impurity;
  • Random seed for reproducibility: random_state = 42;
  • Other hyper parameters retained their default Scikit-learn values.
No additional hyper-parameter optimization (e.g., grid search or randomized search) was performed, as the objective was to demonstrate baseline model behavior integrated with the NDR framework.
Hardware and Software Environment:
The implementation environment consisted of Python 3.10.11 and Scikit-learn 1.5.1. Model training and evaluation were performed on a machine featuring an Intel Core i5-4200M CPU @ 2.5 GHz (2 cores) and 4 GB of RAM, running Windows 10.
Given the relatively small size of the German Credit dataset (1000 instances), computational resource limitations did not impact the training or evaluation performance. The model was trained and evaluated within a reasonable timeframe without any notable delays or constraints.
Model Evaluation and Explanation Generation:
After training, the Random Forest model was used to generate predictions on the test set. These predictions were assessed using standard evaluation metrics (accuracy and confusion matrix). Additionally, a feature importance analysis was performed to identify the most influential variables, supporting the subsequent deductive reasoning.
The predictions were then passed onto the Nomological Deductive Reasoning (NDR) framework, where they were evaluated against a structured financial knowledge base encoded with domain-specific laws. This process generated semantically grounded, human-readable explanations, aligning the model’s decision-making with established banking norms and ontological definitions. The whole process is described by Algorithm 1, and the implementation code is publicly available at https://github.com/HakizimanaGedeon/NDR-Framework-Explainable-AI, accessed on 20 March 2025.
Algorithm 1. NDR Framework: SP Algorithm with Formal Inference
1: Input: I (input data), KB (Knowledge Base), M (Model), F (Feature Extraction function), R (Reasoning function), E (Explanation Generation function)
2: Output: Explanation Ei for each prediction h i y ^ set of predictions
3: Step 1: Preprocess the input data
4: I′ ← Preprocess(I) {Clean and format input data}
5: Step 2: Model prediction ~ Extract features from the preprocessed input
6: F ← ExtractFeatures (I′) {Feature extraction using ML model}
7: Step 3: Retrieve relevant laws from the Knowledge Base
8: L ← GetLaws(KB) {Retrieve laws from Knowledge Base}
9: Step 4: Retrieve the conditions or antecedents for the relevant laws
10: C ← GetConditions(KB) {Retrieve conditions that must hold true for laws}
11: Step 5: Perform reasoning based on the extracted features and laws
12: C′ ← Reason(F, L, C) {Reason using features, laws and conditions}
13: Step 6: Perform prediction based on reasoning and extracted features
14: y ^ ← Predict(M, C′) {Model prediction based on reasoning output and features}
15: Step 7: Generate the explanation based on reasoning, features, and prediction
16: Ei ← E(L′, C′, I′, y ^ ) {Explanation generation based on laws, conditions, data, and prediction}
17: Step 8: Output the generated explanation
18: Return Ei

5.3.3. Experimental Results

To assess the effectiveness of the proposed Nomological Deductive Reasoning (NDR) framework, we conducted a series of experiments using the Statlog (German Credit Data) dataset [33]. Our evaluation focused on three key dimensions: predictive performance, quality of the rule-based explanations, and alignment with the domain-specific knowledge encoded in the Knowledge Base (KB).

Prediction Accuracy

The NDR-augmented model achieved a high prediction accuracy of 97%, correctly classifying individuals as either “Good Credit” or “Bad Credit”. This demonstrates strong statistical performance on the credit-risk classification task.

Rule-Based Explanation Generation

For each prediction, the NDR framework applied relevant rules from the KB to generate structured explanations. These explanations link individual feature values to domain-specific rules grounded in banking regulations, ensuring that decisions are causally interpretable and auditable.
  • Prediction Case 1: “Good Credit”
  • Explanation: “The model predicts a ‘Good credit’ because based on banking laws, the applicant has a stable family or marital status, which is positive for creditworthiness; manageable installment rates, which indicate financial stability; and the loan duration is for more than 12 months, which indicates a positive sign of credit”.
  • Model Reasoning: (i) Personal Status: Stable marital/family status → positively influences creditworthiness, (ii) Installment Rate: Manageable installment amount → indicates financial responsibility, and (iii) Duration: Loan duration > 12 months → considered favorable under domain rules.
  • Prediction Case 2: “Bad Credit”.
  • Explanation: “The model predicts a bad credit because based on banking laws, the person has multiple financial obligations which increases financial risk, and the credit purpose is considered risky, contributing to a negative credit prediction”.
  • Model Reasoning: (i) Other Debtors: Multiple financial obligations → higher perceived risk and (ii) Purpose: Loan purpose considered high-risk → contributes to a negative classification.

Alignment with Domain Knowledge

The deductive consistency of the model’s outputs was validated by comparing its generated explanations with the logic encoded in the Knowledge Base. The following were found:
  • Rule Consistency: All 300 predictions were logically consistent with the rules defined in the Knowledge Base.
  • Rule Coverage: Every prediction was fully accounted for by one or more symbolic rules, resulting in a Rule Coverage = 1.00, meaning 100% of rule coverage before decision-making.
  • Mismatch Penalty: There were no mismatches between the model’s explanations and the rule base, indicating a Mismatch Penalty of 0, reflecting 100% loyalty to the rule base.

Summary of Results

Table 3 summarizes the key evaluation metrics, highlighting the NDR framework’s ability to combine high predictive performance with complete logical consistency and rule-based explainability.
These results demonstrate that NDR not only preserves predictive accuracy but also guarantees that every decision is logically grounded, transparent, and normatively aligned—a crucial requirement for AI adoption in high-stakes domains such as finance.

5.3.4. Evaluation of the Explanatory Quality of the NDR Framework

To assess the quality of explanatory outputs generated by the NDR framework, we focused on three key dimensions of explainability: trustworthiness, human-readability, and actionability. These criteria were selected based on a comprehensive review of the literature and the practical needs of high-stakes domains like finance, where decisions must be interpretable, transparent, and actionable. Consider the following explanation produced by NDR for a credit application decision: “The model predicts a *Bad Credit* because based on banking laws, the applicant has multiple financial obligations which increases the financial risk, and the credit purpose is considered risky, contributing to a negative credit prediction”. This explanation scores highly on trustworthiness, as it aligns with established lending criteria and domain-specific reasoning. It also exhibits strong human-readability, using plain language and avoiding technical jargon, which makes it accessible to non-technical users. Finally, the explanation is clearly actionable, offering concrete steps the applicant can take to influence future outcomes (reducing financial obligations such as clearing his existing debts and properly motivating the purpose of the loan). This kind of output reflects the core strengths of the NDR framework and illustrates how it outperforms alternative methods that may produce less coherent, overly technical, or non-instructive explanations vis-à-vis the end user, who may be a simple lay consumer. A comparative qualitative analysis of NDR alongside other emerging methods—LIME, causal inference models, neuro-symbolic reasoning, knowledge graph-based approaches, and LLM-driven narratives—is presented in Figure 10a,b, using consistent instances to highlight interpretive differences. The full code of this comparison is available at https://github.com/HakizimanaGedeon/NDR-Framework-Explainable-AI.
On the other hand, as illustrated in Figure 10b, the prediction of “Bad Credit” for Instance #4 is accompanied by explanations generated from six different interpretability frameworks, each offering varying degrees of clarity and utility. Among these, the Nomological Deductive Reasoning (NDR) approach delivers the most structured and contextually grounded explanation. It explicitly attributes the adverse credit decision to the absence of a checking account, referencing domain-specific banking norms that associate such financial behavior with increased risk. This reasoning aligns closely with human cognitive expectations, offering both traceability and normative justification. In contrast, the Causal Inference explanation identifies statistically relevant factors—such as younger age (<25) being linked to a higher default risk and stable employment mitigating risk—but lacks decisiveness in synthesizing these competing signals into a coherent conclusion. Similarly, the Neuro-Symbolic explanation emphasizes employment stability as a favorable indicator but does not resolve the tension between this positive factor and the final negative prediction. The Knowledge Graph explanation constructs a semantic path that includes both risk (absence of a checking account) and security (property ownership) yet stops short of articulating how these are weighed in the final decision. The LLM-style explanation, while human-readable, primarily restates input features without embedding them in a domain-specific interpretive framework. Lastly, the LIME explanation provides local feature-importance scores, showing that a “ForeignWorker” status exerted the strongest negative influence, but it does not offer a global rationale or domain-based interpretation. Overall, the NDR explanation distinguishes itself by integrating formal logic with domain-relevant rules, enabling it to generate actionable, transparent justifications. This makes NDR particularly suitable for high-stakes applications, where explanation fidelity and alignment with institutional norms are critical.
From the illustrations in Figure 10a,b, the trustworthiness of NDR-generated explanations stems, first, from their foundation in formal deductive reasoning. Unlike post hoc or correlation-based techniques, NDR explanations are derived from domain-specific causal laws and antecedent conditions encoded in a structured knowledge base. This ensures that each prediction is not only statistically plausible but also epistemically justified. The deductive inference engine enforces consistency by penalizing predictions that violate predefined rules, thereby improving reliability. The empirical results demonstrated that a high proportion of the model outputs remained consistent with the knowledge base, validating the explanatory rigor of the framework.
Second, human-readability is achieved through structured natural-language explanations that mirror human causal reasoning. Explanations are delivered in an “if–then” format, with clearly articulated reasons based on interpretable features. For example:
Prediction: “Good Credit”
Explanation:
  • “The model predicts a ‘Good Credit’ because, based on banking laws, the duration of the loan is longer than 12 months, which indicates a positive sign for credit, and the installment rate is manageable, which suggests financial stability”.
    which is the more human-friendly form adapted by NDR explanation layer from:
  • “If the duration of the loan is longer than 12 months, which indicates a positive sign for credit, and if the installment rate is manageable, suggesting financial stability, then the model predicts a ‘Good Credit’ based on banking laws”.
These explanations are designed to highlight key variables, offer concise reasoning, and use plain language free of technical jargon. Although a formal user study is planned for future work, the current framework was developed with interpretability and end-user accessibility as guiding principles.
Third, the actionability of NDR outputs is supported by their alignment with modifiable input features and real-world financial decision logic. By identifying contributing factors (e.g., installment rate, loan duration, personal status) and explaining their relevance through legal or policy-based rules, the explanations offer users a pathway for improving or justifying outcomes. This makes them particularly useful in domains where decision support is essential.
To summarize the quality of explanations, Table 4 outlines the relationship between NDR’s design mechanisms and the corresponding dimensions of explainability:
These findings reinforce that NDR not only enhances model transparency but also improves the practical utility of AI systems in real-world contexts. By combining formal logic with domain-grounded reasoning, the framework closes the transparency gap identified in current XAI methods.

6. Discussion and Future Work

The initial experimental findings demonstrate the efficacy of the Nomological Deductive Reasoning (NDR) framework in integrating structured domain knowledge into machine learning models for credit scoring. By formalizing and embedding financial principles as logical rules, the NDR framework provides transparent, interpretable, and human-aligned explanations for model predictions. This approach not only bridges the gap between data-driven modeling and domain expertise but also offers pathways to enhance the trustworthiness of AI systems by offering explanations that are logically consistent and verifiable.
A key strength of the NDR framework lies in its ability to generate explanations that align with human reasoning, particularly in high-stakes domains such as finance, where regulatory compliance, user trust, and interpretability are paramount. From the financial experiment, the explanations derived from the NDR model are not only readable and actionable but also grounded in established financial logic, offering stakeholders clarity on why certain predictions are made. This contributes significantly to the field of explainable AI (XAI) by moving beyond statistical approximations toward logic-based, nomologically valid justification. The perfect Rule Coverage and absence of any Mismatch Penalty further underscore the alignment between model behavior and the domain knowledge encoded in the framework.
From a practical standpoint, the NDR framework offers a valuable tool for institutions seeking to improve transparency in their credit-risk assessments, enabling easier auditability and more defensible decision-making. Furthermore, its domain-agnostic architecture opens up possibilities for deployment in other critical areas such as healthcare diagnostics, criminal justice, insurance underwriting, and legal reasoning, where similar demands for explainability and rule adherence exist.
From a theoretical perspective, this work contributes to the ongoing discourse on hybrid AI systems by demonstrating how symbolic reasoning and machine learning can coexist in a mutually reinforcing manner. It also emphasizes the importance of ontology-driven explanations in contexts where decision-making must be not only correct but justifiable.
However, several limitations point to promising avenues for future research. The current implementation of NDR assumes a moderately sized, static rule base, which may face scalability challenges in domains with large, dynamic, or probabilistic rule sets. Enhancing the framework’s ability to operate efficiently with extensive or evolving ontologies remains an open challenge. Additionally, while the deductive component excels in rule-based reasoning, it currently lacks mechanisms to accommodate uncertainty or conflicting rules, which are common in real-world decision-making environments. To address these limitations, future research will focus on three key directions:
  • Scalability and Automation: Developing methods for automated rule induction from structured data and text, as well as mechanisms for dynamic rule updates, will be crucial for broadening the applicability of the NDR framework in real-time and large-scale systems.
  • Handling Uncertainty: Integrating probabilistic reasoning or fuzzy logic into the NDR architecture could improve its robustness in domains where knowledge is imprecise, incomplete, or uncertain.
  • Hybrid Integration with LLMs: A promising direction for enhancing the NDR framework lies in its integration with Large Language Models (LLMs), such as GPT-3 and GPT-4. These models have shown substantial potential in explainable AI (XAI) by generating coherent, human-readable explanations [144,145,146,147,148,149]. Their ability to articulate complex model behavior in natural language makes them valuable for broadening accessibility and personalizing explanations for diverse user groups [148,150]. However, LLMs often lack transparency in their internal reasoning, functioning as black boxes without a structured, verifiable logic chain behind their outputs [151]. This opaqueness, combined with tendencies toward hallucinations [24], biases [152], and occasionally vague or overly generic responses, limits their trustworthiness in high-stakes settings. To address this, future work will explore how LLMs can be coupled with the rule-based rigor of NDR—where LLMs serve as natural-language interfaces that translate formal rule-based deductions into user-friendly narratives. Such a hybrid approach aims to preserve the logical integrity of NDR while enhancing user engagement and interpretability through the expressive capabilities of LLMs.
In conclusion, this research offers both practical contributions—a transparent, rule-based credit-scoring framework—and theoretical insights into the integration of symbolic reasoning with machine learning. The NDR approach not only strengthens the foundation of explainable AI but also points toward a future in which AI systems are both powerful and principled, ensuring that decisions are not just accurate but also aligned with human values and domain logic.

Author Contributions

Conceptualization, G.H. and A.L.E.; methodology, G.H. and A.L.E.; investigation, G.H.; writing—original draft preparation, G.H.; writing—review and editing, G.H. and A.L.E.; supervision, A.L.E.; project administration, A.L.E. All authors have read and agreed to the published version of the manuscript.

Funding

The work of Dr. Agapito Ledezma was supported by Grant PID2021-124335OB-C22 funded by MCIN/AEI 10.13039/501100011033 and the European Union.

Data Availability Statement

The original contributions presented in the study are included in the article, and no new datasets were involved. Further inquiries can be directed to the corresponding author.

Acknowledgments

We thank all the anonymous reviewers for their supportive comments for further improving this article.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A

The literature up until the end of the twentieth century reflects a rich landscape of theoretical models aimed at characterizing scientific explanation. Among the most influential are the deductive–nomological (D-N) model proposed by Carl Hempel [134], widely regarded as the most prominent philosophical account of explanation, and Wesley Salmon’s Statistical Relevance (SR) [153] and Causal Mechanical (CM) [154] models. Additional noteworthy contributions include the Unificationist models advanced by Michael Friedman [155] and Philip Kitcher [156], as well as the Pragmatic theory of explanation developed by Bas van Fraassen [157].
Despite the inevitable divergences inherent in philosophical discourse, a common evaluative criterion can be discerned across these models: adherence to empiricist standards of intelligibility and testability with respect to the concepts employed in explanatory statements. Notably, all these theories—each with its own unique formal structure—converge on the fundamental notion that explanation involves identifying factors or causes that underlie the occurrence of an event E. This is perhaps most explicitly formalized in Hempel’s model, which frames explanation through general laws and antecedent conditions serving as the premises of explanatory arguments.
Table A1 provides a comparative overview of the principal tenets of each model. It also underscores the shared conceptual foundation across these diverse theories, thereby supporting the proposition that a generic framework of explanation may be viable, particularly within the emerging field of explainable Artificial Intelligence (XAI).
Table A1. The most prominent and competing explanation models and their intersection.
Table A1. The most prominent and competing explanation models and their intersection.
Explanation TheoryAuthor(s)Main Idea of the ModelCommon Element (with All Other Theories)
Deductive Model of ExplanationCarl G. Hempel and
Paul Oppenheim [134]
Explanation involves showing that the phenomenon to be explained follows deductively from general laws or premises. It emphasizes logical reasoning and general laws.Logical structure and relationships: Like other theories, it connects a phenomenon to its cause/effect, focusing on structure and relationships (causal, statistical, etc.).
Statistical Relevance
(SR Model)
Wesley Salmon [153]Explanations are based on showing the statistical relevance of factors. A phenomenon is explained by demonstrating how certain statistical factors increase the probability of the event.Cause–effect relationships: Emphasizes how statistical relevance or probability connects events, similar to the focus on causal mechanisms in other theories.
Causal
Mechanical Model
Wesley Salmon [154]Explanations are based on identifying and describing the causal mechanisms that lead to the phenomenon. It focuses on the underlying processes or mechanisms.Causal relationships and mechanisms: Like the deductive model, it aims to logically link causes to effects while focusing on mechanisms, as seen in other theories.
Unificationist ModelMichael Friedman and Philip Kitcher [155,156]Explanations aim to unify a range of phenomena under a common principle or theory. The goal is to reduce complexity by explaining multiple phenomena in terms of a single framework.Simplification and coherence: Like other theories, it seeks to organize complex phenomena, reducing complexity while maintaining causal or statistical relationships.
Pragmatic Model of ExplanationBas van Fraassen [157]Explanations are context-dependent, and their effectiveness is judged by their utility in a given context. They are considered good if they meet practical needs and goalsContextual and practical utility: Shares with other theories the need for the explanation to be useful and relevant for understanding or predictions, with a focus on practical applications.

References

  1. Adadi, A.; Berrada, M. Peeking Inside the Black-Box: A Survey on Explainable Artificial Intelligence (XAI). IEEE Access 2018, 6, 52138–52160. [Google Scholar] [CrossRef]
  2. Pasquale, F. The Black Box Society: The Secret Algorithms That Control Money and Information; Harvard University Press: Cambridge, MA, USA, 2015. [Google Scholar]
  3. Zednik, C. Solving the Black Box Problem: A Normative Framework for Explainable Artificial Intelligence. Philos. Technol. 2021, 34, 265–288. [Google Scholar] [CrossRef]
  4. Hassija, V.; Chamola, V.; Mahapatra, A.; Singal, A.; Goel, D.; Huang, K.; Scardapane, S.; Spinelli, I.; Mahmud, M.; Hussain, A. Interpreting Black Box Models: A Review on Explainable Artificial Intelligence. Cogn. Comput. 2024, 16, 45–74. [Google Scholar] [CrossRef]
  5. Peters, U. Explainable AI Lacks Regulative Reasons: Why AI and Human Decision-Making Are Not Equally Opaque. AI Ethics 2023, 3, 963–974. [Google Scholar] [CrossRef]
  6. Minh, D.; Wang, H.X.; Li, Y.F.; Nguyen, T.N. Explainable Artificial Intelligence: A Comprehensive Review. Artif. Intell. Rev. 2022, 55, 3503–3568. [Google Scholar] [CrossRef]
  7. Evans, T.; Retzlaff, C.O.; Geißler, C.; Kargl, M.; Plass, M.; Müller, H.; Kiehl, T.-R.; Zerbe, N.; Holzinger, A. The Explainability Paradox: Challenges for xAI in Digital Pathology. Fut. Genet. Comput. Syst. 2022, 133, 281–296. [Google Scholar] [CrossRef]
  8. Černevičienė, J.; Kabašinskas, A. Explainable Artificial Intelligence (XAI) in Finance: A Systematic Literature Review. Artif. Intell. Rev. 2024, 57, 216. [Google Scholar] [CrossRef]
  9. Weber, P.; Carl, K.V.; Hinz, O. Applications of Explainable Artificial Intelligence in Finance—A Systematic Review of Finance, Information Systems, and Computer Science Literature. Manag. Rev. Q. 2024, 74, 867–907. [Google Scholar] [CrossRef]
  10. Atakishiyev, S.; Salameh, M.; Yao, H.; Goebel, R. Explainable artificial intelligence for autonomous driving: A comprehensive overview and field guide for future research directions. IEEE Access 2024, 12, 101603–101625. [Google Scholar] [CrossRef]
  11. European Union. General Data Protection Regulation (GDPR). 2016. Available online: https://gdpr.eu/ (accessed on 25 January 2025).
  12. OECD. OECD Principles on Artificial Intelligence. 2019. Available online: https://www.oecd.org/going-digital/ai/principles/ (accessed on 25 January 2025).
  13. European Commission. Artificial Intelligence Act. 2021. Available online: https://digital-strategy.ec.europa.eu/en/policies/regulatory-framework-ai (accessed on 25 January 2025).
  14. State Council of China. Ethical Guidelines for Artificial Intelligence. 2022. Available online: https://www.gov.cn/zhengce/2022-01-01/ (accessed on 25 January 2025).
  15. Ministry of Internal Affairs and Communications of Japan. AI R&D Guidelines. 2020. Available online: https://www.soumu.go.jp/main_content/000730466.pdf (accessed on 25 January 2025).
  16. Australian Government. AI Ethics Framework. 2020. Available online: https://www.industry.gov.au/data-and-publications/australias-artificial-intelligence-ethics-framework (accessed on 25 January 2025).
  17. South African Government. South Africa’s Artificial Intelligence Policy. 2023. Available online: https://www.dcdt.gov.za/sa-national-ai-policy-framework/file/338-sa-national-ai-policy-framework.html (accessed on 25 January 2025).
  18. Nigerian Government. Nigeria’s Artificial Intelligence Policy. 2023. Available online: https://www.nitda.gov.ng (accessed on 25 January 2025).
  19. Rwanda Government. Rwanda’s AI Policy. 2023. Available online: https://www.minict.gov.rw/index.php?eID=dumpFile&t=f&f=67550&token=6195a53203e197efa47592f40ff4aaf24579640e (accessed on 25 January 2025).
  20. Ribeiro, M.T.; Singh, S.; Guestrin, C. “Why should I trust you?” Explaining the Predictions of Any Classifier. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; pp. 1135–1144. [Google Scholar] [CrossRef]
  21. Lundberg, S.M.; Lee, S.-I. A Unified Approach to Interpreting Model Predictions. Adv. Neural Inf. Process. Syst. 2017, 30. [Google Scholar] [CrossRef]
  22. Ribeiro, M.T.; Singh, S.; Guestrin, C. Anchors: High-Precision Model-Agnostic Explanations. In Proceedings of the 32nd International Conference on Neural Information Processing Systems (NeurIPS 2018); pp. 1525–1535. Available online: https://arxiv.org/abs/1802.07681 (accessed on 18 December 2024).
  23. Chen, J.; Song, L.; Wang, S.; Xie, L.; Wang, X.; Zhang, M. Towards Prototype-Based Explanations of Deep Neural Networks. In Proceedings of the 36th International Conference on Machine Learning (ICML 2019), 2019. Available online: https://arxiv.org/abs/1905.11742 (accessed on 14 November 2024).
  24. Ehsan, U.; Riedl, M. Explainable AI Reloaded: Challenging the XAI Status Quo in the Era of Large Language Models. In Proceedings of the Halfway to the Future Symposium, Santa Cruz, CA, USA, 21–23 October 2024; pp. 1–8. [Google Scholar]
  25. Giudici, P.; Raffinetti, E. Shapley-Lorenz Explainable Artificial Intelligence. Expert Syst. Appl. 2021, 167, 114104. [Google Scholar] [CrossRef]
  26. Moradi, M.; Samwald, M. Post-Hoc Explanation of Black-Box Classifiers Using Confident Itemsets. Expert Syst. Appl. 2021, 165, 113941. [Google Scholar] [CrossRef]
  27. Vilone, G.; Longo, L. Explainable Artificial Intelligence: A Systematic Review. arXiv 2020, arXiv:2006.00093. Available online: https://arxiv.org/abs/2006.00093 (accessed on 25 January 2025).
  28. Miller, T. Explanation in Artificial Intelligence: Insights from the Social Sciences. arXiv 2018, arXiv:1706.07222. Available online: https://arxiv.org/abs/1706.07222 (accessed on 25 January 2025).
  29. Longo, L.; Brcic, M.; Cabitza, F.; Choi, J.; Confalonieri, R.; Del Ser, J.; Guidotti, R.; Hayashi, Y.; Herrera, F.; Holzinger, A.; et al. Explainable Artificial Intelligence (XAI) 2.0: A Manifesto of Open Challenges and Interdisciplinary Research Directions. Inf. Fusion 2024, 106, 102301. [Google Scholar] [CrossRef]
  30. Lipton, Z.C. The Mythos of Model Interpretability. Commun. ACM 2018, 61, 36–43. [Google Scholar] [CrossRef]
  31. Rago, A.; Palfi, B.; Sukpanichnant, P.; Nabli, H.; Vivek, K.; Kostopoulou, O.; Kinross, J.; Toni, F. Exploring the Effect of Explanation Content and Format on User Comprehension and Trust. arXiv 2024, arXiv:2408.17401. Available online: https://arxiv.org/abs/2408.17401 (accessed on 25 January 2025).
  32. Ali, S.; Abuhmed, T.; El-Sappagh, S.; Muhammad, K.; Alonso-Moral, J.M.; Confalonieri, R.; Guidotti, R.; Del Ser, J.; Díaz-Rodríguez, N.; Herrera, F. Explainable Artificial Intelligence (XAI): What We Know and What Is Left to Attain Trustworthy Artificial Intelligence. Inf. Fusion 2023, 99, 101805. [Google Scholar] [CrossRef]
  33. Hofmann, H. Statlog (German Credit Data). UCI Machine Learning Repository. 1994. Available online: https://archive.ics.uci.edu/dataset/144/statlog+german+credit+data (accessed on 10 February 2025).
  34. Breiman, L. Random Forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef]
  35. De Graaf, M.M.; Malle, B.F. How People Explain Action (and Autonomous Intelligent Systems Should Too). In Proceedings of the 2017 AAAI Fall Symposium Series, Arlington, VA, USA, 9–11 November 2017. [Google Scholar]
  36. Harbers, M.; van den Bosch, K.; Meyer, J.-J. A Study into Preferred Explanations of Virtual Agent Behavior. In Intelligent Virtual Agents, Proceedings of the 9th International Workshop, Amsterdam, The Netherlands, 14–16 September 2009; Nijholt, A., Reidsma, D., Hondorp, H., Eds.; Springer: Berlin/Heidelberg, Germany, 2009; pp. 132–145. [Google Scholar]
  37. Vilone, G.; Longo, L. Notions of Explainability and Evaluation Approaches for Explainable Artificial Intelligence. Inf. Fusion 2021, 76, 89–106. [Google Scholar] [CrossRef]
  38. Vilone, G.; Longo, L. Classification of Explainable Artificial Intelligence Methods through Their Output Formats. Mach. Learn. Knowl. Extr. 2021, 3, 615–661. [Google Scholar] [CrossRef]
  39. Love, P.; Fang, W.; Matthews, J.; Porter, S.; Luo, H.; Ding, L. Explainable Artificial Intelligence (XAI): Precepts, Methods, and Opportunities for Research in Construction. arXiv 2022, arXiv:2211.06579. [Google Scholar] [CrossRef]
  40. Malaviya, C.; Lee, S.; Roth, D.; Yatskar, M. What If You Said That Differently? How Explanation Formats Affect Human Feedback Efficacy and User Perception. In Proceedings of the 2024 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 1: Long Papers), Mexico City, Mexico, 16–21 June 2024; pp. 3046–3065. [Google Scholar]
  41. Morrison, K.; Shin, D.; Holstein, K.; Perer, A. Evaluating the Impact of Human Explanation Strategies on Human-AI Visual Decision-Making. Proc. ACM Hum.—Comput. Interact. 2023, 7, 1–37. [Google Scholar] [CrossRef]
  42. Chattopadhay, A.; Sarkar, A.; Howlader, P.; Balasubramanian, V.N. Grad-CAM++: Generalized Gradient-Based Visual Explanations for Deep Convolutional Networks. In Proceedings of the 2018 IEEE Winter Conference on Applications of Computer Vision (WACV), Lake Tahoe, NV, USA, 12–15 March 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 839–847. [Google Scholar]
  43. Bhowal, P.; Sen, S.; Yoon, J.H.; Geem, Z.W.; Sarkar, R. Choquet Integral and Coalition Game-Based Ensemble of Deep Learning Models for COVID-19 Screening From Chest X-Ray Images. IEEE J. Biomed. Health Inform. 2021, 25, 4328–4339. [Google Scholar] [CrossRef]
  44. Selvaraju, R.R.; Cogswell, M.; Das, A.; Vedantam, R.; Parikh, D.; Batra, D. Grad-CAM: Visual Explanations from Deep Networks via Gradient-Based Localization. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 618–626. [Google Scholar]
  45. Simonyan, K.; Vedaldi, A.; Zisserman, A. Deep Inside Convolutional Networks: Visualising Image Classification Models and Saliency Maps. arXiv 2014, arXiv:1312.6034. [Google Scholar]
  46. Erhan, D.; Bengio, Y.; Courville, A.; Vincent, P. Visualizing Higher-Layer Features of a Deep Network. Univ. Montr. 2009, 1341, 1. [Google Scholar]
  47. Shrikumar, A.; Greenside, P.; Kundaje, A. Learning Important Features Through Propagating Activation Differences. arXiv 2017, arXiv:1704.02685v2. [Google Scholar] [CrossRef]
  48. Bach, S.; Binder, A.; Montavon, G.; Klauschen, F.; Müller, K.R.; Samek, W. On Pixel-Wise Explanations for Non-Linear Classifier Decisions by Layer-Wise Relevance Propagation. PLoS ONE 2015, 10, e0130140. [Google Scholar] [CrossRef] [PubMed] [PubMed Central]
  49. Sundararajan, M.; Taly, A.; Yan, Q. Axiomatic Attribution for Deep Networks. In Proceedings of the 34th International Conference on Machine Learning (ICML 2017), Sydney, Australia, 6–11 August 2017; Precup, D., Teh, Y.W., Eds.; PMLR: Cambridge, MA, USA, 2017; pp. 3319–3328. [Google Scholar]
  50. Mostafa, S.; Mondal, D.; Beck, M.A.; Bidinosti, C.P.; Henry, C.J.; Stavness, I. Leveraging Guided Backpropagation to Select Convolutional Neural Networks for Plant Classification. Front. Artif. Intell. 2022, 5, 871162. [Google Scholar] [CrossRef]
  51. Petsiuk, V.; Das, A.; Saenko, K. RISE: Randomized Input Sampling for Explanation of Black-Box Models. arXiv 2018, arXiv:1806.07421. [Google Scholar]
  52. Kumar, D.; Wong, A.; Taylor, G.W. Explaining the Unexplained: A Class-Enhanced Attentive Response (CLEAR) Approach to Understanding Deep Neural Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Honolulu, HI, USA, 21–26 July 2017; pp. 36–44. [Google Scholar]
  53. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An Image Is Worth 16x16 Words: Transformers for Image Recognition at Scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  54. Poli, J.-P.; Ouerdane, W.; Pierrard, R. Generation of Textual Explanations in XAI: The Case of Semantic Annotation. In Proceedings of the 2021 IEEE International Conference on Fuzzy Systems (FUZZ-IEEE), Luxembourg, 11–14 July 2021; pp. 1–6. [Google Scholar] [CrossRef]
  55. Selvaraju, R.R.; Chattopadhyay, P.; Elhoseiny, M.; Sharma, T.; Batra, D.; Parikh, D.; Lee, S. Choose Your Neuron: Incorporating Domain Knowledge through Neuron-Importance. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 526–541. [Google Scholar]
  56. Bellini, V.; Schiavone, A.; Di Noia, T.; Ragone, A.; Di Sciascio, E. Knowledge-Aware Autoencoders for Explainable Recommender Systems. In Proceedings of the 3rd Workshop on Deep Learning for Recommender Systems, Vancouver, BC, Canada, 6 October 2018; pp. 24–31. [Google Scholar]
  57. Zhang, W.; Paudel, B.; Zhang, W.; Bernstein, A.; Chen, H. Interaction Embeddings for Prediction and Explanation in Knowledge Graphs. In Proceedings of the Twelfth ACM International Conference on Web Search and Data Mining, Melbourne, Australia, 11–15 January 2019; pp. 96–104. [Google Scholar]
  58. Lei, T.; Barzilay, R.; Jaakkola, T. Rationalizing Neural Predictions. arXiv 2016, arXiv:1606.04155. [Google Scholar]
  59. Barratt, S. InterpNet: Neural Introspection for Interpretable Deep Learning. arXiv 2017, arXiv:1710.09511. [Google Scholar]
  60. Hendricks, L.A.; Akata, Z.; Rohrbach, M.; Donahue, J.; Schiele, B.; Darrell, T. Generating Visual Explanations. In Computer Vision–ECCV 2016: Proceedings of the 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016, Proceedings, Part IV; Springer International Publishing: Berlin/Heidelberg, Germany, 2016; Volume 14, pp. 3–19. [Google Scholar]
  61. Garcia-Magarino, I.; Muttukrishnan, R.; Lloret, J. Human-Centric AI for Trustworthy IoT Systems with Explainable Multilayer Perceptrons. IEEE Access 2019, 7, 125562–125574. [Google Scholar] [CrossRef]
  62. Bennetot, A.; Laurent, J.L.; Chatila, R.; Díaz-Rodríguez, N. Towards Explainable Neural-Symbolic Visual Reasoning. arXiv 2019, arXiv:1909.09065. [Google Scholar]
  63. Letham, B.; Rudin, C.; McCormick, T.H.; Madigan, D. Interpretable Classifiers Using Rules and Bayesian Analysis: Building a Better Stroke Prediction Model. J. Am. Stat. Assoc. 2015, 110, 1350–1371. [Google Scholar] [CrossRef]
  64. Fürnkranz, J. Rule-Based Methods. In Encyclopedia of Systems Biology; Dubitzky, W., Wolkenhauer, O., Cho, K.H., Yokota, H., Eds.; Springer: New York, NY, USA, 2013; pp. 1703–1706. [Google Scholar] [CrossRef]
  65. Quinlan, J.R. Induction of Decision Trees. Mach. Learn. 1986, 1, 81–106. [Google Scholar] [CrossRef]
  66. Ribeiro, M.T.; Singh, S.; Guestrin, C. Anchors: High-Precision Model-Agnostic Explanations. In Proceedings of the Thirty-Second AAAI Conference on Artificial Intelligence, the Thirtieth Innovative Applications of Artificial Intelligence Conference, and the Eighth AAAI Symposium on Educational Advances in Artificial Intelligence (AAAI’18/IAAI’18/EAAI’18), New Orleans, LA, USA, 2–7 February 2018; AAAI Press: Palo Alto, CA, USA, 2018; pp. 1527–1535. [Google Scholar]
  67. Yang, H.; Rudin, C.; Seltzer, M. Scalable Bayesian Rule Lists. In Proceedings of the 34th International Conference on Machine Learning (ICML 2017), Sydney, Australia, 6–11 August 2017; Precup, D., Teh, Y.W., Eds.; PMLR: Cambridge, MA, USA, 2017; pp. 3921–3930. [Google Scholar]
  68. Almutairi, M.; Stahl, F.; Bramer, M. Reg-Rules: An Explainable Rule-Based Ensemble Learner for Classification. IEEE Access 2021, 9, 52015–52035. [Google Scholar] [CrossRef]
  69. Lakkaraju, H.; Bach, S.H.; Leskovec, J. Interpretable Decision Sets: A Joint Framework for Description and Prediction. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; pp. 1675–1684. [Google Scholar]
  70. Gao, J.; Liu, N.; Lawley, M.; Hu, X. An Interpretable Classification Framework for Information Extraction from Online Healthcare Forums. J. Healthc. Eng. 2017, 2017, 9275318. [Google Scholar] [CrossRef]
  71. Shortliffe, E.H.; Davis, R.; Axline, S.G.; Buchanan, B.G.; Green, C.C.; Cohen, S.N. Computer-Based Consultations in Clinical Therapeutics: Explanation and Rule Acquisition Capabilities of the MYCIN System. Comput. Biomed. Res. 1975, 8, 303–320. [Google Scholar] [CrossRef]
  72. Keneni, B.M.; Kaur, D.; Al Bataineh, A.; Devabhaktuni, V.K.; Javaid, A.Y.; Zaientz, J.D.; Marinier, R.P. Evolving Rule-Based Explainable Artificial Intelligence for Unmanned Aerial Vehicles. IEEE Access 2019, 7, 17001–17016. [Google Scholar] [CrossRef]
  73. Bride, H.; Dong, J.; Dong, J.S.; Hóu, Z. Towards Dependable and Explainable Machine Learning Using Automated Reasoning. In Proceedings of the Formal Methods and Software Engineering: 20th International Conference on Formal Engineering Methods, ICFEM 2018, Gold Coast, QLD, Australia, 12–16 November 2018; Proceedings 20. Springer International Publishing: Berlin/Heidelberg, Germany, 2018; pp. 412–416. [Google Scholar]
  74. Johansson, U.; Niklasson, L.; König, R. Accuracy vs. Comprehensibility in Data Mining Models. In Proceedings of the Seventh International Conference on Information Fusion, Stockholm, Sweden, 28 June–1 July 2004; Elsevier: Amsterdam, The Netherlands, 2004; Volume 1, pp. 295–300. [Google Scholar]
  75. Setzu, M.; Guidotti, R.; Monreale, A.; Turini, F.; Pedreschi, D.; Giannotti, F. GlocalX—From Local to Global Explanations of Black Box AI Models. Artif. Intell. 2021, 294, 103457. [Google Scholar] [CrossRef]
  76. Asano, K.; Chun, J. Post-Hoc Explanation Using a Mimic Rule for Numerical Data. In Proceedings of the ICAART (2), Virtual, Online, 4–6 February 2021; pp. 768–774. [Google Scholar]
  77. Friedman, J.H. Greedy Function Approximation: A Gradient Boosting Machine. Ann. Stat. 2001, 29, 1189–1232. [Google Scholar] [CrossRef]
  78. Hastie, T.; Tibshirani, R.; Friedman, J. The Elements of Statistical Learning: Data Mining, Inference, and Prediction; Springer: New York, NY, USA, 2001. [Google Scholar]
  79. Altmann, A.; Toloşi, L.; Sander, O.; Lengauer, T. Permutation Importance: A Corrected Feature Importance Measure. Bioinformatics 2010, 26, 1340–1347. [Google Scholar] [CrossRef]
  80. Marcílio, W.E.; Eler, D.M. From Explanations to Feature Selection: Assessing SHAP Values as Feature Selection Mechanism. In Proceedings of the 2020 33rd SIBGRAPI Conference on Graphics, Patterns and Images (SIBGRAPI), Recife/Porto de Galinhas, Brazil, 7–10 November 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 340–347. [Google Scholar] [CrossRef]
  81. Goldstein, A.; Kapelner, A.; Bleich, J.; Pitkin, E. Peeking Inside the Black Box: Visualizing Statistical Learning with Plots of Individual Conditional Expectation. J. Comput. Graph. Stat. 2015, 24, 44–65. [Google Scholar] [CrossRef]
  82. Kim, B.; Wattenberg, M.; Gilmer, J.; Cai, C.; Wexler, J.; Viegas, F. Interpretability Beyond Feature Attribution: Quantitative Testing with Concept Activation Vectors (TCAV). In Proceedings of the 35th International Conference on Machine Learning (ICML 2018), Stockholm, Sweden, 10–15 July 2018; Caruana, R., Lou, Y., Eds.; PMLR: Cambridge, MA, USA, 2018; pp. 2668–2677. [Google Scholar]
  83. Tan, S.; Caruana, R.; Hooker, G.; Lou, Y. Distill-and-Compare: Auditing Black-Box Models Using Transparent Model Distillation. In Proceedings of the 2018 AAAI/ACM Conference on AI, Ethics, and Society, New Orleans, LA, USA, 2–7 February 2018; pp. 303–310. [Google Scholar]
  84. Strumbelj, E.; Kononenko, I. An Efficient Explanation of Individual Classifications Using Game Theory. J. Mach. Learn. Res. 2010, 11, 1–18. [Google Scholar]
  85. Štrumbelj, E.; Kononenko, I. Towards a Model Independent Method for Explaining Classification for Individual Instances. In Proceedings of the Data Warehousing and Knowledge Discovery: 10th International Conference, DaWaK 2008, Turin, Italy, 2–5 September 2008; Proceedings 10. Springer: Berlin/Heidelberg, Germany, 2008; pp. 273–282. [Google Scholar]
  86. Adler, P.; Falk, C.; Friedler, S.A.; Nix, T.; Rybeck, G.; Scheidegger, C.; Smith, B.; Venkatasubramanian, S. Auditing Black-Box Models for Indirect Influence. Knowl. Inf. Syst. 2018, 54, 95–122. [Google Scholar] [CrossRef]
  87. Alain, G.; Bengio, Y. Understanding Intermediate Layers Using Linear Classifier Probes. arXiv 2016, arXiv:1610.01644. [Google Scholar] [CrossRef]
  88. Främling, K. Explaining Results of Neural Networks by Contextual Importance and Utility. In Proceedings of the AISB, Brighton, UK, 1–2 April 1996; Volume 96. [Google Scholar]
  89. Juscafresa, A.N. An Introduction to Explainable Artificial Intelligence with LIME and SHAP. Bachelor’s Thesis, Degree in Computer Engineering, Universitat de Barcelona, Barcelona, Spain, 29 June 2022. Supervised by Dr. Albert Clapés and Dr. Sergio Escalera. Available online: https://www.google.com/url?sa=t&source=web&rct=j&opi=89978449&url=https://sergioescalera.com/wp-content/uploads/2022/06/presentacio_tfg_nieto_juscafresa_aleix.pdf&ved=2ahUKEwjni4PW4sCNAxXF7TgGHbtUAJsQFnoECBkQAQ&usg=AOvVaw35mUA85cyJvPfQ2SaFsHTS (accessed on 27 January 2025).
  90. Chen, J.; Song, L.; Wainwright, M.; Jordan, M. Learning to Explain: An Information-Theoretic Perspective on Model Interpretation. In Proceedings of the 35th International Conference on Machine Learning 2018, Stockholm, Sweden, 10–15 July 2018; pp. 883–892. [Google Scholar]
  91. Verma, S.; Boonsanong, V.; Hoang, M.; Hines, K.E.; Dickerson, J.P.; Shah, C. Counterfactual Explanations and Algorithmic Recourses for Machine Learning: A Review. arXiv 2020, arXiv:2010.10596. [Google Scholar] [CrossRef]
  92. Xu, K.; Ba, J.; Kiros, R.; Cho, K.; Courville, A.; Salakhutdinov, R.; Zemel, R.; Bengio, Y. Show, Attend and Tell: Neural Image Caption Generation with Visual Attention. In Proceedings of the 32nd International Conference on Machine Learning (ICML 2015), Lille, France, 6–11 July 2015; Bach, F., Blei, D., Eds.; PMLR: Cambridge, MA, USA, 2015; pp. 2048–2057. [Google Scholar]
  93. Yang, S.C.H.; Shafto, P. Explainable Artificial Intelligence via Bayesian Teaching. In Proceedings of the NIPS 2017 Workshop on Teaching Machines, Robots, and Humans, Long Beach, CA, USA, 9 December 2017; Volume 2. [Google Scholar]
  94. Caruana, R.; Kangarloo, H.; Dionisio, J.D.; Sinha, U.; Johnson, D. Case-Based Explanation of Non-Case-Based Learning Methods. In Proceedings of the AMIA Symposium; American Medical Informatics Association: Bethesda, MD, USA, 1999; pp. 212–215. [Google Scholar] [PubMed] [PubMed Central]
  95. Spinner, T.; Schlegel, U.; Schäfer, H.; El-Assady, M. explAIner: A Visual Analytics Framework for Interactive and Explainable Machine Learning. IEEE Trans. Vis. Comput. Graph. 2019, 26, 1064–1074. [Google Scholar] [CrossRef]
  96. Khanna, R.; Kim, B.; Ghosh, J.; Koyejo, S. Interpreting Black Box Predictions Using Fisher Kernels. In Proceedings of the 22nd International Conference on Artificial Intelligence and Statistics (AISTATS 2019), Naha, Okinawa, Japan, 16–18 April 2019; Chaudhuri, K., Sugiyama, M., Eds.; PMLR: Cambridge, MA, USA, 2019; pp. 3382–3390. [Google Scholar]
  97. Dhurandhar, A.; Chen, P.-Y.; Luss, R.; Tu, C.-C.; Ting, P.; Shanmugam, K.; Das, P. Explanations Based on the Missing: Towards Contrastive Explanations with Pertinent Negatives. In Advances in Neural Information Processing Systems 31 (NeurIPS 2018); Bengio, S., Wallach, H., Larochelle, H., Grauman, K., Cesa-Bianchi, N., Garnett, R., Eds.; Curran Associates, Inc.: Red Hook, NY, USA, 2018; pp. 592–603. [Google Scholar]
  98. Agostinho, C.; Dikopoulou, Z.; Lavasa, E.; Perakis, K.; Pitsios, S.; Branco, R.; Reji, S.; Hetterich, J.; Biliri, E.; Lampathaki, F.; et al. Explainability as the Key Ingredient for AI Adoption in Industry 5.0 Settings. Front. Artif. Intell. 2023, 6, 1264372. [Google Scholar] [CrossRef] [PubMed]
  99. Hanif, A.; Zhang, F.; Wang, H.; Barhamgi, M.; Chen, L.; Zhou, R. A Comprehensive Survey of Explainable Artificial Intelligence (XAI) Methods: Exploring Transparency and Interpretability. In Web Information Systems Engineering—WISE 2023; Zhang, F., Wang, H., Barhamgi, M., Chen, L., Zhou, R., Eds.; Lecture Notes in Computer Science; Springer: Singapore, 2023; Volume 14306, pp. 1–13. [Google Scholar] [CrossRef]
  100. Dvorak, J.; Kopp, T.; Kinkel, S.; Lanza, G. Explainable AI: A Key Driver for AI Adoption, a Mistaken Concept, or a Practically Irrelevant Feature? In Proceedings of the 4th UR-AI Symposium, Villingen-Schwenningen, Germany, 14–16 September 2022; pp. 88–101. [Google Scholar]
  101. Adebayo, J.; Gilmer, J.; Muelly, M.; Goodfellow, I.; Hardt, M.; Kim, B. Sanity Checks for Saliency Maps. arXiv 2018, arXiv:1810.03292. [Google Scholar] [CrossRef]
  102. Kim, B.; Seo, J.; Jeon, S.; Koo, J.; Choe, J.; Jeon, T. Why Are Saliency Maps Noisy? Cause of and Solution to Noisy Saliency Maps. arXiv 2019, arXiv:1902.04893. [Google Scholar] [CrossRef]
  103. Slack, D.; Song, L.; Koyejo, S.O.; Padhye, J.; Dhurandhar, A.; Zhang, Y.; Sattigeri, P.; Hughes, T.; Mojsilović, A.; Varshney, K.R.; et al. Fooling LIME and SHAP: Adversarial Attacks on Post-Hoc Explanation Methods. In Proceedings of the 2020 Conference on Fairness, Accountability, and Transparency (FAccT), Barcelona, Spain, 27–30 January 2020; ACM: New York, NY, USA, 2020; pp. 1–12. [Google Scholar] [CrossRef]
  104. Ghorbani, A.; Abid, A.; Zou, J. Interpretation of Neural Networks is Fragile. In Proceedings of the 36th International Conference on Machine Learning (ICML 2019), Long Beach, CA, USA, 9–15 June 2019; Chaudhuri, K., Salakhutdinov, R., Eds.; PMLR: Cambridge, MA, USA, 2019; Volume 97, pp. 2242–2251. Available online: http://proceedings.mlr.press/v97/ghorbani19c.html (accessed on 27 January 2025).
  105. Caruana, R.; Lou, Y.; Gehrke, J.; Koch, P.; Sturm, M.; Elhadad, N. Intelligible models for healthcare: Predicting pneumonia risk and hospital 30-day readmission. In Proceedings of the 21st ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Sydney, Australia, 10–13 August 2015; pp. 1721–1730. [Google Scholar]
  106. Carter, S.; Kim, B.; Brown, R.; Doshi-Velez, F. Visualizing and Understanding High-Dimensional Models in Machine Learning. In Proceedings of the 2019 IEEE International Conference on Data Science and Advanced Analytics (DSAA), Washington, DC, USA, 5–8 October 2019; pp. 230–239. [Google Scholar] [CrossRef]
  107. Molnar, C. Interpretable Machine Learning: A Guide for Making Black Box Models Explainable; Springer: Berlin/Heidelberg, Germany, 2020; pp. 173–174. [Google Scholar]
  108. Van der Laan, M.J.; Woutersen, D. Ethics of Explainability: The Social and Psychological Implications of Algorithmic Decisions. J. Ethics 2021, 25, 121–146. [Google Scholar]
  109. Dave, P.; Dastin, J. Money, Mimicry, and Mind Control: Big Tech Slams Ethics Brakes on AI. Reuters. 2021. Available online: https://news.trust.org/item/20210908095953-jtdiz (accessed on 20 March 2025).
  110. Hoffman, R.R.; Mueller, S.T.; Klein, G. Explaining Explanations: A Taxonomy of AI Interpretability and Its Implications for Trust and User Behavior. ACM Trans. Interact. Intell. Syst. 2018, 8, 25–34. [Google Scholar]
  111. Anderson, J.R. Cognitive Psychology and Its Implications, 6th ed.; W.H. Freeman: New York, NY, USA, 2005. [Google Scholar]
  112. Newell, A.; Simon, H.A. Human Problem Solving; Prentice-Hall: Englewood Cliffs, NJ, USA, 1972. [Google Scholar]
  113. Plato; Jowett, B. The Republic; Dover Publications: Mineola, NY, USA, 1999. [Google Scholar]
  114. Descartes, R. Meditations on First Philosophy; Cottingham, J., Translator; Cambridge University Press: Cambridge, UK, 1996. [Google Scholar]
  115. Piaget, J. The Psychology of the Child; Basic Books: New York, NY, USA, 1972. [Google Scholar]
  116. Vygotsky, L.S. Mind in Society: The Development of Higher Psychological Processes; Cole, M., John-Steiner, V., Scribner, S., Souberman, E., Eds.; Harvard University Press: Cambridge, MA, USA, 1978. [Google Scholar]
  117. Minsky, M. A Framework for Representing Knowledge; Technical Report; Massachusetts Institute of Technology: Cambridge, MA, USA, 1974. [Google Scholar]
  118. McCarthy, J.J.; Minsky, M.L.; Rochester, N. Artificial Intelligence; Research Laboratory of Electronics (RLE), Massachusetts Institute of Technology (MIT): Cambridge, MA, USA, 1959; Available online: https://dspace.mit.edu/bitstream/handle/1721.1/52263/RLE_QPR_053_XIII.pdf (accessed on 17 February 2025).
  119. McCarthy, J. Programs with Common Sense. In Mechanization of Thought Processes; Brain, M., Ed.; Her Majesty’s Stationery Office: London, UK, 1959; Volume I, Available online: https://stacks.stanford.edu/file/druid:yt623dt2417/yt623dt2417.pdf (accessed on 15 February 2025).
  120. Goldstein, I.; Papert, S. Artificial Intelligence, Language, and the Study of Knowledge. Cogn. Sci. 1977, 1, 84–123. [Google Scholar]
  121. Sowa, J.F. Knowledge Representation: Logical, Philosophical, and Computational Foundations; Brooks/Cole: Pacific Grove, CA, USA, 2000; Volume 13. [Google Scholar]
  122. Di Maio, M. Mindful Technology. Buddhist Door, Online Article, Hong Kong, 2019. Available online: https://www.buddhistdoor.net/features/knowledge-representation-in-the-nalanda-buddhist-tradition (accessed on 17 February 2025).
  123. Guarino, N. The Ontological Level: Revisiting 30 Years of Knowledge Representation. In Conceptual Modeling: Foundations and Applications; Springer: Berlin/Heidelberg, Germany, 2009; pp. 52–67. [Google Scholar]
  124. Di Maio, P. Neurosymbolic Knowledge Representation for Explainable and Trustworthy AI. Preprints 2020, 2020010163. [Google Scholar] [CrossRef]
  125. Besold, T.R.; d’Avila Garcez, A.; Bader, S.; Bowman, H.; Domingos, P.; Hitzler, P.; Kühnberger, K.-U.; Lamb, L.C.; Lowd, D.; Moura, J.M.F.; et al. Neural-Symbolic Learning and Reasoning: A Survey and Interpretation 1. In Neuro-Symbolic Artificial Intelligence: The State of the Art; IOS Press: Amsterdam, The Netherlands, 2021; pp. 1–51. [Google Scholar]
  126. Tiddi, I.; Schlobach, S. Knowledge Graphs as Tools for Explainable Machine Learning: A Survey. Artif. Intell. 2022, 302, 103627. [Google Scholar] [CrossRef]
  127. Pearl, J. The Seven Tools of Causal Inference, with Reflections on Machine Learning. Commun. ACM 2019, 62, 54–60. [Google Scholar] [CrossRef]
  128. Pearl, J. Probabilistic Reasoning in Intelligent Systems: Networks of Plausible Inference; Morgan Kaufmann: San Mateo, CA, USA, 1988. [Google Scholar]
  129. Granger, C.W.J. Investigating Causal Relations by Econometric Models and Cross-Spectral Methods. Econometrica 1969, 37, 424–438. [Google Scholar] [CrossRef]
  130. Spirtes, P.; Glymour, C.; Scheines, R. Causation, Prediction, and Search, 2nd ed.; MIT Press: Cambridge, MA, USA, 2000. [Google Scholar] [CrossRef]
  131. Pearl, J. Causality: Models, Reasoning, and Inference, 2nd ed.; Cambridge University Press: Cambridge, UK, 2009. [Google Scholar] [CrossRef]
  132. Merton, R.K. The Sociology of Science: Theoretical and Empirical Investigations; University of Chicago Press: Chicago, IL, USA, 1967. [Google Scholar]
  133. Popper, K. The Logic of Scientific Discovery; Routledge: London, UK, 2002. [Google Scholar]
  134. Hempel, C.G.; Oppenheim, P. Studies in the Logic of Explanation. Philos. Sci. 1948, 15, 135–175. [Google Scholar] [CrossRef]
  135. Yang, S.-H.C.; Folke, N.E.T.; Shafto, P. A Psychological Theory of Explainability. In Proceedings of the 39th International Conference on Machine Learning (ICML 2022); Chaudhuri, K., Salakhutdinov, R., Eds.; PMLR: Cambridge, MA, USA, 2022; Volume 162, pp. 25007–25021. [Google Scholar]
  136. Johnson-Laird, P.N. Mental Models: Towards a Cognitive Science of Language and Reasoning; Cambridge University Press: Cambridge, UK, 1983. [Google Scholar]
  137. Chater, N.; Oaksford, M. The Rational Analysis of Deductive Reasoning. Psychol. Rev. 1999, 106, 443–467. [Google Scholar]
  138. Baron, J. Thinking and Deciding, 4th ed.; Cambridge University Press: Cambridge, UK, 2008. [Google Scholar]
  139. Newton, I. Philosophiæ Naturalis Principia Mathematica (Mathematical Principles of Natural Philosophy); Royal Society: London, UK, 1687. [Google Scholar]
  140. Hume, D. An Enquiry Concerning Human Understanding; A. Millar: London, UK, 1748. [Google Scholar]
  141. Hempel, C.G. Aspects of Scientific Explanation and Other Essays in the Philosophy of Science; Free Press: New York, NY, USA, 1965. [Google Scholar]
  142. Mitchell, T.M. Machine Learning; McGraw-Hill: New York, NY, USA, 1997. [Google Scholar]
  143. Goodfellow, I.; Bengio, Y.; Courville, A. Deep Learning; MIT Press: Cambridge, MA, USA, 2016; Available online: https://www.deeplearningbook.org (accessed on 20 February 2025).
  144. Bhattacharjee, A.; Moraffah, R.; Garland, J.; Liu, H. Towards LLM-Guided Causal Explainability for Black-Box Text Classifiers. arXiv 2024, arXiv:2309.13340. [Google Scholar]
  145. Kroeger, N.; Ley, D.; Krishna, S.; Agarwal, C.; Lakkaraju, H. In-context explainers: Harnessing LLMs for explaining black box models. arXiv 2023, arXiv:2310.05797. [Google Scholar]
  146. Nguyen, V.B.; Schlötterer, J.; Seifert, C. From Black Boxes to Conversations: Incorporating XAI in a Conversational Agent. In Proceedings of the World Conference on Explainable Artificial Intelligence, Lisboa, Portugal, 26–28 July 2023; pp. 71–96. [Google Scholar]
  147. Slack, D.; Krishna, S.; Lakkaraju, H.; Singh, S. Explaining machine learning models with interactive natural language conversations using TalkToModel. Nat. Mach. Intell. 2023, 5, 873–883. [Google Scholar] [CrossRef]
  148. Zytek, A.; Pidò, S.; Veeramachaneni, K. LLMs for XAI: Future directions for explaining explanations. arXiv 2024, arXiv:2405.06064. [Google Scholar]
  149. Burton, J.; Al Moubayed, N.; Enshaei, A. Natural Language Explanations for Machine Learning Classification Decisions. In Proceedings of the 2023 International Joint Conference on Neural Networks (IJCNN), Queensland, Australia, 18–23 June 2023; pp. 1–9. [Google Scholar]
  150. Mavrepis, P.; Makridis, G.; Fatouros, G.; Koukos, V.; Separdani, M.M.; Kyriazis, D. XAI for all: Can large language models simplify explainable AI? arXiv 2024, arXiv:2401.13110. [Google Scholar]
  151. Cambria, E.; Malandri, L.; Mercorio, F.; Nobani, N.; Seveso, A. XAI meets LLMs: A survey of the relation between explainable AI and large language models. arXiv 2024, arXiv:2407.15248. [Google Scholar]
  152. Lareo, X. Large Language Models (LLM). European Data Protection Supervisor. 2023. Available online: https://edps.europa.eu (accessed on 2 March 2025).
  153. Salmon, W.C. Statistical Explanation and Statistical Relevance; University of Pittsburgh Press: Pittsburgh, PA, USA, 1971; Volume 69. [Google Scholar]
  154. Salmon, W.C. Scientific Explanation and the Causal Structure of the World; Princeton University Press: Princeton, NJ, USA, 1984. [Google Scholar]
  155. Friedman, M. Explanation and scientific understanding. J. Philos. 1974, 71, 5–19. [Google Scholar] [CrossRef]
  156. Kitcher, P. Explanatory unification. Philos. Sci. 1981, 48, 507–531. [Google Scholar] [CrossRef]
  157. Van Fraassen, B.C. The pragmatics of explanation. Am. Philos. Q. 1977, 14, 143–150. [Google Scholar]
Figure 1. Methodological workflow for developing and evaluating the NDR framework.
Figure 1. Methodological workflow for developing and evaluating the NDR framework.
Algorithms 18 00306 g001
Figure 2. Chest X-ray examples labeled as (a) COVID-19, (b) Pneumonia, and (c) Normal. Each row (from left to right) shows the input image, AI prediction, and Grad-CAM heatmap. Highlights reflect regions influencing the model’s decision. Source: [43].
Figure 2. Chest X-ray examples labeled as (a) COVID-19, (b) Pneumonia, and (c) Normal. Each row (from left to right) shows the input image, AI prediction, and Grad-CAM heatmap. Highlights reflect regions influencing the model’s decision. Source: [43].
Algorithms 18 00306 g002
Figure 3. Example of a saliency map. The pictures on the right are the saliency maps of the left image and show the regions that are more attentive parts for the CNN. (Source: geeksforgeeks.org/what-is-saliency-map; accessed on 25 January 2025).
Figure 3. Example of a saliency map. The pictures on the right are the saliency maps of the left image and show the regions that are more attentive parts for the CNN. (Source: geeksforgeeks.org/what-is-saliency-map; accessed on 25 January 2025).
Algorithms 18 00306 g003
Figure 4. Example of rule-based explanations. Source: [64].
Figure 4. Example of rule-based explanations. Source: [64].
Algorithms 18 00306 g004
Figure 5. Illustration of an SP-LIME algorithm calculated from 500 candidate explanations sampled from the data uniformly at random. Red indicates a contribution to non-diabetes and green indicates a contribution to diabetes. Source: [89].
Figure 5. Illustration of an SP-LIME algorithm calculated from 500 candidate explanations sampled from the data uniformly at random. Red indicates a contribution to non-diabetes and green indicates a contribution to diabetes. Source: [89].
Algorithms 18 00306 g005
Figure 6. The two facets of XAI challenges in the field.
Figure 6. The two facets of XAI challenges in the field.
Algorithms 18 00306 g006
Figure 7. Proposed architecture for an NDR-based explanation approach.
Figure 7. Proposed architecture for an NDR-based explanation approach.
Algorithms 18 00306 g007
Figure 8. Flowchart diagram of NDR-based explanations.
Figure 8. Flowchart diagram of NDR-based explanations.
Algorithms 18 00306 g008
Figure 9. Sample explanations generated by the NDR framework for credit-risk predictions based on the German Credit dataset [33]. Each explanation illustrates how specific input features are mapped to domain rules defined in the Knowledge Base, enabling transparent, causal reasoning behind the model’s decision. The examples demonstrate both positive and negative classifications, with logical justifications grounded in financial regulations.
Figure 9. Sample explanations generated by the NDR framework for credit-risk predictions based on the German Credit dataset [33]. Each explanation illustrates how specific input features are mapped to domain rules defined in the Knowledge Base, enabling transparent, causal reasoning behind the model’s decision. The examples demonstrate both positive and negative classifications, with logical justifications grounded in financial regulations.
Algorithms 18 00306 g009
Figure 10. (a) Comparative explanation outputs for an instance number # predicted as **Good Credit** using six interpretability methods: Nomological Deductive Reasoning (NDR), Causal Inference, Neuro-Symbolic Reasoning, Knowledge Graph Paths, LLM-style explanations, and LIME. NDR stands out by providing a structured, law-based rationale for predictions, offering enhanced transparency compared with the data-driven feature attributions and descriptive summaries from other approaches. (b) Comparative explanation outputs for an instance number # predicted as **Bad Credit** using six interpretability methods: Nomological Deductive Reasoning (NDR), Causal Inference, Neuro-Symbolic Reasoning, Knowledge Graph Paths, LLM-style explanations, and LIME. While the LIME and LLM approaches highlight influential features and descriptive summaries, only NDR provides a normative, rule-based justification anchored in domain knowledge, thereby enhancing interpretability and decision transparency.
Figure 10. (a) Comparative explanation outputs for an instance number # predicted as **Good Credit** using six interpretability methods: Nomological Deductive Reasoning (NDR), Causal Inference, Neuro-Symbolic Reasoning, Knowledge Graph Paths, LLM-style explanations, and LIME. NDR stands out by providing a structured, law-based rationale for predictions, offering enhanced transparency compared with the data-driven feature attributions and descriptive summaries from other approaches. (b) Comparative explanation outputs for an instance number # predicted as **Bad Credit** using six interpretability methods: Nomological Deductive Reasoning (NDR), Causal Inference, Neuro-Symbolic Reasoning, Knowledge Graph Paths, LLM-style explanations, and LIME. While the LIME and LLM approaches highlight influential features and descriptive summaries, only NDR provides a normative, rule-based justification anchored in domain knowledge, thereby enhancing interpretability and decision transparency.
Algorithms 18 00306 g010
Table 1. Examples of textual explanations as found in the XAI literature.
Table 1. Examples of textual explanations as found in the XAI literature.
Learning TaskExplanation ExampleReference
Animal’s image classification
Algorithms 18 00306 i001
Prediction: Yellow-headed blackbird
has_eye_color = black
has_underparts_color = white
has_belly_color = white
has_breast_color = white
has_breast_pattern = solid
[55]
Movie
Recommendation

Prediction:
Terminator 2
We guess you would like to watch Terminator 2: Judgement Day (1991) more than Transformers: Revenge of the Fallen (2009) because you may prefer:
  • (subject) 1990s science-fiction films
  • (subject) Science-fiction adventure films
  • (subject) Films using computer-generated imagery
  • (subject) Drone films
  • (subject) Cyberpunk films
Over:
  • (subject) Films set in Egypt
  • (subject) Robot films
  • (subject) Films shot in Arizona
  • (subject) Ancient astronauts in fiction
  • (subject) IMAX films
[56]
Knowledge graph completion (link prediction)Prediction:
World War I—entity involved—German empire
Explanation:
World War I—commanders—Erich Ludendorff
Erich Ludendorff—commands—German Empire
Supported by:
Falkland Wars—entities involved—United Kingdom
Falkland Wars—commanders—Margaret Thatcher
Margaret Thatcher—commands—United Kingdom
[57]
Table 2. Premises of explanation arguments.
Table 2. Premises of explanation arguments.
Explanation ConstituentsDescriptionExample
ExplanandumSentence describing the phenomenon to be explained (not the phenomenon itself)The leaves of the tea plant are yellowing.
ExplanansClass of those sentences which are adduced to account for the phenomenon. These sentences fall into two subclasses:

Certain sentences C 1 ,   C 2 ,   ,   C n , which state specific antecedent conditions,
and
a set of sentences L 1 ,   L 2 ,   ,   L m , which represent general laws in the domain of the problem at hand.
While the weather and soil conditions seem to be normal, the leaves of tea plant are yellowing

Nitrogen deficiency leads to the yellowing of leaves (chlorosis) in tea plants
Table 3. Summary of the NDR framework experiment results.
Table 3. Summary of the NDR framework experiment results.
MetricValueDescription
Accuracy97%Correct classification of creditworthiness.
Rule Coverage1.00Fraction of predictions supported by domain rules in the KB.
Rule-consistent
Predictions
300/300Number of predictions fully explainable by symbolic logic.
Mismatch Penalty0Instances where statistical outputs conflicted with KB rules.
Table 4. Justification matrix: explanation quality in the NDR framework.
Table 4. Justification matrix: explanation quality in the NDR framework.
DimensionDesign Mechanism in NDRJustification
TrustworthyDeductive reasoning; causal law constraints; rule-consistency checkEnsures epistemic soundness; aligns predictions with domain knowledge
Human-readablePlain-language “if–then” format; feature-based explanationsClear, structured, and easy to interpret by non-expert users.
ActionableHighlights modifiable inputs; grounded in formal policies/lawsEnables users to take informed actions; supports operational decision-making.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Hakizimana, G.; Ledezma Espino, A. Nomological Deductive Reasoning for Trustworthy, Human-Readable, and Actionable AI Outputs. Algorithms 2025, 18, 306. https://doi.org/10.3390/a18060306

AMA Style

Hakizimana G, Ledezma Espino A. Nomological Deductive Reasoning for Trustworthy, Human-Readable, and Actionable AI Outputs. Algorithms. 2025; 18(6):306. https://doi.org/10.3390/a18060306

Chicago/Turabian Style

Hakizimana, Gedeon, and Agapito Ledezma Espino. 2025. "Nomological Deductive Reasoning for Trustworthy, Human-Readable, and Actionable AI Outputs" Algorithms 18, no. 6: 306. https://doi.org/10.3390/a18060306

APA Style

Hakizimana, G., & Ledezma Espino, A. (2025). Nomological Deductive Reasoning for Trustworthy, Human-Readable, and Actionable AI Outputs. Algorithms, 18(6), 306. https://doi.org/10.3390/a18060306

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop